WorldWideScience

Sample records for based grid job

  1. Application of rule-based data mining techniques to real time ATLAS Grid job monitoring data

    CERN Document Server

    Ahrens, R; The ATLAS collaboration; Kalinin, S; Maettig, P; Sandhoff, M; dos Santos, T; Volkmer, F

    2012-01-01

    The Job Execution Monitor (JEM) is a job-centric grid job monitoring software developed at the University of Wuppertal and integrated into the pilot-based “PanDA” job brokerage system leveraging physics analysis and Monte Carlo event production for the ATLAS experiment on the Worldwide LHC Computing Grid (WLCG). With JEM, job progress and grid worker node health can be supervised in real time by users, site admins and shift personnel. Imminent error conditions can be detected early and countermeasures can be initiated by the Job’s owner immideatly. Grid site admins can access aggregated data of all monitored jobs to infer the site status and to detect job and Grid worker node misbehaviour. Shifters can use the same aggregated data to quickly react to site error conditions and broken production tasks. In this work, the application of novel data-centric rule based methods and data-mining techniques to the real time monitoring data is discussed. The usage of such automatic inference techniques on monitorin...

  2. Mediated definite delegation - Certified Grid jobs in ALICE and beyond

    Science.gov (United States)

    Schreiner, Steffen; Grigoras, Costin; Litmaath, Maarten; Betev, Latchezar; Buchmann, Johannes

    2012-12-01

    Grid computing infrastructures need to provide traceability and accounting of their users’ activity and protection against misuse and privilege escalation, where the delegation of privileges in the course of a job submission is a key concern. This work describes an improved handling of Multi-user Grid Jobs in the ALICE Grid Services. A security analysis of the ALICE Grid job model is presented with derived security objectives, followed by a discussion of existing approaches of unrestricted delegation based on X.509 proxy certificates and the Grid middleware gLExec. Unrestricted delegation has severe security consequences and limitations, most importantly allowing for identity theft and forgery of jobs and data. These limitations are discussed and formulated, both in general and with respect to an adoption in line with Multi-user Grid Jobs. A new general model of mediated definite delegation is developed, allowing a broker to dynamically process and assign Grid jobs to agents while providing strong accountability and long-term traceability. A prototype implementation allowing for fully certified Grid jobs is presented as well as a potential interaction with gLExec. The achieved improvements regarding system security, malicious job exploitation, identity protection, and accountability are emphasized, including a discussion of non-repudiation in the face of malicious Grid jobs.

  3. Mediated definite delegation - Certified Grid jobs in ALICE and beyond

    International Nuclear Information System (INIS)

    Schreiner, Steffen; Buchmann, Johannes; Grigoras, Costin; Litmaath, Maarten; Betev, Latchezar

    2012-01-01

    Grid computing infrastructures need to provide traceability and accounting of their users’ activity and protection against misuse and privilege escalation, where the delegation of privileges in the course of a job submission is a key concern. This work describes an improved handling of Multi-user Grid Jobs in the ALICE Grid Services. A security analysis of the ALICE Grid job model is presented with derived security objectives, followed by a discussion of existing approaches of unrestricted delegation based on X.509 proxy certificates and the Grid middleware gLExec. Unrestricted delegation has severe security consequences and limitations, most importantly allowing for identity theft and forgery of jobs and data. These limitations are discussed and formulated, both in general and with respect to an adoption in line with Multi-user Grid Jobs. A new general model of mediated definite delegation is developed, allowing a broker to dynamically process and assign Grid jobs to agents while providing strong accountability and long-term traceability. A prototype implementation allowing for fully certified Grid jobs is presented as well as a potential interaction with gLExec. The achieved improvements regarding system security, malicious job exploitation, identity protection, and accountability are emphasized, including a discussion of non-repudiation in the face of malicious Grid jobs.

  4. A Mediated Definite Delegation Model allowing for Certified Grid Job Submission

    CERN Document Server

    Schreiner, Steffen; Grigoras, Costin; Litmaath, Maarten

    2012-01-01

    Grid computing infrastructures need to provide traceability and accounting of their users" activity and protection against misuse and privilege escalation. A central aspect of multi-user Grid job environments is the necessary delegation of privileges in the course of a job submission. With respect to these generic requirements this document describes an improved handling of multi-user Grid jobs in the ALICE ("A Large Ion Collider Experiment") Grid Services. A security analysis of the ALICE Grid job model is presented with derived security objectives, followed by a discussion of existing approaches of unrestricted delegation based on X.509 proxy certificates and the Grid middleware gLExec. Unrestricted delegation has severe security consequences and limitations, most importantly allowing for identity theft and forgery of delegated assignments. These limitations are discussed and formulated, both in general and with respect to an adoption in line with multi-user Grid jobs. Based on the architecture of the ALICE...

  5. Job scheduling in a heterogenous grid environment

    Energy Technology Data Exchange (ETDEWEB)

    Oliker, Leonid; Biswas, Rupak; Shan, Hongzhang; Smith, Warren

    2004-02-11

    Computational grids have the potential for solving large-scale scientific problems using heterogeneous and geographically distributed resources. However, a number of major technical hurdles must be overcome before this potential can be realized. One problem that is critical to effective utilization of computational grids is the efficient scheduling of jobs. This work addresses this problem by describing and evaluating a grid scheduling architecture and three job migration algorithms. The architecture is scalable and does not assume control of local site resources. The job migration policies use the availability and performance of computer systems, the network bandwidth available between systems, and the volume of input and output data associated with each job. An extensive performance comparison is presented using real workloads from leading computational centers. The results, based on several key metrics, demonstrate that the performance of our distributed migration algorithms is significantly greater than that of a local scheduling framework and comparable to a non-scalable global scheduling approach.

  6. Performance of R-GMA based grid job monitoring system for CMS data production

    CERN Document Server

    Byrom, Robert; Fisher, Steve M; Grandi, Claudio; Hobson, Peter R; Kyberd, Paul; MacEvoy, Barry; Nebrensky, Jindrich Josef; Tallini, Hugh; Traylen, Stephen

    2004-01-01

    High Energy Physics experiments, such as the Compact Muon Solenoid (CMS) at the CERN laboratory in Geneva, have large-scale data processing requirements, with stored data accumulating at a rate of 1 Gbyte/s. This load comfortably exceeds any previous processing requirements and we believe it may be most efficiently satisfied through Grid computing. Management of large Monte Carlo productions (~3000 jobs) or data analyses and the quality assurance of the results requires careful monitoring and bookkeeping, and an important requirement when using the Grid is the ability to monitor transparently the large number of jobs that are being executed simultaneously at multiple remote sites. R-GMA is a monitoring and information management service for distributed resources based on the Grid Monitoring Architecture of the Global Grid Forum. We have previously developed a system allowing us to test its performance under a heavy load while using few real Grid resources. We present the latest results on this system and comp...

  7. Grid workflow job execution service 'Pilot'

    Science.gov (United States)

    Shamardin, Lev; Kryukov, Alexander; Demichev, Andrey; Ilyin, Vyacheslav

    2011-12-01

    'Pilot' is a grid job execution service for workflow jobs. The main goal for the service is to automate computations with multiple stages since they can be expressed as simple workflows. Each job is a directed acyclic graph of tasks and each task is an execution of something on a grid resource (or 'computing element'). Tasks may be submitted to any WS-GRAM (Globus Toolkit 4) service. The target resources for the tasks execution are selected by the Pilot service from the set of available resources which match the specific requirements from the task and/or job definition. Some simple conditional execution logic is also provided. The 'Pilot' service is built on the REST concepts and provides a simple API through authenticated HTTPS. This service is deployed and used in production in a Russian national grid project GridNNN.

  8. Grid workflow job execution service 'Pilot'

    International Nuclear Information System (INIS)

    Shamardin, Lev; Kryukov, Alexander; Demichev, Andrey; Ilyin, Vyacheslav

    2011-01-01

    'Pilot' is a grid job execution service for workflow jobs. The main goal for the service is to automate computations with multiple stages since they can be expressed as simple workflows. Each job is a directed acyclic graph of tasks and each task is an execution of something on a grid resource (or 'computing element'). Tasks may be submitted to any WS-GRAM (Globus Toolkit 4) service. The target resources for the tasks execution are selected by the Pilot service from the set of available resources which match the specific requirements from the task and/or job definition. Some simple conditional execution logic is also provided. The 'Pilot' service is built on the REST concepts and provides a simple API through authenticated HTTPS. This service is deployed and used in production in a Russian national grid project GridNNN.

  9. Minimizing draining waste through extending the lifetime of pilot jobs in Grid environments

    International Nuclear Information System (INIS)

    Sfiligoi, I; Martin, T; Würthwein, F; Bockelman, B P; Bradley, D C

    2014-01-01

    The computing landscape is moving at an accelerated pace to many-core computing. Nowadays, it is not unusual to get 32 cores on a single physical node. As a consequence, there is increased pressure in the pilot systems domain to move from purely single-core scheduling and allow multi-core jobs as well. In order to allow for a gradual transition from single-core to multi-core user jobs, it is envisioned that pilot jobs will have to handle both kinds of user jobs at the same time, by requesting several cores at a time from Grid providers and then partitioning them between the user jobs at runtime. Unfortunately, the current Grid ecosystem only allows for relatively short lifetime of pilot jobs, requiring frequent draining, with the relative waste of compute resources due to varying lifetimes of the user jobs. Significantly extending the lifetime of pilot jobs is thus highly desirable, but must come without any adverse effects for the Grid resource providers. In this paper we present a mechanism, based on communication between the pilot jobs and the Grid provider, that allows for pilot jobs to run for extended periods of time when there are available resources, but also allows the Grid provider to reclaim the resources in a short amount of time when needed. We also present the experience of running a prototype system using the above mechanism on a few US-based Grid sites.

  10. GridCom, Grid Commander: graphical interface for Grid jobs and data management

    International Nuclear Information System (INIS)

    Galaktionov, V.V.

    2011-01-01

    GridCom - the software package for maintenance of automation of access to means of distributed system Grid (jobs and data). The client part, executed in the form of Java-applets, realises the Web-interface access to Grid through standard browsers. The executive part Lexor (LCG Executor) is started by the user in UI (User Interface) machine providing performance of Grid operations

  11. Job Flow Distribution and Ranked Jobs Scheduling in Grid Virtual Organizations

    CERN Document Server

    Toporkov, Victor; Tselishchev, Alexey; Yemelyanov, Dmitry; Potekhin, Petr

    2015-01-01

    In this work, we consider the problems of job flow distribution and ranked job framework forming within a model of cycle scheduling in Grid virtual organizations. The problem of job flow distribution is solved in terms of jobs and computing resource domains compatibility. A coefficient estimating such compatibility is introduced and studied experimentally. Two distribution strategies are suggested. Job framework forming is justified with such quality of service indicators as an average job execution time, a number of required scheduling cycles, and a number of job execution declines. Two methods for job selection and scheduling are proposed and compared: the first one is based on the knapsack problem solution, while the second one utilizes the mentioned compatibility coefficient. Along with these methods we present experimental results demonstrating the efficiency of proposed approaches and compare them with random job selection.

  12. Job execution in virtualized runtime environments in grid

    International Nuclear Information System (INIS)

    Shamardin, Lev; Demichev, Andrey; Gorbunov, Ilya; Ilyin, Slava; Kryukov, Alexander

    2010-01-01

    Grid systems are used for calculations and data processing in various applied areas such as biomedicine, nanotechnology and materials science, cosmophysics and high energy physics as well as in a number of industrial and commercial areas. Traditional method of execution of jobs in grid is running jobs directly on the cluster nodes. This puts restrictions on the choice of the operational environment to the operating system of the node and also does not allow to enforce resource sharing policies or jobs isolation nor guarantee minimal level of available system resources. We propose a new approach to running jobs on the cluster nodes when each grid job runs in its own virtual environment. This allows to use different operating systems for different jobs on the same nodes in cluster, provides better isolation between running jobs and allows to enforce resource sharing policies. The implementation of the proposed approach was made in the framework of gLite middleware of the EGEE/WLCG project and was successfully tested in SINP MSU. The implementation is transparent for the grid user and allows to submit binaries compiled for various operating systems using exactly the same gLite interface. Virtual machine images with the standard gLite worker node software and sample MS Windows execution environment were created.

  13. Real Time Monitor of Grid job executions

    International Nuclear Information System (INIS)

    Colling, D J; Martyniak, J; McGough, A S; Krenek, A; Sitera, J; Mulac, M; Dvorak, F

    2010-01-01

    In this paper we describe the architecture and operation of the Real Time Monitor (RTM), developed by the Grid team in the HEP group at Imperial College London. This is arguably the most popular dissemination tool within the EGEE [1] Grid. Having been used, on many occasions including GridFest and LHC inauguration events held at CERN in October 2008. The RTM gathers information from EGEE sites hosting Logging and Bookkeeping (LB) services. Information is cached locally at a dedicated server at Imperial College London and made available for clients to use in near real time. The system consists of three main components: the RTM server, enquirer and an apache Web Server which is queried by clients. The RTM server queries the LB servers at fixed time intervals, collecting job related information and storing this in a local database. Job related data includes not only job state (i.e. Scheduled, Waiting, Running or Done) along with timing information but also other attributes such as Virtual Organization and Computing Element (CE) queue - if known. The job data stored in the RTM database is read by the enquirer every minute and converted to an XML format which is stored on a Web Server. This decouples the RTM server database from the client removing the bottleneck problem caused by many clients simultaneously accessing the database. This information can be visualized through either a 2D or 3D Java based client with live job data either being overlaid on to a 2 dimensional map of the world or rendered in 3 dimensions over a globe map using OpenGL.

  14. Data location-aware job scheduling in the grid. Application to the GridWay metascheduler

    International Nuclear Information System (INIS)

    Delgado Peris, Antonio; Hernandez, Jose; Huedo, Eduardo; Llorente, Ignacio M

    2010-01-01

    Grid infrastructures constitute nowadays the core of the computing facilities of the biggest LHC experiments. These experiments produce and manage petabytes of data per year and run thousands of computing jobs every day to process that data. It is the duty of metaschedulers to allocate the tasks to the most appropriate resources at the proper time. Our work reviews the policies that have been proposed for the scheduling of grid jobs in the context of very data-intensive applications. We indicate some of the practical problems that such models will face and describe what we consider essential characteristics of an optimum scheduling system: aim to minimise not only job turnaround time but also data replication, flexibility to support different virtual organisation requirements and capability to coordinate the tasks of data placement and job allocation while keeping their execution decoupled. These ideas have guided the development of an enhanced prototype for GridWay, a general purpose metascheduler, part of the Globus Toolkit and member of the EGEE's RESPECT program. Current GridWay's scheduling algorithm is unaware of data location. Our prototype makes it possible for job requests to set data needs not only as absolute requirements but also as functions for resource ranking. As our tests show, this makes it more flexible than currently used resource brokers to implement different data-aware scheduling algorithms.

  15. The Grid[Way] Job Template Manager, a tool for parameter sweeping

    Science.gov (United States)

    Lorca, Alejandro; Huedo, Eduardo; Llorente, Ignacio M.

    2011-04-01

    Parameter sweeping is a widely used algorithmic technique in computational science. It is specially suited for high-throughput computing since the jobs evaluating the parameter space are loosely coupled or independent. A tool that integrates the modeling of a parameter study with the control of jobs in a distributed architecture is presented. The main task is to facilitate the creation and deletion of job templates, which are the elements describing the jobs to be run. Extra functionality relies upon the GridWay Metascheduler, acting as the middleware layer for job submission and control. It supports interesting features like multi-dimensional sweeping space, wildcarding of parameters, functional evaluation of ranges, value-skipping and job template automatic indexation. The use of this tool increases the reliability of the parameter sweep study thanks to the systematic bookkeeping of job templates and respective job statuses. Furthermore, it simplifies the porting of the target application to the grid reducing the required amount of time and effort. Program summaryProgram title: Grid[Way] Job Template Manager (version 1.0) Catalogue identifier: AEIE_v1_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AEIE_v1_0.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: Apache license 2.0 No. of lines in distributed program, including test data, etc.: 3545 No. of bytes in distributed program, including test data, etc.: 126 879 Distribution format: tar.gz Programming language: Perl 5.8.5 and above Computer: Any (tested on PC x86 and x86_64) Operating system: Unix, GNU/Linux (tested on Ubuntu 9.04, Scientific Linux 4.7, centOS 5.4), Mac OS X (tested on Snow Leopard 10.6) RAM: 10 MB Classification: 6.5 External routines: The GridWay Metascheduler [1]. Nature of problem: To parameterize and manage an application running on a grid or cluster. Solution method: Generation of job templates as a cross product of

  16. A Grid job monitoring system

    Energy Technology Data Exchange (ETDEWEB)

    Dumitrescu, Catalin [Fermi National Accelerator Laboratory (United States); Nowack, Andreas [RWTH Aachen (Germany); Padhi, Sanjay [University of California San Diego (United States); Sarkar, Subir, E-mail: subir.sarkar@cern.c [INFN, Sezione di Pisa and Scuola Normale Superiore, Pisa (Italy)

    2010-04-01

    This paper presents a web-based Job Monitoring framework for individual Grid sites that allows users to follow in detail their jobs in quasi-real time. The framework consists of several independent components : (a) a set of sensors that run on the site CE and worker nodes and update a database, (b) a simple yet extensible web services framework and (c) an Ajax powered web interface having a look-and-feel and control similar to a desktop application. The monitoring framework supports LSF, Condor and PBS-like batch systems. This is one of the first monitoring systems where an X.509 authenticated web interface can be seamlessly accessed by both end-users and site administrators. While a site administrator has access to all the possible information, a user can only view the jobs for the Virtual Organizations (VO) he/she is a part of. The monitoring framework design supports several possible deployment scenarios. For a site running a supported batch system, the system may be deployed as a whole, or existing site sensors can be adapted and reused with the web services components. A site may even prefer to build the web server independently and choose to use only the Ajax powered web interface. Finally, the system is being used to monitor a glideinWMS instance. This broadens the scope significantly, allowing it to monitor jobs over multiple sites.

  17. A grid job monitoring system

    International Nuclear Information System (INIS)

    Dumitrescu, Catalin; Nowack, Andreas; Padhi, Sanjay; Sarkar, Subir

    2010-01-01

    This paper presents a web-based Job Monitoring framework for individual Grid sites that allows users to follow in detail their jobs in quasi-real time. The framework consists of several independent components: (a) a set of sensors that run on the site CE and worker nodes and update a database, (b) a simple yet extensible web services framework and (c) an Ajax powered web interface having a look-and-feel and control similar to a desktop application. The monitoring framework supports LSF, Condor and PBS-like batch systems. This is one of the first monitoring systems where an X.509 authenticated web interface can be seamlessly accessed by both end-users and site administrators. While a site administrator has access to all the possible information, a user can only view the jobs for the Virtual Organizations (VO) he/she is a part of. The monitoring framework design supports several possible deployment scenarios. For a site running a supported batch system, the system may be deployed as a whole, or existing site sensors can be adapted and reused with the web services components. A site may even prefer to build the web server independently and choose to use only the Ajax powered web interface. Finally, the system is being used to monitor a glideinWMS instance. This broadens the scope significantly, allowing it to monitor jobs over multiple sites.

  18. Scalability tests of R-GMA based Grid job monitoring system for CMS Monte Carlo data production

    CERN Document Server

    Bonacorsi, D; Field, L; Fisher, S; Grandi, C; Hobson, P R; Kyberd, P; MacEvoy, B; Nebrensky, J J; Tallini, H; Traylen, S

    2004-01-01

    High Energy Physics experiments such as CMS (Compact Muon Solenoid) at the Large Hadron Collider have unprecedented, large-scale data processing computing requirements, with data accumulating at around 1 Gbyte/s. The Grid distributed computing paradigm has been chosen as the solution to provide the requisite computing power. The demanding nature of CMS software and computing requirements, such as the production of large quantities of Monte Carlo simulated data, makes them an ideal test case for the Grid and a major driver for the development of Grid technologies. One important challenge when using the Grid for large-scale data analysis is the ability to monitor the large numbers of jobs that are being executed simultaneously at multiple remote sites. R-GMA is a monitoring and information management service for distributed resources based on the Grid Monitoring Architecture of the Global Grid Forum. In this paper we report on the first measurements of R-GMA as part of a monitoring architecture to be used for b...

  19. GridCom, Grid Commander: graphical interface for Grid jobs and data management; GridCom, Grid Commander: graficheskij interfejs dlya raboty s zadachami i dannymi v gride

    Energy Technology Data Exchange (ETDEWEB)

    Galaktionov, V V

    2011-07-01

    GridCom - the software package for maintenance of automation of access to means of distributed system Grid (jobs and data). The client part, executed in the form of Java-applets, realises the Web-interface access to Grid through standard browsers. The executive part Lexor (LCG Executor) is started by the user in UI (User Interface) machine providing performance of Grid operations

  20. Remote data access in computational jobs on the ATLAS data grid

    CERN Document Server

    Begy, Volodimir; The ATLAS collaboration; Lassnig, Mario

    2018-01-01

    This work describes the technique of remote data access from computational jobs on the ATLAS data grid. In comparison to traditional data movement and stage-in approaches it is well suited for data transfers which are asynchronous with respect to the job execution. Hence, it can be used for optimization of data access patterns based on various policies. In this study, remote data access is realized with the HTTP and WebDAV protocols, and is investigated in the context of intra- and inter-computing site data transfers. In both cases, the typical scenarios for application of remote data access are identified. The paper also presents an analysis of parameters influencing the data goodput between heterogeneous storage element - worker node pairs on the grid.

  1. Biomedical applications on the GRID efficient management of parallel jobs

    CERN Document Server

    Moscicki, Jakub T; Lee Hurng Chun; Lin, S C; Pia, Maria Grazia

    2004-01-01

    Distributed computing based on the Master-Worker and PULL interaction model is applicable to a number of applications in high energy physics, medical physics and bio-informatics. We demonstrate a realistic medical physics use-case of a dosimetric system for brachytherapy using distributed Grid resources. We present the efficient techniques for running parallel jobs in a case of the BLAST, a gene sequencing application, as well as for the Monte Carlo simulation based on Geant4. We present a strategy for improving the runtime performance and robustness of the jobs as well as for the minimization of the development time needed to migrate the applications to a distributed environment.

  2. Exploring virtualisation tools with a new virtualisation provisioning method to test dynamic grid environments for ALICE grid jobs over ARC grid middleware

    International Nuclear Information System (INIS)

    Wagner, B; Kileng, B

    2014-01-01

    The Nordic Tier-1 centre for LHC is distributed over several computing centres. It uses ARC as the internal computing grid middleware. ALICE uses its own grid middleware AliEn to distribute jobs and the necessary software application stack. To make use of most of the AliEn infrastructure and software deployment methods for running ALICE grid jobs on ARC, we are investigating different possible virtualisation technologies. For this a testbed and possible framework for bridging different middleware systems is under development. It allows us to test a variety of virtualisation methods and software deployment technologies in the form of different virtual machines.

  3. Multiobjective Variable Neighborhood Search algorithm for scheduling independent jobs on computational grid

    Directory of Open Access Journals (Sweden)

    S. Selvi

    2015-07-01

    Full Text Available Grid computing solves high performance and high-throughput computing problems through sharing resources ranging from personal computers to super computers distributed around the world. As the grid environments facilitate distributed computation, the scheduling of grid jobs has become an important issue. In this paper, an investigation on implementing Multiobjective Variable Neighborhood Search (MVNS algorithm for scheduling independent jobs on computational grid is carried out. The performance of the proposed algorithm has been evaluated with Min–Min algorithm, Simulated Annealing (SA and Greedy Randomized Adaptive Search Procedure (GRASP algorithm. Simulation results show that MVNS algorithm generally performs better than other metaheuristics methods.

  4. A Web portal for CMS Grid job submission and management

    Energy Technology Data Exchange (ETDEWEB)

    Braun, David [Department of Physics, Purdue University, W. Lafayette, IN 47907 (United States); Neumeister, Norbert, E-mail: neumeist@purdue.ed [Rosen Center for Advanced Computing, Purdue University, W. Lafayette, IN 47907 (United States)

    2010-04-01

    We present a Web portal for CMS Grid submission and management. The portal is built using a JBoss application server. It has a three tier architecture; presentation, business logic and data. Bean based business logic interacts with the underlying Grid infrastructure and pre-existing external applications, while the presentation layer uses AJAX to offer an intuitive, functional interface to the back-end. Application data aggregating information from the portal as well as the external applications is persisted to the server memory cache and then to a backend database. We describe how the portal exploits standard, off-the-shelf commodity software together with existing Grid infrastructures in order to facilitate job submission and monitoring for the CMS collaboration. This paper describes the design, development, current functionality and plans for future enhancements of the portal.

  5. A Web portal for CMS Grid job submission and management

    International Nuclear Information System (INIS)

    Braun, David; Neumeister, Norbert

    2010-01-01

    We present a Web portal for CMS Grid submission and management. The portal is built using a JBoss application server. It has a three tier architecture; presentation, business logic and data. Bean based business logic interacts with the underlying Grid infrastructure and pre-existing external applications, while the presentation layer uses AJAX to offer an intuitive, functional interface to the back-end. Application data aggregating information from the portal as well as the external applications is persisted to the server memory cache and then to a backend database. We describe how the portal exploits standard, off-the-shelf commodity software together with existing Grid infrastructures in order to facilitate job submission and monitoring for the CMS collaboration. This paper describes the design, development, current functionality and plans for future enhancements of the portal.

  6. Job schedul in Grid batch farms

    International Nuclear Information System (INIS)

    Gellrich, Andreas

    2014-01-01

    We present here a study for a scheduler which cooperates with the queueing system TORQUE and is tailored to the needs of a HEP-dominated large Grid site with around 10000 jobs slots. Triggered by severe scaling problems of MAUI, a scheduler, referred to as MYSCHED, was developed and put into operation. We discuss conceptional aspects as well as experiences after almost two years of running.

  7. Renewable Energy Jobs. Status, prospects and policies. Biofuels and grid-connected electricity generation

    Energy Technology Data Exchange (ETDEWEB)

    Lucas, H; Ferroukhi, R [et al.; IRENA Policy Advisory Services and Capacity Building Directorate, Abu Dhabi (United Arab Emirates)

    2012-01-15

    Over the past years, interest has grown in the potential for the renewable energy industry to create jobs. Governments are seeking win-win solutions to the dual challenge of high unemployment and climate change. By 2010, USD 51 billion had been pledged to renewables in stimulus packages, and by early 2011 there were 119 countries with some kind of policy target and/or support policy for renewable energy, such as feed-in tariffs, quota obligations, favourable tax treatment and public loans or grants, many of which explicitly target job creation as a policy goal. Policy-makers in many countries are now designing renewable energy policies that aim to create new jobs, build industries and benefit particular geographic areas. But how much do we know for certain about the job creation potential for renewable energy? This working paper aims to provide an overview of current knowledge on five questions: (1) How can jobs in renewable energy be characterised?; (2) How are they shared out across the technology value chain and what skill levels are required?; (3) How many jobs currently exist and where are they in the world?; (4) How many renewable energy jobs could there be in the future?; and (5) What policy frameworks can be used to promote employment benefits from renewable energy? This paper focuses on grid-connected electricity generation technologies and biofuels. Since the employment potential of off-grid applications is large, it will be covered by a forthcoming study by IRENA on job creation in the context of energy access, based on a number of case studies.

  8. Grid Service for User-Centric Job

    Energy Technology Data Exchange (ETDEWEB)

    Lauret, Jerome

    2009-07-31

    The User Centric Monitoring (UCM) project was aimed at developing a toolkit that provides the Virtual Organization (VO) with tools to build systems that serve a rich set of intuitive job and application monitoring information to the VO’s scientists so that they can be more productive. The tools help collect and serve the status and error information through a Web interface. The proposed UCM toolkit is composed of a set of library functions, a database schema, and a Web portal that will collect and filter available job monitoring information from various resources and present it to users in a user-centric view rather than and administrative-centric point of view. The goal is to create a set of tools that can be used to augment grid job scheduling systems, meta-schedulers, applications, and script sets in order to provide the UCM information. The system provides various levels of an application programming interface that is useful through out the Grid environment and at the application level for logging messages, which are combined with the other user-centric monitoring information in a abstracted “data store”. A planned monitoring portal will also dynamically present the information to users in their web browser in a secure manor, which is also easily integrated into any JSR-compliant portal deployment that a VO might employ. The UCM is meant to be flexible and modular in the ways that it can be adopted to give the VO many choices to build a solution that works for them with special attention to the smaller VOs that do not have the resources to implement home-grown solutions.

  9. The swiss army knife of job submission tools: grid-control

    Science.gov (United States)

    Stober, F.; Fischer, M.; Schleper, P.; Stadie, H.; Garbers, C.; Lange, J.; Kovalchuk, N.

    2017-10-01

    grid-control is a lightweight and highly portable open source submission tool that supports all common workflows in high energy physics (HEP). It has been used by a sizeable number of HEP analyses to process tasks that sometimes consist of up to 100k jobs. grid-control is built around a powerful plugin and configuration system, that allows users to easily specify all aspects of the desired workflow. Job submission to a wide range of local or remote batch systems or grid middleware is supported. Tasks can be conveniently specified through the parameter space that will be processed, which can consist of any number of variables and data sources with complex dependencies on each other. Dataset information is processed through a configurable pipeline of dataset filters, partition plugins and partition filters. The partition plugins can take the number of files, size of the work units, metadata or combinations thereof into account. All changes to the input datasets or variables are propagated through the processing pipeline and can transparently trigger adjustments to the parameter space and the job submission. While the core functionality is completely experiment independent, full integration with the CMS computing environment is provided by a small set of plugins.

  10. A framework supporting the development of a Grid portal for analysis based on ROI.

    Science.gov (United States)

    Ichikawa, K; Date, S; Kaishima, T; Shimojo, S

    2005-01-01

    In our research on brain function analysis, users require two different simultaneous types of processing: interactive processing to a specific part of data and high-performance batch processing to an entire dataset. The difference between these two types of processing is in whether or not the analysis is for data in the region of interest (ROI). In this study, we propose a Grid portal that has a mechanism to freely assign computing resources to the users on a Grid environment according to the users' two different types of processing requirements. We constructed a Grid portal which integrates interactive processing and batch processing by the following two mechanisms. First, a job steering mechanism controls job execution based on user-tagged priority among organizations with heterogeneous computing resources. Interactive jobs are processed in preference to batch jobs by this mechanism. Second, a priority-based result delivery mechanism that administrates a rank of data significance. The portal ensures a turn-around time of interactive processing by the priority-based job controlling mechanism, and provides the users with quality of services (QoS) for interactive processing. The users can access the analysis results of interactive jobs in preference to the analysis results of batch jobs. The Grid portal has also achieved high-performance computation of MEG analysis with batch processing on the Grid environment. The priority-based job controlling mechanism has been realized to freely assign computing resources to the users' requirements. Furthermore the achievement of high-performance computation contributes greatly to the overall progress of brain science. The portal has thus made it possible for the users to flexibly include the large computational power in what they want to analyze.

  11. Ganga - an Optimiser and Front-End for Grid Job Submission (Demo)

    CERN Document Server

    Maier, A; Egede, U; Elmsheuser, J; Gaidioz, B; Harrison, K; Hurng-Chun Lee; Liko, D; Mosckicki, J; Muraru, A; Romanovsky, V; Soroko, A; Tan, C L; Koblitz, B

    2007-01-01

    The presentation will introduce the Ganga job-management system (http://cern.ch/ganga), developed as an ATLAS/LHCb common project. The main goal of Ganga is to provide a simple and consistent way of preparing, organising and executing analysis tasks, allowing physicists to concentrate on the algorithmic part without having to worry about technical details. Ganga provides a clean Python API that reduces and simplifies the work involved in preparing an application, organizing the submission, and gathering results. Technical details of submitting a job to the Grid, for example the preparation of a job-description file, are factored out and taken care of transparently by the system. By changing the parameter that identifies the execution backend, a user can trivially switch between running an application on a portable PC, running higherstatistics tests on a local batch system, and analysing all available statistics on the Grid. Although Ganga is being developed for LHCb and ATLAS, it is not limited to use with HE...

  12. Processing moldable tasks on the grid: Late job binding with lightweight user-level overlay

    CERN Document Server

    Moscicki, J T; Sloot, P M A; Lamanna, M

    2011-01-01

    Independent observations and everyday user experience indicate that performance and reliability of large grid infrastructures may suffer from large and unpredictable variations. In this paper we study the impact of the job queuing time on processing of moldable tasks which are commonly found in large-scale production grids. We use the mean value and variance of makespan as the quality of service indicators. We develop a general task processing model to provide a quantitative comparison between two models: early and late job binding in a user-level overlay applied to the EGEE Grid infrastructure. We find that the late-binding model effectively defines a transformation of the distribution of makespan according to the Central Limit Theorem. As demonstrated by Monte Carlo simulations using real job traces, this transformation allows to substantially reduce the mean value and variance of makespan. For certain classes of applications task granularity may be adjusted such that a speedup of an order of magnitude or m...

  13. An improved ant colony optimization algorithm with fault tolerance for job scheduling in grid computing systems.

    Directory of Open Access Journals (Sweden)

    Hajara Idris

    Full Text Available The Grid scheduler, schedules user jobs on the best available resource in terms of resource characteristics by optimizing job execution time. Resource failure in Grid is no longer an exception but a regular occurring event as resources are increasingly being used by the scientific community to solve computationally intensive problems which typically run for days or even months. It is therefore absolutely essential that these long-running applications are able to tolerate failures and avoid re-computations from scratch after resource failure has occurred, to satisfy the user's Quality of Service (QoS requirement. Job Scheduling with Fault Tolerance in Grid Computing using Ant Colony Optimization is proposed to ensure that jobs are executed successfully even when resource failure has occurred. The technique employed in this paper, is the use of resource failure rate, as well as checkpoint-based roll back recovery strategy. Check-pointing aims at reducing the amount of work that is lost upon failure of the system by immediately saving the state of the system. A comparison of the proposed approach with an existing Ant Colony Optimization (ACO algorithm is discussed. The experimental results of the implemented Fault Tolerance scheduling algorithm show that there is an improvement in the user's QoS requirement over the existing ACO algorithm, which has no fault tolerance integrated in it. The performance evaluation of the two algorithms was measured in terms of the three main scheduling performance metrics: makespan, throughput and average turnaround time.

  14. Job system generation in grid taking into account user preferences

    Directory of Open Access Journals (Sweden)

    D. M. Yemelyanov

    2016-01-01

    Full Text Available Distributed computing environments like Grid are characterized by heterogeneity, low cohesion and dynamic structure of computing nodes. This is why the task of resource scheduling in such environments is complex. Different approaches to job scheduling in grid exist. Some of them use economic principles. Economic approaches to scheduling have shown their efficiency. One of such approaches is cyclic scheduling scheme which is considered in this paper.Cyclic scheduling scheme takes into account the preferences of computing environment users by means of an optimization criterion, which is included in the resource request. Besides, the scheme works cyclically by scheduling a certain job batch at each scheduling step. This is why there is a preliminary scheduling step which is job batch generation.The purpose of this study was to estimate the infl uence of job batch structure by the user criterion on the degree of its satisfaction. In other words we had to find the best way to form the batch with relation to the user optimization criterion. For example if it is more efficient to form the batch with jobs with the same criterion value or with different criterion values. Also we wanted to find the combination of criterion values which would give the most efficient scheduling results.To achieve this purpose an experiment in a simulation environment was conducted. The experiment consisted of scheduling of job batches with different values of the user criterion, other parameters of the resource request and the characteristics of the computing environment being the same. Three job batch generation strategies were considered. In the first strategy the batch consisted of jobs with the same criterion value. In the second strategy the batch consisted of jobs with all the considered criteria equally likely. The third strategy was similar to the second one, but only two certain criteria were considered. The third strategy was considered in order to find the most

  15. On the Optimization of GLite-Based Job Submission

    International Nuclear Information System (INIS)

    Misurelli, Giuseppe; Veronesi, Paolo; Palmieri, Francesco; Pardi, Silvio

    2011-01-01

    A Grid is a very dynamic, complex and heterogeneous system, whose reliability can be adversely conditioned by several different factors such as communications and hardware faults, middleware bugs or wrong configurations due to human errors. As the infrastructure scales, spanning a large number of sites, each hosting hundreds or thousands of hosts/resources, the occurrence of runtime faults following job submission becomes a very frequent and phenomenon. Therefore, fault avoidance becomes a fundamental aim in modern Grids since the dependability of individual resources spread upon widely distributed computing infrastructures and often used outside of their native organizational boundaries, cannot be guaranteed in any systematic way. Accordingly, we propose a simple job optimization solution based on a user-driven fault avoidance strategy. Such strategy starts from the introduction within the grid information system of several on-line service-monitoring metrics that can be used as specific hints to the workload management system for driving resource discovery operations according to a fault-free resource-scheduling plan. This solution, whose main goal is to minimize the execution time by avoiding execution failures, demonstrated to be very effective in incrementing both the user perceivable quality and the overall grid performance.

  16. A security architecture for the ALICE grid services

    CERN Document Server

    Schreiner, Steffen; Buchmann, Johannes; Betev, Latchezar; Grigoras, Alina

    2012-01-01

    Globally distributed research cyberinfrastructures, like the ALICE Grid Services, need to provide traceability and accountability of operations and internal interactions. This document presents a new security architecture for the ALICE Grid Services, allowing to establish non-repudiation with respect to creatorship and ownership of Grid files and jobs. It is based on mutually authenticated and encrypted communication using X.509 Public Key Infrastructure and the Transport Layer Security (TLS) protocol. Introducing certified Grid file entries and signed Grid jobs by implementing a model of Mediated Definite Delegation it allows to establish long-term accountability concerning Grid jobs and files. Initial submissions as well as any alteration of Grid jobs are becoming verifiable and can be traced back to the originator. The architecture has been implemented as a prototype along with the development of a new central Grid middleware, called jAliEn.

  17. The LHCb Grid Simulation

    CERN Multimedia

    Baranov, Alexander

    2016-01-01

    The LHCb Grid access if based on the LHCbDirac system. It provides access to data and computational resources to researchers with different geographical locations. The Grid has a hierarchical topology with multiple sites distributed over the world. The sites differ from each other by their number of CPUs, amount of disk storage and connection bandwidth. These parameters are essential for the Grid work. Moreover, job scheduling and data distribution strategy have a great impact on the grid performance. However, it is hard to choose an appropriate algorithm and strategies as they need a lot of time to be tested on the real grid. In this study, we describe the LHCb Grid simulator. The simulator reproduces the LHCb Grid structure with its sites and their number of CPUs, amount of disk storage and bandwidth connection. We demonstrate how well the simulator reproduces the grid work, show its advantages and limitations. We show how well the simulator reproduces job scheduling and network anomalies, consider methods ...

  18. Efficient job handling in the GRID short deadline, interactivity, fault tolerance and parallelism

    CERN Document Server

    Moscicki, Jakub

    2006-01-01

    The major GRID infastructures are designed mainly for batch-oriented computing with coarse-grained jobs and relatively high job turnaround time. However many practical applications in natural and physical sciences may be easily parallelized and run as a set of smaller tasks which require little or no synchronization and which may be scheduled in a more efficient way. The Distributed Analysis Environment Framework (DIANE), is a Master-Worker execution skeleton for applications, which complements the GRID middleware stack. Automatic failure recovery and task dispatching policies enable an easy customization of the behaviour of the framework in a dynamic and non-reliable computing environment. We demonstrate the experience of using the framework with several diverse real-life applications, including Monte Carlo Simulation, Physics Data Analysis and Biotechnology. The interfacing of existing sequential applications from the point of view of non-expert user is made easy, also for legacy applications. We analyze th...

  19. Jobs masonry in LHCb with elastic Grid Jobs

    Science.gov (United States)

    Stagni, F.; Charpentier, Ph

    2015-12-01

    In any distributed computing infrastructure, a job is normally forbidden to run for an indefinite amount of time. This limitation is implemented using different technologies, the most common one being the CPU time limit implemented by batch queues. It is therefore important to have a good estimate of how much CPU work a job will require: otherwise, it might be killed by the batch system, or by whatever system is controlling the jobs’ execution. In many modern interwares, the jobs are actually executed by pilot jobs, that can use the whole available time in running multiple consecutive jobs. If at some point the available time in a pilot is too short for the execution of any job, it should be released, while it could have been used efficiently by a shorter job. Within LHCbDIRAC, the LHCb extension of the DIRAC interware, we developed a simple way to fully exploit computing capabilities available to a pilot, even for resources with limited time capabilities, by adding elasticity to production MonteCarlo (MC) simulation jobs. With our approach, independently of the time available, LHCbDIRAC will always have the possibility to execute a MC job, whose length will be adapted to the available amount of time: therefore the same job, running on different computing resources with different time limits, will produce different amounts of events. The decision on the number of events to be produced is made just in time at the start of the job, when the capabilities of the resource are known. In order to know how many events a MC job will be instructed to produce, LHCbDIRAC simply requires three values: the CPU-work per event for that type of job, the power of the machine it is running on, and the time left for the job before being killed. Knowing these values, we can estimate the number of events the job will be able to simulate with the available CPU time. This paper will demonstrate that, using this simple but effective solution, LHCb manages to make a more efficient use of

  20. MaGate Simulator: A Simulation Environment for a Decentralized Grid Scheduler

    Science.gov (United States)

    Huang, Ye; Brocco, Amos; Courant, Michele; Hirsbrunner, Beat; Kuonen, Pierre

    This paper presents a simulator for of a decentralized modular grid scheduler named MaGate. MaGate’s design emphasizes scheduler interoperability by providing intelligent scheduling serving the grid community as a whole. Each MaGate scheduler instance is able to deal with dynamic scheduling conditions, with continuously arriving grid jobs. Received jobs are either allocated on local resources, or delegated to other MaGates for remote execution. The proposed MaGate simulator is based on GridSim toolkit and Alea simulator, and abstracts the features and behaviors of complex fundamental grid elements, such as grid jobs, grid resources, and grid users. Simulation of scheduling tasks is supported by a grid network overlay simulator executing distributed ant-based swarm intelligence algorithms to provide services such as group communication and resource discovery. For evaluation, a comparison of behaviors of different collaborative policies among a community of MaGates is provided. Results support the use of the proposed approach as a functional ready grid scheduler simulator.

  1. Jobs masonry in LHCb with elastic Grid Jobs

    CERN Document Server

    Stagni, F

    2015-01-01

    In any distributed computing infrastructure, a job is normally forbidden to run for an indefinite amount of time. This limitation is implemented using different technologies, the most common one being the CPU time limit implemented by batch queues. It is therefore important to have a good estimate of how much CPU work a job will require: otherwise, it might be killed by the batch system, or by whatever system is controlling the jobs' execution. In many modern interwares, the jobs are actually executed by pilot jobs, that can use the whole available time in running multiple consecutive jobs. If at some point the available time in a pilot is too short for the execution of any job, it should be released, while it could have been used efficiently by a shorter job. Within LHCbDIRAC, the LHCb extension of the DIRAC interware, we developed a simple way to fully exploit computing capabilities available to a pilot, even for resources with limited time capabilities, by adding elasticity to production MonteCarlo (MC) si...

  2. The U.S. Smart Grid Revolution. KEMA's Perspectives for Job Creation

    Energy Technology Data Exchange (ETDEWEB)

    NONE

    2009-01-13

    Smart Grid is a vision for the electric delivery system of the future. The Smart Grid envisions an entirely transformed electrical infrastructure. It will embody a network of devices as vast, interconnected, automated, and interactive as the Internet. Utilities and consumers will accrue returns through the convergence of power delivery and information technologies to achieve improved reliability, reduced O and M costs, avoidance of new capacity, and increased customer satisfaction. Smart Grid includes advanced sensing, control, communications, and analytic technologies such as Advanced Metering and T and D Automation. The GridWise Alliance believes that Smart Grid will: Utilize information technologies to revolutionize energy systems as they have revolutionized other aspects of U.S. business; Create value for all participants by developing and deploying technology solutions that cross enterprise and regulatory boundaries; Enhance security and reliability through an information-rich power grid that is flexible and adaptive; and Empower consumers to benefit from their participation in the operation of the power grid. This job creation analysis focuses in detail on the Advanced Metering and T and D automation aspects of Smart Grid, but also touches on the related new business opportunities around energy services, renewable installation and services, and home automation.

  3. HLRmon: a role-based grid accounting report web tool

    International Nuclear Information System (INIS)

    Pra, S D; Fattibene, E; Misurelli, G; Pescarmona, F; Gaido, L

    2008-01-01

    Both Grid users and Grid operators need ways to get CPU usage statistics about jobs executed in a given time period at various different levels, depending on their specific Grid's role and rights. While a Grid user is interested in reports about its own jobs and should not get access to other's data, Site or Virtual Organization (VO) or Regional Operation Centre (ROC) manager would also like to see how resources are used through the Grid in a per Site or per VO basis, or both. The whole set of different reports turns out to be quite large, and various existing tools made to create them tend to better satisfy a single user's category, eventually despite of another. HLRmon results from our efforts to generate suitable reports for all existing categories and has been designed to serve them within a unified layout. Thanks to its ability to authenticate clients through certificate and related authorization rights, it can a-priori restrict the selectable items range offered to the web user, so that sensitive information can only be provided to specifically enabled people. Information are gathered by HLRmon from a Home Location Register (HLR) which stores complete accounting data in a per job basis. Depending on the kind of reports that are to be generated, it directly queries the HLR server using an ad-hoc Distributed Grid Accounting System (DGAS) query tool (tipically user's level detail info), or a local RDBMS table with daily aggregate information in a per Day, Site, VO basis, thus saving connection delay time and needless load on the HLR server

  4. Scheduling Network Traffic for Grid Purposes

    DEFF Research Database (Denmark)

    Gamst, Mette

    This thesis concerns scheduling of network traffic in grid context. Grid computing consists of a number of geographically distributed computers, which work together for solving large problems. The computers are connected through a network. When scheduling job execution in grid computing, data...... transmission has so far not been taken into account. This causes stability problems, because data transmission takes time and thus causes delays to the execution plan. This thesis proposes the integration of job scheduling and network routing. The scientific contribution is based on methods from operations...... research and consists of six papers. The first four considers data transmission in grid context. The last two solves the data transmission problem, where the number of paths per data connection is bounded from above. The thesis shows that it is possible to solve the integrated job scheduling and network...

  5. The Anatomy of a Grid portal

    International Nuclear Information System (INIS)

    Licari, Daniele; Calzolari, Federico

    2011-01-01

    In this paper we introduce a new way to deal with Grid portals referring to our implementation. L-GRID is a light portal to access the EGEE/EGI Grid infrastructure via Web, allowing users to submit their jobs from a common Web browser in a few minutes, without any knowledge about the Grid infrastructure. It provides the control over the complete lifecycle of a Grid Job, from its submission and status monitoring, to the output retrieval. The system, implemented as client-server architecture, is based on the Globus Grid middleware. The client side application is based on a java applet; the server relies on a Globus User Interface. There is no need of user registration on the server side, and the user needs only his own X.509 personal certificate. The system is user-friendly, secure (it uses SSL protocol, mechanism for dynamic delegation and identity creation in public key infrastructures), highly customizable, open source, and easy to install. The X.509 personal certificate does not get out from the local machine. It allows to reduce the time spent for the job submission, granting at the same time a higher efficiency and a better security level in proxy delegation and management.

  6. The Anatomy of a Grid portal

    Science.gov (United States)

    Licari, Daniele; Calzolari, Federico

    2011-12-01

    In this paper we introduce a new way to deal with Grid portals referring to our implementation. L-GRID is a light portal to access the EGEE/EGI Grid infrastructure via Web, allowing users to submit their jobs from a common Web browser in a few minutes, without any knowledge about the Grid infrastructure. It provides the control over the complete lifecycle of a Grid Job, from its submission and status monitoring, to the output retrieval. The system, implemented as client-server architecture, is based on the Globus Grid middleware. The client side application is based on a java applet; the server relies on a Globus User Interface. There is no need of user registration on the server side, and the user needs only his own X.509 personal certificate. The system is user-friendly, secure (it uses SSL protocol, mechanism for dynamic delegation and identity creation in public key infrastructures), highly customizable, open source, and easy to install. The X.509 personal certificate does not get out from the local machine. It allows to reduce the time spent for the job submission, granting at the same time a higher efficiency and a better security level in proxy delegation and management.

  7. A History-based Estimation for LHCb job requirements

    Science.gov (United States)

    Rauschmayr, Nathalie

    2015-12-01

    The main goal of a Workload Management System (WMS) is to find and allocate resources for the given tasks. The more and better job information the WMS receives, the easier will be to accomplish its task, which directly translates into higher utilization of resources. Traditionally, the information associated with each job, like expected runtime, is defined beforehand by the Production Manager in best case and fixed arbitrary values by default. In the case of LHCb's Workload Management System no mechanisms are provided which automate the estimation of job requirements. As a result, much more CPU time is normally requested than actually needed. Particularly, in the context of multicore jobs this presents a major problem, since single- and multicore jobs shall share the same resources. Consequently, grid sites need to rely on estimations given by the VOs in order to not decrease the utilization of their worker nodes when making multicore job slots available. The main reason for going to multicore jobs is the reduction of the overall memory footprint. Therefore, it also needs to be studied how memory consumption of jobs can be estimated. A detailed workload analysis of past LHCb jobs is presented. It includes a study of job features and their correlation with runtime and memory consumption. Following the features, a supervised learning algorithm is developed based on a history based prediction. The aim is to learn over time how jobs’ runtime and memory evolve influenced due to changes in experiment conditions and software versions. It will be shown that estimation can be notably improved if experiment conditions are taken into account.

  8. NetJobs: A new approach to network monitoring for the Grid using Grid jobs

    OpenAIRE

    Pagano, Alfredo

    2011-01-01

    With grid computing, the far-fl�ung and disparate IT resources act as a single "virtual datacenter". Grid computing interfaces heterogeneous IT resources so they are available when and where we need them. Grid allows us to provision applications and allocate capacity among research and business groups that are geographically and organizationally dispersed. Building a high availability Grid is hold as the next goal to achieve: protecting against computer failures and site failures to avoid dow...

  9. BaBar MC production on the Canadian grid using a web services approach

    Science.gov (United States)

    Agarwal, A.; Armstrong, P.; Desmarais, R.; Gable, I.; Popov, S.; Ramage, S.; Schaffer, S.; Sobie, C.; Sobie, R.; Sulivan, T.; Vanderster, D.; Mateescu, G.; Podaima, W.; Charbonneau, A.; Impey, R.; Viswanathan, M.; Quesnel, D.

    2008-07-01

    The present paper highlights the approach used to design and implement a web services based BaBar Monte Carlo (MC) production grid using Globus Toolkit version 4. The grid integrates the resources of two clusters at the University of Victoria, using the ClassAd mechanism provided by the Condor-G metascheduler. Each cluster uses the Portable Batch System (PBS) as its local resource management system (LRMS). Resource brokering is provided by the Condor matchmaking process, whereby the job and resource attributes are expressed as ClassAds. The important features of the grid are automatic registering of resource ClassAds to the central registry, ClassAds extraction from the registry to the metascheduler for matchmaking, and the incorporation of input/output file staging. Web-based monitoring is employed to track the status of grid resources and the jobs for an efficient operation of the grid. The performance of this new grid for BaBar jobs, and the existing Canadian computational grid (GridX1) based on Globus Toolkit version 2 is found to be consistent.

  10. BaBar MC production on the Canadian grid using a web services approach

    International Nuclear Information System (INIS)

    Agarwal, A; Armstrong, P; Desmarais, R; Gable, I; Popov, S; Ramage, S; Schaffer, S; Sobie, C; Sobie, R; Sulivan, T; Vanderster, D; Mateescu, G; Podaima, W; Charbonneau, A; Impey, R; Viswanathan, M; Quesnel, D

    2008-01-01

    The present paper highlights the approach used to design and implement a web services based BaBar Monte Carlo (MC) production grid using Globus Toolkit version 4. The grid integrates the resources of two clusters at the University of Victoria, using the ClassAd mechanism provided by the Condor-G metascheduler. Each cluster uses the Portable Batch System (PBS) as its local resource management system (LRMS). Resource brokering is provided by the Condor matchmaking process, whereby the job and resource attributes are expressed as ClassAds. The important features of the grid are automatic registering of resource ClassAds to the central registry, ClassAds extraction from the registry to the metascheduler for matchmaking, and the incorporation of input/output file staging. Web-based monitoring is employed to track the status of grid resources and the jobs for an efficient operation of the grid. The performance of this new grid for BaBar jobs, and the existing Canadian computational grid (GridX1) based on Globus Toolkit version 2 is found to be consistent

  11. FAULT TOLERANCE IN JOB SCHEDULING THROUGH FAULT MANAGEMENT FRAMEWORK USING SOA IN GRID

    Directory of Open Access Journals (Sweden)

    V. Indhumathi

    2017-01-01

    Full Text Available The rapid development in computing resources has enhanced the recital of computers and abridged their costs. This accessibility of low cost prevailing computers joined with the fame of the Internet and high-speed networks has leaded the computing surroundings to be mapped from dispersed to grid environments. Grid is a kind of dispersed system which supports the allotment and harmonized exploit of geographically dispersed and multi-owner resources, autonomously from their physical form and site, in vibrant practical organizations that carve up the similar objective of decipher large-scale applications. Thus any type of failure can happen at any point of time and job running in grid environment might fail. Therefore fault tolerance is an imperative and demanding concern in grid computing as the steadiness of individual grid resources may not be guaranteed. In order to build computational grids more effectual and consistent fault tolerant system is required. In order to accomplish the user prospect in terms of recital and competence, the Grid system desires SOA Fault Management Framework for the sharing of tasks with fault tolerance. A Fault Management Framework endeavor to pick up the response time of user’s proposed applications by ensures maximal exploitation of obtainable resources. The main aim is to avert, if probable, the stipulation where some processors are congested by means of a set of tasks while others are flippantly loaded or even at leisure.

  12. gLExec: gluing grid computing to the Unix world

    Science.gov (United States)

    Groep, D.; Koeroo, O.; Venekamp, G.

    2008-07-01

    The majority of compute resources in todays scientific grids are based on Unix and Unix-like operating systems. In this world, user and user-group management are based around the concepts of a numeric 'user ID' and 'group ID' that are local to the resource. In contrast, grid concepts of user and group management are centered around globally assigned identifiers and VO membership, structures that are independent of any specific resource. At the fabric boundary, these 'grid identities' have to be translated to Unix user IDs. New job submission methodologies, such as job-execution web services, community-deployed local schedulers, and the late binding of user jobs in a grid-wide overlay network of 'pilot jobs', push this fabric boundary ever further down into the resource. gLExec, a light-weight (and thereby auditable) credential mapping and authorization system, addresses these issues. It can be run both on fabric boundary, as part of an execution web service, and on the worker node in a late-binding scenario. In this contribution we describe the rationale for gLExec, how it interacts with the site authorization and credential mapping frameworks such as LCAS, LCMAPS and GUMS, and how it can be used to improve site control and traceability in a pilot-job system.

  13. gLExec: gluing grid computing to the Unix world

    International Nuclear Information System (INIS)

    Groep, D; Koeroo, O; Venekamp, G

    2008-01-01

    The majority of compute resources in todays scientific grids are based on Unix and Unix-like operating systems. In this world, user and user-group management are based around the concepts of a numeric 'user ID' and 'group ID' that are local to the resource. In contrast, grid concepts of user and group management are centered around globally assigned identifiers and VO membership, structures that are independent of any specific resource. At the fabric boundary, these 'grid identities' have to be translated to Unix user IDs. New job submission methodologies, such as job-execution web services, community-deployed local schedulers, and the late binding of user jobs in a grid-wide overlay network of 'pilot jobs', push this fabric boundary ever further down into the resource. gLExec, a light-weight (and thereby auditable) credential mapping and authorization system, addresses these issues. It can be run both on fabric boundary, as part of an execution web service, and on the worker node in a late-binding scenario. In this contribution we describe the rationale for gLExec, how it interacts with the site authorization and credential mapping frameworks such as LCAS, LCMAPS and GUMS, and how it can be used to improve site control and traceability in a pilot-job system

  14. Optimizing Resource Utilization in Grid Batch Systems

    International Nuclear Information System (INIS)

    Gellrich, Andreas

    2012-01-01

    On Grid sites, the requirements of the computing tasks (jobs) to computing, storage, and network resources differ widely. For instance Monte Carlo production jobs are almost purely CPU-bound, whereas physics analysis jobs demand high data rates. In order to optimize the utilization of the compute node resources, jobs must be distributed intelligently over the nodes. Although the job resource requirements cannot be deduced directly, jobs are mapped to POSIX UID/GID according to the VO, VOMS group and role information contained in the VOMS proxy. The UID/GID then allows to distinguish jobs, if users are using VOMS proxies as planned by the VO management, e.g. ‘role=production’ for Monte Carlo jobs. It is possible to setup and configure batch systems (queuing system and scheduler) at Grid sites based on these considerations although scaling limits were observed with the scheduler MAUI. In tests these limitations could be overcome with a home-made scheduler.

  15. Ganga: User-friendly Grid job submission and management tool for LHC and beyond

    International Nuclear Information System (INIS)

    Vanderster, D C; Gaidoz, B; Maier, A; Moscicki, J T; Muraru, A; Brochu, F; Cowan, G; Egede, U; Reece, W; Williams, M; Elmsheuser, J; Harrison, K; Slater, M; Tan, C L; Lee, H C; Liko, D; Pajchel, K; Samset, B; Soroko, A

    2010-01-01

    Ganga has been widely used for several years in ATLAS, LHCb and a handful of other communities. Ganga provides a simple yet powerful interface for submitting and managing jobs to a variety of computing backends. The tool helps users configuring applications and keeping track of their work. With the major release of version 5 in summer 2008, Ganga's main user-friendly features have been strengthened. Examples include a new configuration interface, enhanced support for job collections, bulk operations and easier access to subjobs. In addition to the traditional batch and Grid backends such as Condor, LSF, PBS, gLite/EDG a point-to-point job execution via ssh on remote machines is now supported. Ganga is used as an interactive job submission interface for end-users, and also as a job submission component for higher-level tools. For example GangaRobot is used to perform automated, end-to-end testing of distributed data analysis. Ganga comes with an extensive test suite covering more than 350 test cases. The development model involves all active developers in the release management shifts which is an important and novel approach for the distributed software collaborations. Ganga 5 is a mature, stable and widely-used tool with long-term support from the HEP community.

  16. Analyzing data flows of WLCG jobs at batch job level

    Science.gov (United States)

    Kuehn, Eileen; Fischer, Max; Giffels, Manuel; Jung, Christopher; Petzold, Andreas

    2015-05-01

    With the introduction of federated data access to the workflows of WLCG, it is becoming increasingly important for data centers to understand specific data flows regarding storage element accesses, firewall configurations, as well as the scheduling of batch jobs themselves. As existing batch system monitoring and related system monitoring tools do not support measurements at batch job level, a new tool has been developed and put into operation at the GridKa Tier 1 center for monitoring continuous data streams and characteristics of WLCG jobs and pilots. Long term measurements and data collection are in progress. These measurements already have been proven to be useful analyzing misbehaviors and various issues. Therefore we aim for an automated, realtime approach for anomaly detection. As a requirement, prototypes for standard workflows have to be examined. Based on measurements of several months, different features of HEP jobs are evaluated regarding their effectiveness for data mining approaches to identify these common workflows. The paper will introduce the actual measurement approach and statistics as well as the general concept and first results classifying different HEP job workflows derived from the measurements at GridKa.

  17. Optimal Grid Scheduling Using Improved Artificial Bee Colony Algorithm

    OpenAIRE

    T. Vigneswari; M. A. Maluk Mohamed

    2015-01-01

    Job Scheduling plays an important role for efficient utilization of grid resources available across different domains and geographical zones. Scheduling of jobs is challenging and NPcomplete. Evolutionary / Swarm Intelligence algorithms have been extensively used to address the NP problem in grid scheduling. Artificial Bee Colony (ABC) has been proposed for optimization problems based on foraging behaviour of bees. This work proposes a modified ABC algorithm, Cluster Hete...

  18. Porting of Scientific Applications to Grid Computing on GridWay

    Directory of Open Access Journals (Sweden)

    J. Herrera

    2005-01-01

    Full Text Available The expansion and adoption of Grid technologies is prevented by the lack of a standard programming paradigm to port existing applications among different environments. The Distributed Resource Management Application API has been proposed to aid the rapid development and distribution of these applications across different Distributed Resource Management Systems. In this paper we describe an implementation of the DRMAA standard on a Globus-based testbed, and show its suitability to express typical scientific applications, like High-Throughput and Master-Worker applications. The DRMAA routines are supported by the functionality offered by the GridWay2 framework, which provides the runtime mechanisms needed for transparently executing jobs on a dynamic Grid environment based on Globus. As cases of study, we consider the implementation with DRMAA of a bioinformatics application, a genetic algorithm and the NAS Grid Benchmarks.

  19. Parallel Monte Carlo simulations on an ARC-enabled computing grid

    International Nuclear Information System (INIS)

    Nilsen, Jon K; Samset, Bjørn H

    2011-01-01

    Grid computing opens new possibilities for running heavy Monte Carlo simulations of physical systems in parallel. The presentation gives an overview of GaMPI, a system for running an MPI-based random walker simulation on grid resources. Integrating the ARC middleware and the new storage system Chelonia with the Ganga grid job submission and control system, we show that MPI jobs can be run on a world-wide computing grid with good performance and promising scaling properties. Results for relatively communication-heavy Monte Carlo simulations run on multiple heterogeneous, ARC-enabled computing clusters in several countries are presented.

  20. A History-based Estimation for LHCb job requirements

    CERN Document Server

    Rauschmayr, Nathalie

    2015-01-01

    The main goal of a Workload Management System (WMS) is to find and allocate resources for the given tasks. The more and better job information the WMS receives, the easier will be to accomplish its task, which directly translates into higher utilization of resources. Traditionally, the information associated with each job, like expected runtime, is defined beforehand by the Production Manager in best case and fixed arbitrary values by default. In the case of LHCb's Workload Management System no mechanisms are provided which automate the estimation of job requirements. As a result, much more CPU time is normally requested than actually needed. Particularly, in the context of multicore jobs this presents a major problem, since single- and multicore jobs shall share the same resources. Consequently, grid sites need to rely on estimations given by the VOs in order to not decrease the utilization of their worker nodes when making multicore job slots available. The main reason for going to multicore jobs is the red...

  1. Using Grid for the BABAR Experiment

    International Nuclear Information System (INIS)

    Bozzi, C.

    2005-01-01

    The BaBar experiment has been taking data since 1999. In 2001 the computing group started to evaluate the possibility to evolve toward a distributed computing model in a grid environment. We built a prototype system, based on the European Data Grid (EDG), to submit full-scale analysis and Monte Carlo simulation jobs. Computing elements, storage elements, and worker nodes have been installed at SLAC and at various European sites. A BaBar virtual organization (VO) and a test replica catalog (RC) are maintained in Manchester, U.K., and the experiment is using three EDG testbed resource brokers in the U.K. and in Italy. First analysis tests were performed under the assumption that a standard BaBar software release was available at the grid target sites, using RC to register information about the executable and the produced n-tuples. Hundreds of analysis jobs accessing either Objectivity or Root data files ran on the grid. We tested the Monte Carlo production using a farm of the INFN-grid testbed customized to install an Objectivity database and run BaBar simulation software. First simulation production tests were performed using standard Job Description Language commands and the output files were written on the closest storage element. A package that can be officially distributed to grid sites not specifically customized for BaBar has been prepared. We are studying the possibility to add a user friendly interface to access grid services for BaBar

  2. The Grid2003 Production Grid Principles and Practice

    CERN Document Server

    Foster, I; Gose, S; Maltsev, N; May, E; Rodríguez, A; Sulakhe, D; Vaniachine, A; Shank, J; Youssef, S; Adams, D; Baker, R; Deng, W; Smith, J; Yu, D; Legrand, I; Singh, S; Steenberg, C; Xia, Y; Afaq, A; Berman, E; Annis, J; Bauerdick, L A T; Ernst, M; Fisk, I; Giacchetti, L; Graham, G; Heavey, A; Kaiser, J; Kuropatkin, N; Pordes, R; Sekhri, V; Weigand, J; Wu, Y; Baker, K; Sorrillo, L; Huth, J; Allen, M; Grundhoefer, L; Hicks, J; Luehring, F C; Peck, S; Quick, R; Simms, S; Fekete, G; Van den Berg, J; Cho, K; Kwon, K; Son, D; Park, H; Canon, S; Jackson, K; Konerding, D E; Lee, J; Olson, D; Sakrejda, I; Tierney, B; Green, M; Miller, R; Letts, J; Martin, T; Bury, D; Dumitrescu, C; Engh, D; Gardner, R; Mambelli, M; Smirnov, Y; Voeckler, J; Wilde, M; Zhao, Y; Zhao, X; Avery, P; Cavanaugh, R J; Kim, B; Prescott, C; Rodríguez, J; Zahn, A; McKee, S; Jordan, C; Prewett, J; Thomas, T; Severini, H; Clifford, B; Deelman, E; Flon, L; Kesselman, C; Mehta, G; Olomu, N; Vahi, K; De, K; McGuigan, P; Sosebee, M; Bradley, D; Couvares, P; De Smet, A; Kireyev, C; Paulson, E; Roy, A; Koranda, S; Moe, B; Brown, B; Sheldon, P

    2004-01-01

    The Grid2003 Project has deployed a multi-virtual organization, application-driven grid laboratory ("GridS") that has sustained for several months the production-level services required by physics experiments of the Large Hadron Collider at CERN (ATLAS and CMS), the Sloan Digital Sky Survey project, the gravitational wave search experiment LIGO, the BTeV experiment at Fermilab, as well as applications in molecular structure analysis and genome analysis, and computer science research projects in such areas as job and data scheduling. The deployed infrastructure has been operating since November 2003 with 27 sites, a peak of 2800 processors, work loads from 10 different applications exceeding 1300 simultaneous jobs, and data transfers among sites of greater than 2 TB/day. We describe the principles that have guided the development of this unique infrastructure and the practical experiences that have resulted from its creation and use. We discuss application requirements for grid services deployment and configur...

  3. Integrated Job Scheduling and Network Routing

    DEFF Research Database (Denmark)

    Gamst, Mette; Pisinger, David

    2013-01-01

    We consider an integrated job scheduling and network routing problem which appears in Grid Computing and production planning. The problem is to schedule a number of jobs at a finite set of machines, such that the overall profit of the executed jobs is maximized. Each job demands a number of resou...... indicate that the algorithm can be used as an actual scheduling algorithm in the Grid or as a tool for analyzing Grid performance when adding extra machines or jobs. © 2012 Wiley Periodicals, Inc.......We consider an integrated job scheduling and network routing problem which appears in Grid Computing and production planning. The problem is to schedule a number of jobs at a finite set of machines, such that the overall profit of the executed jobs is maximized. Each job demands a number...... of resources which must be sent to the executing machine through a network with limited capacity. A job cannot start before all of its resources have arrived at the machine. The scheduling problem is formulated as a Mixed Integer Program (MIP) and proved to be NP-hard. An exact solution approach using Dantzig...

  4. Pilot factory - a Condor-based system for scalable Pilot Job generation in the Panda WMS framework

    International Nuclear Information System (INIS)

    Chiu, Po-Hsiang; Potekhin, Maxim

    2010-01-01

    The Panda Workload Management System is designed around the concept of the Pilot Job - a 'smart wrapper' for the payload executable that can probe the environment on the remote worker node before pulling down the payload from the server and executing it. Such design allows for improved logging and monitoring capabilities as well as flexibility in Workload Management. In the Grid environment (such as the Open Science Grid), Panda Pilot Jobs are submitted to remote sites via mechanisms that ultimately rely on Condor-G. As our experience has shown, in cases where a large number of Panda jobs are simultaneously routed to a particular remote site, the increased load on the head node of the cluster, which is caused by the Pilot Job submission, may lead to overall lack of scalability. We have developed a Condor-inspired solution to this problem, which is using the schedd-based glidein, whose mission is to redirect pilots to the native batch system. Once a glidein schedd is installed and running, it can be utilized exactly the same way as local schedds and therefore, from the user's perspective, Pilots thus submitted are quite similar to jobs submitted to the local Condor pool.

  5. Application of remote debugging techniques in user-centric job monitoring

    International Nuclear Information System (INIS)

    Dos Santos, T; Mättig, P; Harenberg, T; Volkmer, F; Beermann, T; Kalinin, S; Ahrens, R; Wulff, N

    2012-01-01

    With the Job Execution Monitor, a user-centric job monitoring software developed at the University of Wuppertal and integrated into the job brokerage systems of the WLCG, job progress and grid worker node health can be supervised in real time. Imminent error conditions can thus be detected early by the submitter and countermeasures can be taken. Grid site admins can access aggregated data of all monitored jobs to infer the site status and to detect job misbehaviour. To remove the last 'blind spot' from this monitoring, a remote debugging technique based on the GNU C compiler suite was developed and integrated into the software; its design concept and architecture is described in this paper and its application discussed.

  6. RGLite, an interface between ROOT and gLite—proof on the grid

    Science.gov (United States)

    Malzacher, P.; Manafov, A.; Schwarz, K.

    2008-07-01

    Using the gLitePROOF package it is possible to perform PROOF-based distributed data analysis on the gLite Grid. The LHC experiments managed to run globally distributed Monte Carlo productions on the Grid, now the development of tools for data analysis is in the foreground. To grant access interfaces must be provided. The ROOT/PROOF framework is used as a starting point. Using abstract ROOT classes (TGrid, ...) interfaces can be implemented, via which Grid access from ROOT can be accomplished. A concrete implementation exists for the ALICE Grid environment AliEn. Within the D-Grid project an interface to the common Grid middleware of all LHC experiments, gLite, has been created. Therefore it is possible to query Grid File Catalogues from ROOT for the location of the data to be analysed. Grid jobs can be submitted into a gLite based Grid. The status of the jobs can be asked for, and their results can be obtained.

  7. RGLite, an interface between ROOT and gLite-proof on the grid

    International Nuclear Information System (INIS)

    Malzacher, P; Manafov, A; Schwarz, K

    2008-01-01

    Using the gLitePROOF package it is possible to perform PROOF-based distributed data analysis on the gLite Grid. The LHC experiments managed to run globally distributed Monte Carlo productions on the Grid, now the development of tools for data analysis is in the foreground. To grant access interfaces must be provided. The ROOT/PROOF framework is used as a starting point. Using abstract ROOT classes (TGrid, ...) interfaces can be implemented, via which Grid access from ROOT can be accomplished. A concrete implementation exists for the ALICE Grid environment AliEn. Within the D-Grid project an interface to the common Grid middleware of all LHC experiments, gLite, has been created. Therefore it is possible to query Grid File Catalogues from ROOT for the location of the data to be analysed. Grid jobs can be submitted into a gLite based Grid. The status of the jobs can be asked for, and their results can be obtained

  8. Employing peer-to-peer software distribution in ALICE Grid Services to enable opportunistic use of OSG resources

    CERN Multimedia

    CERN. Geneva; Sakrejda, Iwona

    2012-01-01

    The ALICE Grid infrastructure is based on AliEn, a lightweight open source framework built on Web Services and a Distributed Agent Model in which job agents are submitted onto a grid site to prepare the environment and pull work from a central task queue located at CERN. In the standard configuration, each ALICE grid site supports an ALICE-specific VO box as a single point of contact between the site and the ALICE central services. VO box processes monitor site utilization and job requests (ClusterMonitor), monitor dynamic job and site properties (MonaLisa), perform job agent submission (CE) and deploy job-specific software (PackMan). In particular, requiring a VO box at each site simplifies deployment of job software, done onto a shared file system at the site, and adds redundancy to the overall Grid system. ALICE offline computing, however, has also implemented a peer-to-peer method (based on BitTorrent) for downloading job software directly onto each worker node as needed. By utilizing both this peer-...

  9. Experience with Remote Job Execution

    International Nuclear Information System (INIS)

    Lynch, Vickie E.; Cobb, John W; Green, Mark L.; Kohl, James Arthur; Miller, Stephen D.; Ren, Shelly; Smith, Bradford C.; Vazhkudai, Sudharshan S.

    2008-01-01

    The Neutron Science Portal at Oak Ridge National Laboratory submits jobs to the TeraGrid for remote job execution. The TeraGrid is a network of high performance computers supported by the US National Science Foundation. There are eleven partner facilities with over a petaflop of peak computing performance and sixty petabytes of long-term storage. Globus is installed on a local machine and used for job submission. The graphical user interface is produced by java coding that reads an XML file. After submission, the status of the job is displayed in a Job Information Service window which queries globus for the status. The output folder produced in the scratch directory of the TeraGrid machine is returned to the portal with globus-url-copy command that uses the gridftp servers on the TeraGrid machines. This folder is copied from the stage-in directory of the community account to the user's results directory where the output can be plotted using the portal's visualization services. The primary problem with remote job execution is diagnosing execution problems. We have daily tests of submitting multiple remote jobs from the portal. When these jobs fail on a computer, it is difficult to diagnose the problem from the globus output. Successes and problems will be presented

  10. Job prioritization in LHCb

    CERN Document Server

    Castellani, G

    2007-01-01

    LHCb is one of the four high-energy experiments running in the near future at the Large Hadron Collider (LHC) at CERN. LHCb will try to answer some fundamental questions about the asymmetry between matter and anti-matter. The experiment is expected to produce about 2PB of data per year. Those will be distributed to several laboratories all over Europe and then analyzed by the Physics community. To achieve this target LHCb fully uses the Grid to reprocess, replicate and analyze data. The access to the Grid happens through LHCb's own distributed production and analysis system, DIRAC (Distributed Infrastructure with Remote Agent Control). Dirac implements the ‘pull’ job scheduling paradigm, where all the jobs are stored in a central task queues and then pulled via generic grid jobs called Pilot Agents. The whole LHCb community (about 600 people) is divided in sets of physicists, developers, production and software managers that have different needs about their jobs on the Grid. While a Monte Carlo simulation...

  11. The Particle Physics Data Grid. Final Report

    International Nuclear Information System (INIS)

    Livny, Miron

    2002-01-01

    The main objective of the Particle Physics Data Grid (PPDG) project has been to implement and evaluate distributed (Grid-enabled) data access and management technology for current and future particle and nuclear physics experiments. The specific goals of PPDG have been to design, implement, and deploy a Grid-based software infrastructure capable of supporting the data generation, processing and analysis needs common to the physics experiments represented by the participants, and to adapt experiment-specific software to operate in the Grid environment and to exploit this infrastructure. To accomplish these goals, the PPDG focused on the implementation and deployment of several critical services: reliable and efficient file replication service, high-speed data transfer services, multisite file caching and staging service, and reliable and recoverable job management services. The focus of the activity was the job management services and the interplay between these services and distributed data access in a Grid environment. Software was developed to study the interaction between HENP applications and distributed data storage fabric. One key conclusion was the need for a reliable and recoverable tool for managing large collections of interdependent jobs. An attached document provides an overview of the current status of the Directed Acyclic Graph Manager (DAGMan) with its main features and capabilities

  12. Grid3: An Application Grid Laboratory for Science

    CERN Multimedia

    CERN. Geneva

    2004-01-01

    level services required by the participating experiments. The deployed infrastructure has been operating since November 2003 with 27 sites, a peak of 2800 processors, work loads from 10 different applications exceeding 1300 simultaneous jobs, and data transfers among sites of greater than 2 TB/day. The Grid3 infrastructure was deployed from grid level services provided by groups and applications within the collaboration. The services were organized into four distinct "grid level services" including: Grid3 Packaging, Monitoring and Information systems, User Authentication and the iGOC Grid Operatio...

  13. Grid-based platform for training in Earth Observation

    Science.gov (United States)

    Petcu, Dana; Zaharie, Daniela; Panica, Silviu; Frincu, Marc; Neagul, Marian; Gorgan, Dorian; Stefanut, Teodor

    2010-05-01

    found in [4]. The Workload Management System (WMS) provides two types of resource managers. The first one will be based on Condor HTC and use Condor as a job manager for task dispatching and working nodes (for development purposes) while the second one will use GT4 GRAM (for production purposes). The WMS main component, the Grid Task Dispatcher (GTD), is responsible for the interaction with other internal services as the composition engine in order to facilitate access to the processing platform. Its main responsibilities are to receive tasks from the workflow engine or directly from user interface, to use a task description language (the ClassAd meta language in case of Condor HTC) for job units, to submit and check the status of jobs inside the workload management system and to retrieve job logs for debugging purposes. More details can be found in [4]. A particular component of the platform is eGLE, the eLearning environment. It provides the functionalities necessary to create the visual appearance of the lessons through the usage of visual containers like tools, patterns and templates. The teacher uses the platform for testing the already created lessons, as well as for developing new lesson resources, such as new images and workflows describing graph-based processing. The students execute the lessons or describe and experiment with new workflows or different data. The eGLE database includes several workflow-based lesson descriptions, teaching materials and lesson resources, selected satellite and spatial data. More details can be found in [5]. A first training event of using the platform was organized in September 2009 during 11th SYNASC symposium (links to the demos, testing interface, and exercises are available on project site [1]). The eGLE component was presented at 4th GPC conference in May 2009. Moreover, the functionality of the platform will be presented as demo in April 2010 at 5th EGEE User forum. References: [1] GiSHEO consortium, Project site, http

  14. FermiGrid - experience and future plans

    International Nuclear Information System (INIS)

    Chadwick, K.; Berman, E.; Canal, P.; Hesselroth, T.; Garzoglio, G.; Levshina, T.; Sergeev, V.; Sfiligoi, I.; Timm, S.; Yocum, D.

    2007-01-01

    Fermilab supports a scientific program that includes experiments and scientists located across the globe. In order to better serve this community, Fermilab has placed its production computer resources in a Campus Grid infrastructure called 'FermiGrid'. The FermiGrid infrastructure allows the large experiments at Fermilab to have priority access to their own resources, enables sharing of these resources in an opportunistic fashion, and movement of work (jobs, data) between the Campus Grid and National Grids such as Open Science Grid and the WLCG. FermiGrid resources support multiple Virtual Organizations (VOs), including VOs from the Open Science Grid (OSG), EGEE and the Worldwide LHC Computing Grid Collaboration (WLCG). Fermilab also makes leading contributions to the Open Science Grid in the areas of accounting, batch computing, grid security, job management, resource selection, site infrastructure, storage management, and VO services. Through the FermiGrid interfaces, authenticated and authorized VOs and individuals may access our core grid services, the 10,000+ Fermilab resident CPUs, near-petabyte (including CMS) online disk pools and the multi-petabyte Fermilab Mass Storage System. These core grid services include a site wide Globus gatekeeper, VO management services for several VOs, Fermilab site authorization services, grid user mapping services, as well as job accounting and monitoring, resource selection and data movement services. Access to these services is via standard and well-supported grid interfaces. We will report on the user experience of using the FermiGrid campus infrastructure interfaced to a national cyberinfrastructure--the successes and the problems

  15. Final Progress Report for 'An Abstract Job Handling Grid Service for Dataset Analysis'

    International Nuclear Information System (INIS)

    David A Alexander

    2005-01-01

    For Phase I of the Job Handling project, Tech-X has built a Grid service for processing analysis requests, as well as a Graphical User Interface (GUI) client that uses the service. The service is designed to generically support High-Energy Physics (HEP) experimental analysis tasks. It has an extensible, flexible, open architecture and language. The service uses the Solenoidal Tracker At RHIC (STAR) experiment as a working example. STAR is an experiment at the Relativistic Heavy Ion Collider (RHIC) at the Brookhaven National Laboratory (BNL). STAR and other experiments at BNL generate multiple Petabytes of HEP data. The raw data is captured as millions of input files stored in a distributed data catalog. Potentially using thousands of files as input, analysis requests are submitted to a processing environment containing thousands of nodes. The Grid service provides a standard interface to the processing farm. It enables researchers to run large-scale, massively parallel analysis tasks, regardless of the computational resources available in their location

  16. FermiGrid-experience and future plans

    International Nuclear Information System (INIS)

    Chadwick, K; Berman, E; Canal, P; Hesselroth, T; Garzoglio, G; Levshina, T; Sergeev, V; Sfiligoi, I; Sharma, N; Timm, S; Yocum, D R

    2008-01-01

    Fermilab supports a scientific program that includes experiments and scientists located across the globe. In order to better serve this community, Fermilab has placed its production computer resources in a Campus Grid infrastructure called 'FermiGrid'. The FermiGrid infrastructure allows the large experiments at Fermilab to have priority access to their own resources, enables sharing of these resources in an opportunistic fashion, and movement of work (jobs, data) between the Campus Grid and National Grids such as Open Science Grid (OSG) and the Worldwide LHC Computing Grid Collaboration (WLCG). FermiGrid resources support multiple Virtual Organizations (VOs), including VOs from the OSG, EGEE, and the WLCG. Fermilab also makes leading contributions to the Open Science Grid in the areas of accounting, batch computing, grid security, job management, resource selection, site infrastructure, storage management, and VO services. Through the FermiGrid interfaces, authenticated and authorized VOs and individuals may access our core grid services, the 10,000+ Fermilab resident CPUs, near-petabyte (including CMS) online disk pools and the multi-petabyte Fermilab Mass Storage System. These core grid services include a site wide Globus gatekeeper, VO management services for several VOs, Fermilab site authorization services, grid user mapping services, as well as job accounting and monitoring, resource selection and data movement services. Access to these services is via standard and well-supported grid interfaces. We will report on the user experience of using the FermiGrid campus infrastructure interfaced to a national cyberinfrastructure - the successes and the problems

  17. Enabling Campus Grids with Open Science Grid Technology

    International Nuclear Information System (INIS)

    Weitzel, Derek; Fraser, Dan; Pordes, Ruth; Bockelman, Brian; Swanson, David

    2011-01-01

    The Open Science Grid is a recognized key component of the US national cyber-infrastructure enabling scientific discovery through advanced high throughput computing. The principles and techniques that underlie the Open Science Grid can also be applied to Campus Grids since many of the requirements are the same, even if the implementation technologies differ. We find five requirements for a campus grid: trust relationships, job submission, resource independence, accounting, and data management. The Holland Computing Center's campus grid at the University of Nebraska-Lincoln was designed to fulfill the requirements of a campus grid. A bridging daemon was designed to bring non-Condor clusters into a grid managed by Condor. Condor features which make it possible to bridge Condor sites into a multi-campus grid have been exploited at the Holland Computing Center as well.

  18. A Job Monitoring and Accounting Tool for the LSF Batch System

    International Nuclear Information System (INIS)

    Sarkar, Subir; Taneja, Sonia

    2011-01-01

    This paper presents a web based job monitoring and group-and-user accounting tool for the LSF Batch System. The user oriented job monitoring displays a simple and compact quasi real-time overview of the batch farm for both local and Grid jobs. For Grid jobs the Distinguished Name (DN) of the Grid users is shown. The overview monitor provides the most up-to-date status of a batch farm at any time. The accounting tool works with the LSF accounting log files. The accounting information is shown for a few pre-defined time periods by default. However, one can also compute the same information for any arbitrary time window. The tool already proved to be an extremely useful means to validate more extensive accounting tools available in the Grid world. Several sites have already been using the present tool and more sites running the LSF batch system have shown interest. We shall discuss the various aspects that make the tool essential for site administrators and end-users alike and outline the current status of development as well as future plans.

  19. Analyzing Grid Log Data with Affinity Propagation

    NARCIS (Netherlands)

    Modena, G.; van Someren, M.W.; Ali, M; Bosse, T.; Hindriks, K.V.; Hoogendoorn, M.; Jonker, C.M; Treur, J.

    2013-01-01

    In this paper we present an unsupervised learning approach to detect meaningful job traffic patterns in Grid log data. Manual anomaly detection on modern Grid environments is troublesome given their increasing complexity, the distributed, dynamic topology of the network and heterogeneity of the jobs

  20. Integrating GRID tools to build a computing resource broker: activities of DataGrid WP1

    International Nuclear Information System (INIS)

    Anglano, C.; Barale, S.; Gaido, L.; Guarise, A.; Lusso, S.; Werbrouck, A.

    2001-01-01

    Resources on a computational Grid are geographically distributed, heterogeneous in nature, owned by different individuals or organizations with their own scheduling policies, have different access cost models with dynamically varying loads and availability conditions. This makes traditional approaches to workload management, load balancing and scheduling inappropriate. The first work package (WP1) of the EU-funded DataGrid project is addressing the issue of optimizing the distribution of jobs onto Grid resources based on a knowledge of the status and characteristics of these resources that is necessarily out-of-date (collected in a finite amount of time at a very loosely coupled site). The authors describe the DataGrid approach in integrating existing software components (from Condor, Globus, etc.) to build a Grid Resource Broker, and the early efforts to define a workable scheduling strategy

  1. User centric monitoring (UCM) information service for the next generation of Grid-enabled scientists

    International Nuclear Information System (INIS)

    Alexander, D A; Li, C; Lauret, J; Fine, V

    2008-01-01

    Nuclear and high-energy physicists routinely execute data processing and data analysis jobs on a Grid and need to be able to easily and remotely monitor the execution of these jobs. Existing Grid monitoring tools provide abundant information about the whole system, but are geared towards production jobs and well suited for Grid administrators, while the information tailored towards an individual user is not readily available in a user-friendly and user-centric way. Such User Centric information includes monitoring information such as the status of the submitted job, queue position, time of the start/finish, percentage of being done, error messages, standard output, and reasons for failure. We proposed to develop a framework based on Grid service technology that allows scientists to track and monitor their jobs easily from a user-centric view. The proposed framework aims to be flexible so that it can be applied by any Grid Virtual Organization (VO) with various ways of collecting the user-centric job monitoring information built into the framework. Furthermore, the framework provides a rich and reusable set of methods of presenting the information to the user from within a Web browser and other clients. In this presentation, we will give an architectural overview of the UCM service, show an example implementation in the RHIC/STAR experiment context and discuss limitations and future collaborative work

  2. User centric monitoring (UCM) information service for the next generation of Grid-enabled scientists

    Energy Technology Data Exchange (ETDEWEB)

    Alexander, D A; Li, C [Tech-X Corporation, 5621 Arapahoe Avenue Suite A, Boulder, CO 80303 (United States); Lauret, J; Fine, V [Physics Department, Brookhaven National Laboratory, Upton, NY 11973 (United States)], E-mail: alexanda@txcorp.com

    2008-07-15

    Nuclear and high-energy physicists routinely execute data processing and data analysis jobs on a Grid and need to be able to easily and remotely monitor the execution of these jobs. Existing Grid monitoring tools provide abundant information about the whole system, but are geared towards production jobs and well suited for Grid administrators, while the information tailored towards an individual user is not readily available in a user-friendly and user-centric way. Such User Centric information includes monitoring information such as the status of the submitted job, queue position, time of the start/finish, percentage of being done, error messages, standard output, and reasons for failure. We proposed to develop a framework based on Grid service technology that allows scientists to track and monitor their jobs easily from a user-centric view. The proposed framework aims to be flexible so that it can be applied by any Grid Virtual Organization (VO) with various ways of collecting the user-centric job monitoring information built into the framework. Furthermore, the framework provides a rich and reusable set of methods of presenting the information to the user from within a Web browser and other clients. In this presentation, we will give an architectural overview of the UCM service, show an example implementation in the RHIC/STAR experiment context and discuss limitations and future collaborative work.

  3. Enabling campus grids with open science grid technology

    Energy Technology Data Exchange (ETDEWEB)

    Weitzel, Derek [Nebraska U.; Bockelman, Brian [Nebraska U.; Swanson, David [Nebraska U.; Fraser, Dan [Argonne; Pordes, Ruth [Fermilab

    2011-01-01

    The Open Science Grid is a recognized key component of the US national cyber-infrastructure enabling scientific discovery through advanced high throughput computing. The principles and techniques that underlie the Open Science Grid can also be applied to Campus Grids since many of the requirements are the same, even if the implementation technologies differ. We find five requirements for a campus grid: trust relationships, job submission, resource independence, accounting, and data management. The Holland Computing Center's campus grid at the University of Nebraska-Lincoln was designed to fulfill the requirements of a campus grid. A bridging daemon was designed to bring non-Condor clusters into a grid managed by Condor. Condor features which make it possible to bridge Condor sites into a multi-campus grid have been exploited at the Holland Computing Center as well.

  4. CDF GlideinWMS usage in Grid computing of high energy physics

    International Nuclear Information System (INIS)

    Zvada, Marian; Sfiligoi, Igor; Benjamin, Doug

    2010-01-01

    Many members of large science collaborations already have specialized grids available to advance their research in the need of getting more computing resources for data analysis. This has forced the Collider Detector at Fermilab (CDF) collaboration to move beyond the usage of dedicated resources and start exploiting Grid resources. Nowadays, CDF experiment is increasingly relying on glidein-based computing pools for data reconstruction. Especially, Monte Carlo production and user data analysis, serving over 400 users by central analysis farm middleware (CAF) on the top of Condor batch system and CDF Grid infrastructure. Condor is designed as distributed architecture and its glidein mechanism of pilot jobs is ideal for abstracting the Grid computing by making a virtual private computing pool. We would like to present the first production use of the generic pilot-based Workload Management System (glideinWMS), which is an implementation of the pilot mechanism based on the Condor distributed infrastructure. CDF Grid computing uses glideinWMS for its data reconstruction on the FNAL campus Grid, user analysis and Monte Carlo production across Open Science Grid (OSG). We review this computing model and setup used including CDF specific configuration within the glideinWMS system which provides powerful scalability and makes Grid computing working like in a local batch environment with ability to handle more than 10000 running jobs at a time.

  5. Monitoring system for the GRID Monte Carlo mass production in the H1 experiment at DESY

    International Nuclear Information System (INIS)

    Bystritskaya, Elena; Fomenko, Alexander; Gogitidze, Nelly; Lobodzinski, Bogdan

    2014-01-01

    The H1 Virtual Organization (VO), as one of the small VOs, employs most components of the EMI or gLite Middleware. In this framework, a monitoring system is designed for the H1 Experiment to identify and recognize within the GRID the best suitable resources for execution of CPU-time consuming Monte Carlo (MC) simulation tasks (jobs). Monitored resources are Computer Elements (CEs), Storage Elements (SEs), WMS-servers (WMSs), CernVM File System (CVMFS) available to the VO HONE and local GRID User Interfaces (UIs). The general principle of monitoring GRID elements is based on the execution of short test jobs on different CE queues using submission through various WMSs and directly to the CREAM-CEs as well. Real H1 MC Production jobs with a small number of events are used to perform the tests. Test jobs are periodically submitted into GRID queues, the status of these jobs is checked, output files of completed jobs are retrieved, the result of each job is analyzed and the waiting time and run time are derived. Using this information, the status of the GRID elements is estimated and the most suitable ones are included in the automatically generated configuration files for use in the H1 MC production. The monitoring system allows for identification of problems in the GRID sites and promptly reacts on it (for example by sending GGUS (Global Grid User Support) trouble tickets). The system can easily be adapted to identify the optimal resources for tasks other than MC production, simply by changing to the relevant test jobs. The monitoring system is written mostly in Python and Perl with insertion of a few shell scripts. In addition to the test monitoring system we use information from real production jobs to monitor the availability and quality of the GRID resources. The monitoring tools register the number of job resubmissions, the percentage of failed and finished jobs relative to all jobs on the CEs and determine the average values of waiting and running time for the

  6. GENIUS: a web portal for the grid

    International Nuclear Information System (INIS)

    Andronico, A.; Barbera, R.; Falzone, A.; Lo Re, G.; Pulvirenti, A.; Rodolico, A.

    2003-01-01

    The architecture and the current implementation of the grid portal GENIUS (Grid Enabled web environment for site Independent User job Submission), jointly developed by INFN and NICE within the context of the INFN Grid and DataGrid Project, is presented and discussed

  7. A performance study of grid workflow engines

    NARCIS (Netherlands)

    Stratan, C.; Iosup, A.; Epema, D.H.J.

    2008-01-01

    To benefit from grids, scientists require grid workflow engines that automatically manage the execution of inter-related jobs on the grid infrastructure. So far, the workflows community has focused on scheduling algorithms and on interface tools. Thus, while several grid workflow engines have been

  8. The Impact of the Implementation Cost of Replication in Data Grid Job Scheduling

    Directory of Open Access Journals (Sweden)

    Babar Nazir

    2018-05-01

    Full Text Available Data Grids deal with geographically-distributed large-scale data-intensive applications. Schemes scheduled for data grids attempt to not only improve data access time, but also aim to improve the ratio of data availability to a node, where the data requests are generated. Data replication techniques manage large data by storing a number of data files efficiently. In this paper, we propose centralized dynamic scheduling strategy-replica placement strategies (CDSS-RPS. CDSS-RPS schedule the data and task so that it minimizes the implementation cost and data transfer time. CDSS-RPS consists of two algorithms, namely (a centralized dynamic scheduling (CDS and (b replica placement strategy (RPS. CDS considers the computing capacity of a node and finds an appropriate location for the job. RPS attempts to improve file access time by using replication on the basis of number of accesses, storage capacity of a computing node, and response time of a requested file. Extensive simulations are carried out to demonstrate the effectiveness of the proposed strategy. Simulation results demonstrate that the replication and scheduling strategies improve the implementation cost and average access time significantly.

  9. CDF experience with monte carlo production using LCG grid

    International Nuclear Information System (INIS)

    Griso, S P; Lucchesi, D; Compostella, G; Sfiligoi, I; Cesini, D

    2008-01-01

    The upgrades of the Tevatron collider and CDF detector have considerably increased the demand on computing resources, in particular for Monte Carlo production. This has forced the collaboration to move beyond the usage of dedicated resources and start exploiting the Grid. The CDF Analysis Farm (CAF) model has been reimplemented into LcgCAF in order to access Grid resources by using the LCG/EGEE middleware. Many sites in Italy and in Europe are accessed through this portal by CDF users mainly to produce Monte Carlo data but also for other analysis jobs. We review here the setup used to submit jobs to Grid sites and retrieve the output, including CDF-specific configuration of some Grid components. We also describe the batch and interactive monitor tools developed to allow users to verify the jobs status during their lifetime in the Grid environment. Finally we analyze the efficiency and typical failure modes of the current Grid infrastructure reporting the performances of different parts of the system used

  10. A principled approach to grid middleware

    DEFF Research Database (Denmark)

    Berthold, Jost; Bardino, Jonas; Vinter, Brian

    2011-01-01

    This paper provides an overview of MiG, a Grid middleware for advanced job execution, data storage and group collaboration in an integrated, yet lightweight solution using standard software. In contrast to most other Grid middlewares, MiG is developed with a particular focus on usability and mini......This paper provides an overview of MiG, a Grid middleware for advanced job execution, data storage and group collaboration in an integrated, yet lightweight solution using standard software. In contrast to most other Grid middlewares, MiG is developed with a particular focus on usability...... and minimal system requirements, applying strict principles to keep the middleware free of legacy burdens and overly complicated design. We provide an overview of MiG and describe its features in view of the Grid vision and its relation to more recent cloud computing trends....

  11. Grid Transmission Expansion Planning Model Based on Grid Vulnerability

    Science.gov (United States)

    Tang, Quan; Wang, Xi; Li, Ting; Zhang, Quanming; Zhang, Hongli; Li, Huaqiang

    2018-03-01

    Based on grid vulnerability and uniformity theory, proposed global network structure and state vulnerability factor model used to measure different grid models. established a multi-objective power grid planning model which considering the global power network vulnerability, economy and grid security constraint. Using improved chaos crossover and mutation genetic algorithm to optimize the optimal plan. For the problem of multi-objective optimization, dimension is not uniform, the weight is not easy given. Using principal component analysis (PCA) method to comprehensive assessment of the population every generation, make the results more objective and credible assessment. the feasibility and effectiveness of the proposed model are validated by simulation results of Garver-6 bus system and Garver-18 bus.

  12. A Development of Lightweight Grid Interface

    International Nuclear Information System (INIS)

    Iwai, G; Kawai, Y; Sasaki, T; Watase, Y

    2011-01-01

    In order to help a rapid development of Grid/Cloud aware applications, we have developed API to abstract the distributed computing infrastructures based on SAGA (A Simple API for Grid Applications). SAGA, which is standardized in the OGF (Open Grid Forum), defines API specifications to access distributed computing infrastructures, such as Grid, Cloud and local computing resources. The Universal Grid API (UGAPI), which is a set of command line interfaces (CLI) and APIs, aims to offer simpler API to combine several SAGA interfaces with richer functionalities. These CLIs of the UGAPI offer typical functionalities required by end users for job management and file access to the different distributed computing infrastructures as well as local computing resources. We have also built a web interface for the particle therapy simulation and demonstrated the large scale calculation using the different infrastructures at the same time. In this paper, we would like to present how the web interface based on UGAPI and SAGA achieve more efficient utilization of computing resources over the different infrastructures with technical details and practical experiences.

  13. Running CMS remote analysis builder jobs on advanced resource connector middleware

    International Nuclear Information System (INIS)

    Edelmann, E; Happonen, K; Koivumäki, J; Lindén, T; Välimaa, J

    2011-01-01

    CMS user analysis jobs are distributed over the grid with the CMS Remote Analysis Builder application (CRAB). According to the CMS computing model the applications should run transparently on the different grid flavours in use. In CRAB this is handled with different plugins that are able to submit to different grids. Recently a CRAB plugin for submitting to the Advanced Resource Connector (ARC) middleware has been developed. The CRAB ARC plugin enables simple and fast job submission with full job status information available. CRAB can be used with a server which manages and monitors the grid jobs on behalf of the user. In the presentation we will report on the CRAB ARC plugin and on the status of integrating it with the CRAB server and compare this with using the gLite ARC interoperability method for job submission.

  14. The pilot way to Grid resources using glideinWMS

    CERN Document Server

    Sfiligoi, Igor; Holzman, Burt; Mhashilkar, Parag; Padhi, Sanjay; Wurthwrin, Frank

    Grid computing has become very popular in big and widespread scientific communities with high computing demands, like high energy physics. Computing resources are being distributed over many independent sites with only a thin layer of grid middleware shared between them. This deployment model has proven to be very convenient for computing resource providers, but has introduced several problems for the users of the system, the three major being the complexity of job scheduling, the non-uniformity of compute resources, and the lack of good job monitoring. Pilot jobs address all the above problems by creating a virtual private computing pool on top of grid resources. This paper presents both the general pilot concept, as well as a concrete implementation, called glideinWMS, deployed in the Open Science Grid.

  15. H1 Grid production tool for large scale Monte Carlo simulation

    Energy Technology Data Exchange (ETDEWEB)

    Lobodzinski, B; Wissing, Ch [DESY, Hamburg (Germany); Bystritskaya, E; Vorobiew, M [ITEP, Moscow (Russian Federation); Karbach, T M [University of Dortmund (Germany); Mitsyn, S [JINR, Moscow (Russian Federation); Mudrinic, M, E-mail: bogdan.lobodzinski@desy.d [VINS, Belgrad (Serbia)

    2010-04-01

    The H1 Collaboration at HERA has entered the period of high precision analyses based on the final data sample. These analyses require a massive production of simulated Monte Carlo (MC) events. The H1 MC framework (H1MC) is a software for mass MC production on the LCG Grid infrastructure and on a local batch system created by H1 Collaboration. The aim of the tool is a full automatisation of the MC production workflow including management of the MC jobs on the Grid down to copying of the resulting files from the Grid to the H1 mass storage tape device. The H1 MC framework has modular structure, delegating a specific task to each module, including task specific to the H1 experiment: Automatic building of steer and input files, simulation of the H1 detector, reconstruction of particle tracks and post processing calculation. Each module provides data or functionality needed by other modules via a local database. The Grid jobs created for detector simulation and reconstruction from generated MC input files are fully independent and fault-tolerant for 32 and 64-bit LCG Grid architecture and in Grid running state they can be continuously monitored using Relational Grid Monitoring Architecture (R-GMA) service. To monitor the full production chain and detect potential problems, regular checks of the job state are performed using the local database and the Service Availability Monitoring (SAM) framework. The improved stability of the system has resulted in a dramatic increase in the production rate, which exceeded two billion MC events in 2008.

  16. The CrossGrid project

    International Nuclear Information System (INIS)

    Kunze, M.

    2003-01-01

    There are many large-scale problems that require new approaches to computing, such as earth observation, environmental management, biomedicine, industrial and scientific modeling. The CrossGrid project addresses realistic problems in medicine, environmental protection, flood prediction, and physics analysis and is oriented towards specific end-users: Medical doctors, who could obtain new tools to help them to obtain correct diagnoses and to guide them during operations; industries, that could be advised on the best timing for some critical operations involving risk of pollution; flood crisis teams, that could predict the risk of a flood on the basis of historical records and actual hydrological and meteorological data; physicists, who could optimize the analysis of massive volumes of data distributed across countries and continents. Corresponding applications will be based on Grid technology and could be complex and difficult to use: the CrossGrid project aims at developing several tools that will make the Grid more friendly for average users. Portals for specific applications will be designed, that should allow for easy connection to the Grid, create a customized work environment, and provide users with all necessary information to get their job done

  17. MrGrid: a portable grid based molecular replacement pipeline.

    Directory of Open Access Journals (Sweden)

    Jason W Schmidberger

    Full Text Available BACKGROUND: The crystallographic determination of protein structures can be computationally demanding and for difficult cases can benefit from user-friendly interfaces to high-performance computing resources. Molecular replacement (MR is a popular protein crystallographic technique that exploits the structural similarity between proteins that share some sequence similarity. But the need to trial permutations of search models, space group symmetries and other parameters makes MR time- and labour-intensive. However, MR calculations are embarrassingly parallel and thus ideally suited to distributed computing. In order to address this problem we have developed MrGrid, web-based software that allows multiple MR calculations to be executed across a grid of networked computers, allowing high-throughput MR. METHODOLOGY/PRINCIPAL FINDINGS: MrGrid is a portable web based application written in Java/JSP and Ruby, and taking advantage of Apple Xgrid technology. Designed to interface with a user defined Xgrid resource the package manages the distribution of multiple MR runs to the available nodes on the Xgrid. We evaluated MrGrid using 10 different protein test cases on a network of 13 computers, and achieved an average speed up factor of 5.69. CONCLUSIONS: MrGrid enables the user to retrieve and manage the results of tens to hundreds of MR calculations quickly and via a single web interface, as well as broadening the range of strategies that can be attempted. This high-throughput approach allows parameter sweeps to be performed in parallel, improving the chances of MR success.

  18. Job submission and management through web services the experience with the CREAM service

    CERN Document Server

    Aiftimiei, C; Bertocco, S; Fina, S D; Ronco, S D; Dorigo, A; Gianelle, A; Marzolla, M; Mazzucato, M; Sgaravatto, M; Verlato, M; Zangrando, L; Corvo, M; Miccio, V; Sciabà, A; Cesini, D; Dongiovanni, D; Grandi, C

    2008-01-01

    Modern Grid middleware is built around components providing basic functionality, such as data storage, authentication, security, job management, resource monitoring and reservation. In this paper we describe the Computing Resource Execution and Management (CREAM) service. CREAM provides a Web service-based job execution and management capability for Grid systems; in particular, it is being used within the gLite middleware. CREAM exposes a Web service interface allowing conforming clients to submit and manage computational jobs to a Local Resource Management System. We developed a special component, called ICE (Interface to CREAM Environment) to integrate CREAM in gLite. ICE transfers job submissions and cancellations from the Workload Management System, allowing users to manage CREAM jobs from the gLite User Interface. This paper describes some recent studies aimed at assessing the performance and reliability of CREAM and ICE; those tests have been performed as part of the acceptance tests for integration of ...

  19. Multi-core job submission and grid resource scheduling for ATLAS AthenaMP

    CERN Document Server

    Crooks, D; The ATLAS collaboration; Harrington, R; Purdie, S; Severini, H; Skipsey, S; Tsulaia, V; Washbrook, A

    2012-01-01

    AthenaMP is the multi-core implementation of the ATLAS software framework and allows the efficient sharing of memory pages between multiple threads of execution. This has now been validated for production and delivers a significant reduction on overall memory footprint with negligible CPU overhead. Before AthenaMP can be routinely run on the LHC Computing Grid, it must be determined how the computing resources available to ATLAS can best exploit the notable improvements delivered by switching to this multi-process model. In particular, there is a need to identify and assess the potential impact of scheduling issues where single core and multi-core job queues have access to the same underlying resources. A study into the effectiveness and scalability of AthenaMP in a production environment will be presented. Submitting AthenaMP tasks to the Tier-0 and candidate Tier-2 sites will allow detailed measurement of worker node performance and also highlight the relative performance of local resource management system...

  20. Reducing Deadline Miss Rate for Grid Workloads running in Virtual Machines: a deadline-aware and adaptive approach

    CERN Document Server

    Khalid, Omer; Anthony, Richard; Petridis, Miltos

    2011-01-01

    This thesis explores three major areas of research; integration of virutalization into sci- entific grid infrastructures, evaluation of the virtualization overhead on HPC grid job’s performance, and optimization of job execution times to increase their throughput by reducing job deadline miss rate. Integration of the virtualization into the grid to deploy on-demand virtual machines for jobs in a way that is transparent to the end users and have minimum impact on the existing system poses a significant challenge. This involves the creation of virtual machines, decompression of the operating system image, adapting the virtual environ- ment to satisfy software requirements of the job, constant update of the job state once it’s running with out modifying batch system or existing grid middleware, and finally bringing the host machine back to a consistent state. To facilitate this research, an existing and in production pilot job framework has been modified to deploy virtual machines on demand on the grid using...

  1. Integrating job scheduling and constrained network routing

    DEFF Research Database (Denmark)

    Gamst, Mette

    2010-01-01

    This paper examines the NP-hard problem of scheduling jobs on resources such that the overall profit of executed jobs is maximized. Job demand must be sent through a constrained network to the resource before execution can begin. The problem has application in grid computing, where a number...

  2. Ecosystem Based Business Model of Smart Grid

    OpenAIRE

    Lundgaard, Morten Raahauge; Ma, Zheng; Jørgensen, Bo Nørregaard

    2015-01-01

    This paper tries to investigate the ecosystem based business model in a smart grid infrastructure and the potential of value capture in the highly complex macro infrastructure such as smart grid. This paper proposes an alternative perspective to study the smart grid business ecosystem to support the infrastructural challenges, such as the interoperability of business components for smart grid. So far little research has explored the business ecosystem in the smart grid concept. The study on t...

  3. ATLAS Grid Workflow Performance Optimization

    CERN Document Server

    Elmsheuser, Johannes; The ATLAS collaboration

    2018-01-01

    The CERN ATLAS experiment grid workflow system manages routinely 250 to 500 thousand concurrently running production and analysis jobs to process simulation and detector data. In total more than 300 PB of data is distributed over more than 150 sites in the WLCG. At this scale small improvements in the software and computing performance and workflows can lead to significant resource usage gains. ATLAS is reviewing together with CERN IT experts several typical simulation and data processing workloads for potential performance improvements in terms of memory and CPU usage, disk and network I/O. All ATLAS production and analysis grid jobs are instrumented to collect many performance metrics for detailed statistical studies using modern data analytics tools like ElasticSearch and Kibana. This presentation will review and explain the performance gains of several ATLAS simulation and data processing workflows and present analytics studies of the ATLAS grid workflows.

  4. The CMS integration grid testbed

    Energy Technology Data Exchange (ETDEWEB)

    Graham, Gregory E.

    2004-08-26

    The CMS Integration Grid Testbed (IGT) comprises USCMS Tier-1 and Tier-2 hardware at the following sites: the California Institute of Technology, Fermi National Accelerator Laboratory, the University of California at San Diego, and the University of Florida at Gainesville. The IGT runs jobs using the Globus Toolkit with a DAGMan and Condor-G front end. The virtual organization (VO) is managed using VO management scripts from the European Data Grid (EDG). Gridwide monitoring is accomplished using local tools such as Ganglia interfaced into the Globus Metadata Directory Service (MDS) and the agent based Mona Lisa. Domain specific software is packaged and installed using the Distribution After Release (DAR) tool of CMS, while middleware under the auspices of the Virtual Data Toolkit (VDT) is distributed using Pacman. During a continuous two month span in Fall of 2002, over 1 million official CMS GEANT based Monte Carlo events were generated and returned to CERN for analysis while being demonstrated at SC2002. In this paper, we describe the process that led to one of the world's first continuously available, functioning grids.

  5. Analysis gets on the starting Grid

    CERN Multimedia

    Roger Jones

    It is vital for ATLAS to have a functioning distributed analysis system to analyse its data. There are three major Grid deployments in ATLAS (Enabling Grids for E-sciencE, EGEE; the US Open Science Grid, OSG; and the Nordic DataGrid Facility, NGDF), and our data and jobs need to work across all of them, as well as on local machines and batch systems. Users must also be able to locate the data they want and register new small datasets so they can be used later. ATLAS has a suite of products to meet these needs, and a series of Distributed Analysis tutorials are training an increasing number of brave early adopters to use the system. Real users are vital to make sure that the tools are fit for their purpose and to refine our computing model. One such tutorial happened on the 1st and 2nd February at the National eScience Centre in Edinburgh, UK, sponsored by the GridPP Collaboration. The first day introduced an international set of tutees to the basic tools for Grid-based distributed analysis. The architecture...

  6. Campus Grids: Bringing Additional Computational Resources to HEP Researchers

    International Nuclear Information System (INIS)

    Weitzel, Derek; Fraser, Dan; Bockelman, Brian; Swanson, David

    2012-01-01

    It is common at research institutions to maintain multiple clusters that represent different owners or generations of hardware, or that fulfill different needs and policies. Many of these clusters are consistently under utilized while researchers on campus could greatly benefit from these unused capabilities. By leveraging principles from the Open Science Grid it is now possible to utilize these resources by forming a lightweight campus grid. The campus grids framework enables jobs that are submitted to one cluster to overflow, when necessary, to other clusters within the campus using whatever authentication mechanisms are available on campus. This framework is currently being used on several campuses to run HEP and other science jobs. Further, the framework has in some cases been expanded beyond the campus boundary by bridging campus grids into a regional grid, and can even be used to integrate resources from a national cyberinfrastructure such as the Open Science Grid. This paper will highlight 18 months of operational experiences creating campus grids in the US, and the different campus configurations that have successfully utilized the campus grid infrastructure.

  7. Climate modelling on the GRID Experiences in the EU-project EELA

    Energy Technology Data Exchange (ETDEWEB)

    Fernandez-Quiruelas, V.; Fernandez, J.; Cofino, A. S.; Gutierrez, J. M.; Baeza Retamal, C.; Abarca del Rio, R.; Miguel San Martin, R.; Carrillo, M.

    2007-07-01

    Recent trends in climate modeling find in GRID computing a powerful way to achieve results by sharing computing and data distributed resources. In particular, ensemble prediction is based on the generation of multiple simulations from perturbed model conditions to sample the existing uncertainties. In this work, we present a GRID application consisting of a sequence of two state-of-the-art climate models (one global model and one regional model), operable through a web portal (based on Genius). The main goal of the application is providing ensemble-based regional predictions. This requires managing a complex work flow involving long-term jobs and job dependencies in a user-transparent way. In doing so, we identified the weaknesses of current middle ware tools and developed a robust work flow by merging the optimal existing applications with an underlying self-developed work flow application based on the communication with metadata catalogs (currently AMGA) storing application status and dynamic model output generation. As an illustrative scientific challenge, the application is applied to study the El Nino phenomenon, by simulating an El Nino year with different forcing conditions and analyzing the precipitation response over south-american countries subject to flooding risk. GRID computing; Climate models; CAM model; WRF model; Work flow. (Author)

  8. Climate modelling on the GRID Experiences in the EU-project EELA

    International Nuclear Information System (INIS)

    Fernandez-Quiruelas, V.; Fernandez, J.; Cofino, A. S.; Gutierrez, J. M.; Baeza Retamal, C.; Abarca del Rio, R.; Miguel San Martin, R.; Carrillo, M.

    2007-01-01

    Recent trends in climate modeling find in GRID computing a powerful way to achieve results by sharing computing and data distributed resources. In particular, ensemble prediction is based on the generation of multiple simulations from perturbed model conditions to sample the existing uncertainties. In this work, we present a GRID application consisting of a sequence of two state-of-the-art climate models (one global model and one regional model), operable through a web portal (based on Genius). The main goal of the application is providing ensemble-based regional predictions. This requires managing a complex work flow involving long-term jobs and job dependencies in a user-transparent way. In doing so, we identified the weaknesses of current middle ware tools and developed a robust work flow by merging the optimal existing applications with an underlying self-developed work flow application based on the communication with metadata catalogs (currently AMGA) storing application status and dynamic model output generation. As an illustrative scientific challenge, the application is applied to study the El Nino phenomenon, by simulating an El Nino year with different forcing conditions and analyzing the precipitation response over south-american countries subject to flooding risk. GRID computing; Climate models; CAM model; WRF model; Work flow. (Author)

  9. DZero data-intensive computing on the Open Science Grid

    International Nuclear Information System (INIS)

    Abbott, B; Baranovski, A; Diesburg, M; Garzoglio, G; Mhashilkar, P; Kurca, T

    2008-01-01

    High energy physics experiments periodically reprocess data, in order to take advantage of improved understanding of the detector and the data processing code. Between February and May 2007, the DZero experiment has reprocessed a substantial fraction of its dataset. This consists of half a billion events, corresponding to about 100 TB of data, organized in 300,000 files. The activity utilized resources from sites around the world, including a dozen sites participating to the Open Science Grid consortium (OSG). About 1,500 jobs were run every day across the OSG, consuming and producing hundreds of Gigabytes of data. Access to OSG computing and storage resources was coordinated by the SAM-Grid system. This system organized job access to a complex topology of data queues and job scheduling to clusters, using a SAM-Grid to OSG job forwarding infrastructure. For the first time in the lifetime of the experiment, a data intensive production activity was managed on a general purpose grid, such as OSG. This paper describes the implications of using OSG, where all resources are granted following an opportunistic model, the challenges of operating a data intensive activity over such large computing infrastructure, and the lessons learned throughout the project

  10. DZero data-intensive computing on the Open Science Grid

    International Nuclear Information System (INIS)

    Abbott, B.; Baranovski, A.; Diesburg, M.; Garzoglio, G.; Kurca, T.; Mhashilkar, P.

    2007-01-01

    High energy physics experiments periodically reprocess data, in order to take advantage of improved understanding of the detector and the data processing code. Between February and May 2007, the DZero experiment has reprocessed a substantial fraction of its dataset. This consists of half a billion events, corresponding to about 100 TB of data, organized in 300,000 files. The activity utilized resources from sites around the world, including a dozen sites participating to the Open Science Grid consortium (OSG). About 1,500 jobs were run every day across the OSG, consuming and producing hundreds of Gigabytes of data. Access to OSG computing and storage resources was coordinated by the SAM-Grid system. This system organized job access to a complex topology of data queues and job scheduling to clusters, using a SAM-Grid to OSG job forwarding infrastructure. For the first time in the lifetime of the experiment, a data intensive production activity was managed on a general purpose grid, such as OSG. This paper describes the implications of using OSG, where all resources are granted following an opportunistic model, the challenges of operating a data intensive activity over such large computing infrastructure, and the lessons learned throughout the project

  11. CMS Monte Carlo production in the WLCG computing grid

    International Nuclear Information System (INIS)

    Hernandez, J M; Kreuzer, P; Hof, C; Khomitch, A; Mohapatra, A; Filippis, N D; Pompili, A; My, S; Abbrescia, M; Maggi, G; Donvito, G; Weirdt, S D; Maes, J; Mulders, P v; Villella, I; Wakefield, S; Guan, W; Fanfani, A; Evans, D; Flossdorf, A

    2008-01-01

    Monte Carlo production in CMS has received a major boost in performance and scale since the past CHEP06 conference. The production system has been re-engineered in order to incorporate the experience gained in running the previous system and to integrate production with the new CMS event data model, data management system and data processing framework. The system is interfaced to the two major computing Grids used by CMS, the LHC Computing Grid (LCG) and the Open Science Grid (OSG). Operational experience and integration aspects of the new CMS Monte Carlo production system is presented together with an analysis of production statistics. The new system automatically handles job submission, resource monitoring, job queuing, job distribution according to the available resources, data merging, registration of data into the data bookkeeping, data location, data transfer and placement systems. Compared to the previous production system automation, reliability and performance have been considerably improved. A more efficient use of computing resources and a better handling of the inherent Grid unreliability have resulted in an increase of production scale by about an order of magnitude, capable of running in parallel at the order of ten thousand jobs and yielding more than two million events per day

  12. Using CREAM and CEMonitor for job submission and management in the gLite middleware

    Energy Technology Data Exchange (ETDEWEB)

    Aiftimiei, C; Andreetto, P; Bertocco, S; Dalla Fina, S; Dorigo, A; Frizziero, E; Gianelle, A; Mazzucato, M; Sgaravatto, M; Traldi, S; Zangrando, L [INFN Padova, Via Marzolo 8, I-35131 Padova (Italy); Marzolla, M [Dipartimento di Scienze dell' Informazione, Universita di Bologna, Mura A. Zamboni 7, I-40127 Bologna (Italy); Lorenzo, P Mendez; Miccio, V [CERN, BAT. 28-1-019, 1211 Geneve (Switzerland)

    2010-04-01

    In this paper we describe the use of CREAM and CEMonitor services for job submission and management within the gLite Grid middleware. Both CREAM and CEMonitor address one of the most fundamental operations of a Grid middleware, that is job submission and management. Specifically, CREAM is a job management service used for submitting, managing and monitoring computational jobs. CEMonitor is an event notification framework, which can be coupled with CREAM to provide the users with asynchronous job status change notifications. Both components have been integrated in the gLite Workload Management System by means of ICE (Interface to CREAM Environment). These software components have been released for production in the EGEE Grid infrastructure and, for what concerns the CEMonitor service, also in the OSG Grid. In this paper we report the current status of these services, the achieved results, and the issues that still have to be addressed.

  13. Using CREAM and CEMonitor for job submission and management in the gLite middleware

    International Nuclear Information System (INIS)

    Aiftimiei, C; Andreetto, P; Bertocco, S; Dalla Fina, S; Dorigo, A; Frizziero, E; Gianelle, A; Mazzucato, M; Sgaravatto, M; Traldi, S; Zangrando, L; Marzolla, M; Lorenzo, P Mendez; Miccio, V

    2010-01-01

    In this paper we describe the use of CREAM and CEMonitor services for job submission and management within the gLite Grid middleware. Both CREAM and CEMonitor address one of the most fundamental operations of a Grid middleware, that is job submission and management. Specifically, CREAM is a job management service used for submitting, managing and monitoring computational jobs. CEMonitor is an event notification framework, which can be coupled with CREAM to provide the users with asynchronous job status change notifications. Both components have been integrated in the gLite Workload Management System by means of ICE (Interface to CREAM Environment). These software components have been released for production in the EGEE Grid infrastructure and, for what concerns the CEMonitor service, also in the OSG Grid. In this paper we report the current status of these services, the achieved results, and the issues that still have to be addressed.

  14. Modeling and Grid impedance Variation Analysis of Parallel Connected Grid Connected Inverter based on Impedance Based Harmonic Analysis

    DEFF Research Database (Denmark)

    Kwon, JunBum; Wang, Xiongfei; Bak, Claus Leth

    2014-01-01

    This paper addresses the harmonic compensation error problem existing with parallel connected inverter in the same grid interface conditions by means of impedance-based analysis and modeling. Unlike the single grid connected inverter, it is found that multiple parallel connected inverters and grid...... impedance can make influence to each other if they each have a harmonic compensation function. The analysis method proposed in this paper is based on the relationship between the overall output impedance and input impedance of parallel connected inverter, where controller gain design method, which can...

  15. Job submission and management through web services: the experience with the CREAM service

    Energy Technology Data Exchange (ETDEWEB)

    Aiftimiei, C; Andreetto, P; Bertocco, S; Fina, S D; Ronco, S D; Dorigo, A; Gianelle, A; Marzolla, M; Mazzucato, M; Sgaravatto, M; Verlato, M; Zangrando, L [INFN Sezione di Padova, Via Marzolo 8, 35131 Padova (Italy); Corvo, M; Miccio, V; Sciaba, A [CERN, BAT. 28-1-019, 1211 Geneve (Switzerland); Cesini, D; Dongiovanni, D [INFN CNAF, viale Berti Pichat 6/2, 40127 Bologna (Italy); Grandi, C [INFN Sezione di Bologna, viale Berti Pichat 6/2, 40127 Bologna (Italy)

    2008-07-15

    Modern Grid middleware is built around components providing basic functionality, such as data storage, authentication, security, job management, resource monitoring and reservation. In this paper we describe the Computing Resource Execution and Management (CREAM) service. CREAM provides a Web service-based job execution and management capability for Grid systems; in particular, it is being used within the gLite middleware. CREAM exposes a Web service interface allowing conforming clients to submit and manage computational jobs to a Local Resource Management System. We developed a special component, called ICE (Interface to CREAM Environment) to integrate CREAM in gLite. ICE transfers job submissions and cancellations from the Workload Management System, allowing users to manage CREAM jobs from the gLite User Interface. This paper describes some recent studies aimed at assessing the performance and reliability of CREAM and ICE; those tests have been performed as part of the acceptance tests for integration of CREAM and ICE in gLite. We also discuss recent work towards enhancing CREAM with a BES and JSDL compliant interface.

  16. Job submission and management through web services: the experience with the CREAM service

    International Nuclear Information System (INIS)

    Aiftimiei, C; Andreetto, P; Bertocco, S; Fina, S D; Ronco, S D; Dorigo, A; Gianelle, A; Marzolla, M; Mazzucato, M; Sgaravatto, M; Verlato, M; Zangrando, L; Corvo, M; Miccio, V; Sciaba, A; Cesini, D; Dongiovanni, D; Grandi, C

    2008-01-01

    Modern Grid middleware is built around components providing basic functionality, such as data storage, authentication, security, job management, resource monitoring and reservation. In this paper we describe the Computing Resource Execution and Management (CREAM) service. CREAM provides a Web service-based job execution and management capability for Grid systems; in particular, it is being used within the gLite middleware. CREAM exposes a Web service interface allowing conforming clients to submit and manage computational jobs to a Local Resource Management System. We developed a special component, called ICE (Interface to CREAM Environment) to integrate CREAM in gLite. ICE transfers job submissions and cancellations from the Workload Management System, allowing users to manage CREAM jobs from the gLite User Interface. This paper describes some recent studies aimed at assessing the performance and reliability of CREAM and ICE; those tests have been performed as part of the acceptance tests for integration of CREAM and ICE in gLite. We also discuss recent work towards enhancing CREAM with a BES and JSDL compliant interface

  17. Ecosystem Based Business Model of Smart Grid

    DEFF Research Database (Denmark)

    Lundgaard, Morten Raahauge; Ma, Zheng; Jørgensen, Bo Nørregaard

    2015-01-01

    This paper tries to investigate the ecosystem based business model in a smart grid infrastructure and the potential of value capture in the highly complex macro infrastructure such as smart grid. This paper proposes an alternative perspective to study the smart grid business ecosystem to support...... the infrastructural challenges, such as the interoperability of business components for smart grid. So far little research has explored the business ecosystem in the smart grid concept. The study on the smart grid with the theory of business ecosystem may open opportunities to understand market catalysts. This study...... contributes an understanding of business ecosystem applicable for smart grid. Smart grid infrastructure is an intricate business ecosystem, which have several intentions to deliver the value proposition and what it should be. The findings help to identify and capture value from markets....

  18. GLOA: A New Job Scheduling Algorithm for Grid Computing

    Directory of Open Access Journals (Sweden)

    Zahra Pooranian

    2013-03-01

    Full Text Available The purpose of grid computing is to produce a virtual supercomputer by using free resources available through widespread networks such as the Internet. This resource distribution, changes in resource availability, and an unreliable communication infrastructure pose a major challenge for efficient resource allocation. Because of the geographical spread of resources and their distributed management, grid scheduling is considered to be a NP-complete problem. It has been shown that evolutionary algorithms offer good performance for grid scheduling. This article uses a new evaluation (distributed algorithm inspired by the effect of leaders in social groups, the group leaders' optimization algorithm (GLOA, to solve the problem of scheduling independent tasks in a grid computing system. Simulation results comparing GLOA with several other evaluation algorithms show that GLOA produces shorter makespans.

  19. Feature combination analysis in smart grid based using SOM for Sudan national grid

    Science.gov (United States)

    Bohari, Z. H.; Yusof, M. A. M.; Jali, M. H.; Sulaima, M. F.; Nasir, M. N. M.

    2015-12-01

    In the investigation of power grid security, the cascading failure in multicontingency situations has been a test because of its topological unpredictability and computational expense. Both system investigations and burden positioning routines have their limits. In this project, in view of sorting toward Self Organizing Maps (SOM), incorporated methodology consolidating spatial feature (distance)-based grouping with electrical attributes (load) to evaluate the vulnerability and cascading impact of various part sets in the force lattice. Utilizing the grouping result from SOM, sets of overwhelming stacked beginning victimized people to perform assault conspires and asses the consequent falling impact of their failures, and this SOM-based approach viably distinguishes the more powerless sets of substations than those from the conventional burden positioning and other bunching strategies. The robustness of power grids is a central topic in the design of the so called "smart grid". In this paper, to analyze the measures of importance of the nodes in a power grid under cascading failure. With these efforts, we can distinguish the most vulnerable nodes and protect them, improving the safety of the power grid. Also we can measure if a structure is proper for power grids.

  20. Experimental Demonstration of a Self-organized Architecture for Emerging Grid Computing Applications on OBS Testbed

    Science.gov (United States)

    Liu, Lei; Hong, Xiaobin; Wu, Jian; Lin, Jintong

    As Grid computing continues to gain popularity in the industry and research community, it also attracts more attention from the customer level. The large number of users and high frequency of job requests in the consumer market make it challenging. Clearly, all the current Client/Server(C/S)-based architecture will become unfeasible for supporting large-scale Grid applications due to its poor scalability and poor fault-tolerance. In this paper, based on our previous works [1, 2], a novel self-organized architecture to realize a highly scalable and flexible platform for Grids is proposed. Experimental results show that this architecture is suitable and efficient for consumer-oriented Grids.

  1. GrEMBOSS: EMBOSS over the EELA GRID

    International Nuclear Information System (INIS)

    Bonavides-Martinez, C.; Murrieta-Leon, E.; Verleyen, J.; Zayas-Lagunas, R.; Hernandez-Alvarez, A.; Rodriguez-Bahena, R.; Valverde, J. R.; Branger, P. A.; Sarachu, M.

    2007-01-01

    With the growth of genome databases and the implied complexity for processing such information within bioinformatics research, there is a need for computing power and massive storage facilities which can be provided by Grid infrastructures. EMBOSS is a free Open Source sequence analysis package specially developed for the needs of the bioinformatics and molecular biology user community. This work describes the deployment of EMBOSS over the EELA and EGEE Grids, both gLite middle ware-based infrastructures. This work is focused on rewriting the I/O EMBOSS libraries (AJAX) to use the GFAL from the LCG/EGEE middle ware. This library allows the use of files registered on the catalog service which are contained in the storage elements of a Grid. Submitting a job into a Grid is not an intuitive task. This work also describes an ad hoc mechanism to allow bioinformaticians to concentrate on the EMBOSS command, instead of acquiring advanced knowledge about Grid usage. The results obtained so far demonstrate the functionality of GrEMBOSS, and represent an efficient and viable alternative for gridifying other bioinformatics applications. (Author)

  2. GrEMBOSS: EMBOSS over the EELA GRID

    Energy Technology Data Exchange (ETDEWEB)

    Bonavides-Martinez, C.; Murrieta-Leon, E.; Verleyen, J.; Zayas-Lagunas, R.; Hernandez-Alvarez, A.; Rodriguez-Bahena, R.; Valverde, J. R.; Branger, P. A.; Sarachu, M.

    2007-07-01

    With the growth of genome databases and the implied complexity for processing such information within bioinformatics research, there is a need for computing power and massive storage facilities which can be provided by Grid infrastructures. EMBOSS is a free Open Source sequence analysis package specially developed for the needs of the bioinformatics and molecular biology user community. This work describes the deployment of EMBOSS over the EELA and EGEE Grids, both gLite middle ware-based infrastructures. This work is focused on rewriting the I/O EMBOSS libraries (AJAX) to use the GFAL from the LCG/EGEE middle ware. This library allows the use of files registered on the catalog service which are contained in the storage elements of a Grid. Submitting a job into a Grid is not an intuitive task. This work also describes an ad hoc mechanism to allow bioinformaticians to concentrate on the EMBOSS command, instead of acquiring advanced knowledge about Grid usage. The results obtained so far demonstrate the functionality of GrEMBOSS, and represent an efficient and viable alternative for gridifying other bioinformatics applications. (Author)

  3. Context-Aware Usage-Based Grid Authorization Framework

    Institute of Scientific and Technical Information of China (English)

    CUI Yongquan; HONG Fan; FU Cai

    2006-01-01

    Due to inherent heterogeneity, multi-domain characteristic and highly dynamic nature, authorization is a critical concern in grid computing. This paper proposes a general authorization and access control architecture, grid usage control (GUCON), for grid computing. It's based on the next generation access control mechanism usage control (UCON) model. The GUCON Framework dynamic grants and adapts permission to the subject based on a set of contextual information collected from the system environments; while retaining the authorization by evaluating access requests based on subject attributes, object attributes and requests. In general, GUCON model provides very flexible approaches to adapt the dynamically security request. GUCON model is being implemented in our experiment prototype.

  4. Grid-system element of LCG-2 in DLNP

    International Nuclear Information System (INIS)

    Dolbilov, A.G.; Ivanov, Yu.P.

    2008-01-01

    We present here the description of the computer cluster at the Dzhelepov Laboratory of Nuclear Problems of JINR, where the second Grid note at JINR was realized. The configuration of the system, which allows effective joint usage of cluster resources both for local users and for others within the framework of the ATLAS collaboration is examined. Examples are given for basic stages of preparing and running ordinary cluster jobs and with the Grid usages, starting from obtaining CA certificates, submitting jobs and retrieving the results. Perspectives of the cluster upgrade are discussed

  5. Scheduling in Heterogeneous Grid Environments: The Effects of DataMigration

    Energy Technology Data Exchange (ETDEWEB)

    Oliker, Leonid; Biswas, Rupak; Shan, Hongzhang; Smith, Warren

    2004-01-01

    Computational grids have the potential for solving large-scale scientific problems using heterogeneous and geographically distributed resources. However, a number of major technical hurdles must be overcome before this goal can be fully realized. One problem critical to the effective utilization of computational grids is efficient job scheduling. Our prior work addressed this challenge by defining a grid scheduling architecture and several job migration strategies. The focus of this study is to explore the impact of data migration under a variety of demanding grid conditions. We evaluate our grid scheduling algorithms by simulating compute servers, various groupings of servers into sites, and inter-server networks, using real workloads obtained from leading supercomputing centers. Several key performance metrics are used to compare the behavior of our algorithms against reference local and centralized scheduling schemes. Results show the tremendous benefits of grid scheduling, even in the presence of input/output data migration - while highlighting the importance of utilizing communication-aware scheduling schemes.

  6. Peer-to-peer Cooperative Scheduling Architecture for National Grid Infrastructure

    Science.gov (United States)

    Matyska, Ludek; Ruda, Miroslav; Toth, Simon

    For some ten years, the Czech National Grid Infrastructure MetaCentrum uses a single central PBSPro installation to schedule jobs across the country. This centralized approach keeps a full track about all the clusters, providing support for jobs spanning several sites, implementation for the fair-share policy and better overall control of the grid environment. Despite a steady progress in the increased stability and resilience to intermittent very short network failures, growing number of sites and processors makes this architecture, with a single point of failure and scalability limits, obsolete. As a result, a new scheduling architecture is proposed, which relies on higher autonomy of clusters. It is based on a peer to peer network of semi-independent schedulers for each site or even cluster. Each scheduler accepts jobs for the whole infrastructure, cooperating with other schedulers on implementation of global policies like central job accounting, fair-share, or submission of jobs across several sites. The scheduling system is integrated with the Magrathea system to support scheduling of virtual clusters, including the setup of their internal network, again eventually spanning several sites. On the other hand, each scheduler is local to one of several clusters and is able to directly control and submit jobs to them even if the connection of other scheduling peers is lost. In parallel to the change of the overall architecture, the scheduling system itself is being replaced. Instead of PBSPro, chosen originally for its declared support of large scale distributed environment, the new scheduling architecture is based on the open-source Torque system. The implementation and support for the most desired properties in PBSPro and Torque are discussed and the necessary modifications to Torque to support the MetaCentrum scheduling architecture are presented, too.

  7. A tool for optimization of the production and user analysis on the Grid, C. Grigoras for the ALICE Collaboration

    Science.gov (United States)

    Grigoras, Costin; Carminati, Federico; Vladimirovna Datskova, Olga; Schreiner, Steffen; Lee, Sehoon; Zhu, Jianlin; Gheata, Mihaela; Gheata, Andrei; Saiz, Pablo; Betev, Latchezar; Furano, Fabrizio; Mendez Lorenzo, Patricia; Grigoras, Alina Gabriela; Bagnasco, Stefano; Peters, Andreas Joachim; Saiz Santos, Maria Dolores

    2011-12-01

    With the LHC and ALICE entering a full operation and production modes, the amount of Simulation and RAW data processing and end user analysis computational tasks are increasing. The efficient management of all these tasks, all of which have large differences in lifecycle, amounts of processed data and methods to analyze the end result, required the development and deployment of new tools in addition to the already existing Grid infrastructure. To facilitate the management of the large scale simulation and raw data reconstruction tasks, ALICE has developed a production framework called a Lightweight Production Manager (LPM). The LPM is automatically submitting jobs to the Grid based on triggers and conditions, for example after a physics run completion. It follows the evolution of the job and publishes the results on the web for worldwide access by the ALICE physicists. This framework is tightly integrated with the ALICE Grid framework AliEn. In addition to the publication of the job status, LPM is also allowing a fully authenticated interface to the AliEn Grid catalogue, to browse and download files, and in the near future will provide simple types of data analysis through ROOT plugins. The framework is also being extended to allow management of end user jobs.

  8. A tool for optimization of the production and user analysis on the Grid, C. Grigoras for the ALICE Collaboration

    International Nuclear Information System (INIS)

    Grigoras, Costin; Carminati, Federico; Vladimirovna Datskova, Olga; Schreiner, Steffen; Gheata, Mihaela; Gheata, Andrei; Saiz, Pablo; Betev, Latchezar; Furano, Fabrizio; Lorenzo, Patricia Mendez; Grigoras, Alina Gabriela; Peters, Andreas Joachim; Saiz Santos, Maria Dolores; Lee, Sehoon; Zhu Jianlin; Bagnasco, Stefano

    2011-01-01

    With the LHC and ALICE entering a full operation and production modes, the amount of Simulation and RAW data processing and end user analysis computational tasks are increasing. The efficient management of all these tasks, all of which have large differences in lifecycle, amounts of processed data and methods to analyze the end result, required the development and deployment of new tools in addition to the already existing Grid infrastructure. To facilitate the management of the large scale simulation and raw data reconstruction tasks, ALICE has developed a production framework called a Lightweight Production Manager (LPM). The LPM is automatically submitting jobs to the Grid based on triggers and conditions, for example after a physics run completion. It follows the evolution of the job and publishes the results on the web for worldwide access by the ALICE physicists. This framework is tightly integrated with the ALICE Grid framework AliEn. In addition to the publication of the job status, LPM is also allowing a fully authenticated interface to the AliEn Grid catalogue, to browse and download files, and in the near future will provide simple types of data analysis through ROOT plugins. The framework is also being extended to allow management of end user jobs.

  9. AliEn - GRID application for ALICE Collaboration

    International Nuclear Information System (INIS)

    Zgura, Ion-Sorin

    2003-01-01

    AliEn (ALICE Environment) is a GRID framework built on top of the latest Internet standards for information exchange and authentication (SOAP, PKI) and common Open Source components. AliEn provides a virtual file catalogue that allows transparent access to distributed data-sets and a number of collaborating Web services which implement the authentication, job execution, file transport, performance monitor and event logging.The ALICE experiment has developed AliEn as an implementation of distributed computing infrastructure needed to simulate, reconstruct and analyze data from the experiment. The sites that belong to the ALICE Virtual Organisation can be seen and used as a single entity - any available node executes jobs and access to logical and datasets is transparent to the user. In developing AliEn common standards and solutions in the form of Open Source components were used. Only 1% (25k physical lines of code in Perl) is native AliEn code while 99% of the code has been imported in form of Open Sources packages and Perl modules. Currently ALICE is using the system for distributed production of Monte Carlo data at over 30 sites on four continents. During the last twelve months more than 30,000 jobs have been successfully run under AliEn control worldwide, totalling 25 CPU years and producing 20 TB of data. The user interface is compatible to EU DataGrid at the level of authentication and job description language. In perspective AliEn will be interfaced to the mainstream Grid infrastructure in HEP and it will remain to serve as interface between ALICE Offline framework and external Grid infrastructure. (authors)

  10. The Construction of Job Families Based on Company Specific PAQ Job Dimensions.

    Science.gov (United States)

    Taylor, L. R.; Colbert, G. A.

    1978-01-01

    Research is presented on the construction of job families based on Position Analysis Questionnaire data. The data were subjected to a component analysis. Results were interpreted as sufficiently encouraging to proceed with analyses of validity generalization within the job families. (Editor/RK)

  11. Forecasting Model for Network Throughput of Remote Data Access in Computing Grids

    CERN Document Server

    Begy, Volodimir; The ATLAS collaboration

    2018-01-01

    Computing grids are one of the key enablers of eScience. Researchers from many fields (e.g. High Energy Physics, Bioinformatics, Climatology, etc.) employ grids to run computational jobs in a highly distributed manner. The current state of the art approach for data access in the grid is data placement: a job is scheduled to run at a specific data center, and its execution starts only when the complete input data has been transferred there. This approach has two major disadvantages: (1) the jobs are staying idle while waiting for the input data; (2) due to the limited infrastructure resources, the distributed data management system handling the data placement, may queue the transfers up to several days. An alternative approach is remote data access: a job may stream the input data directly from storage elements, which may be located at local or remote data centers. Remote data access brings two innovative benefits: (1) the jobs can be executed asynchronously with respect to the data transfer; (2) when combined...

  12. glideinWMS - A generic pilot-based Workload Management System

    International Nuclear Information System (INIS)

    Sfiligoi, Igor

    2007-01-01

    The Grid resources are distributed among hundreds of independent Grid sites, requiring a higher level Workload Management System (WMS) to be used efficiently. Pilot jobs have been used for this purpose by many communities, bringing increased reliability, global fair share and just in time resource matching. GlideinWMS is a WMS based on the Condor glidein concept, i.e. a regular Condor pool, with the Condor daemons (startds) being started by pilot jobs, and real jobs being vanilla, standard or MPI universe jobs. The glideinWMS is composed of a set of Glidein Factories, handling the submission of pilot jobs to a set of Grid sites, and a set of VO Frontends, requesting pilot submission based on the status of user jobs. This paper contains the structural overview of glideinWMS as well as a detailed description of the current implementation and the current scalability limits

  13. glideinWMS-a generic pilot-based workload management system

    International Nuclear Information System (INIS)

    Sfiligoi, I

    2008-01-01

    The Grid resources are distributed among hundreds of independent Grid sites, requiring a higher level Workload Management System (WMS) to be used efficiently. Pilot jobs have been used for this purpose by many communities, bringing increased reliability, global fair share and just in time resource matching. glideinWMS is a WMS based on the Condor glidein concept, i.e. a regular Condor pool, with the Condor daemons (startds) being started by pilot jobs, and real jobs being vanilla, standard or MPI universe jobs. The glideinWMS is composed of a set of Glidein Factories, handling the submission of pilot jobs to a set of Grid sites, and a set of VO Frontends, requesting pilot submission based on the status of user jobs. This paper contains the structural overview of glideinWMS as well as a detailed description of the current implementation and the current scalability limits

  14. The CMS Integration Grid Testbed

    CERN Document Server

    Graham, G E; Aziz, Shafqat; Bauerdick, L.A.T.; Ernst, Michael; Kaiser, Joseph; Ratnikova, Natalia; Wenzel, Hans; Wu, Yu-jun; Aslakson, Erik; Bunn, Julian; Iqbal, Saima; Legrand, Iosif; Newman, Harvey; Singh, Suresh; Steenberg, Conrad; Branson, James; Fisk, Ian; Letts, James; Arbree, Adam; Avery, Paul; Bourilkov, Dimitri; Cavanaugh, Richard; Rodriguez, Jorge Luis; Kategari, Suchindra; Couvares, Peter; DeSmet, Alan; Livny, Miron; Roy, Alain; Tannenbaum, Todd; Graham, Gregory E.; Aziz, Shafqat; Ernst, Michael; Kaiser, Joseph; Ratnikova, Natalia; Wenzel, Hans; Wu, Yujun; Aslakson, Erik; Bunn, Julian; Iqbal, Saima; Legrand, Iosif; Newman, Harvey; Singh, Suresh; Steenberg, Conrad; Branson, James; Fisk, Ian; Letts, James; Arbree, Adam; Avery, Paul; Bourilkov, Dimitri; Cavanaugh, Richard; Rodriguez, Jorge; Kategari, Suchindra; Couvares, Peter; Smet, Alan De; Livny, Miron; Roy, Alain; Tannenbaum, Todd

    2003-01-01

    The CMS Integration Grid Testbed (IGT) comprises USCMS Tier-1 and Tier-2 hardware at the following sites: the California Institute of Technology, Fermi National Accelerator Laboratory, the University of California at San Diego, and the University of Florida at Gainesville. The IGT runs jobs using the Globus Toolkit with a DAGMan and Condor-G front end. The virtual organization (VO) is managed using VO management scripts from the European Data Grid (EDG). Gridwide monitoring is accomplished using local tools such as Ganglia interfaced into the Globus Metadata Directory Service (MDS) and the agent based Mona Lisa. Domain specific software is packaged and installed using the Distrib ution After Release (DAR) tool of CMS, while middleware under the auspices of the Virtual Data Toolkit (VDT) is distributed using Pacman. During a continuo us two month span in Fall of 2002, over 1 million official CMS GEANT based Monte Carlo events were generated and returned to CERN for analysis while being demonstrated at SC2002. ...

  15. Advanced technologies for scalable ATLAS conditions database access on the grid

    International Nuclear Information System (INIS)

    Basset, R; Canali, L; Girone, M; Hawkings, R; Valassi, A; Viegas, F; Dimitrov, G; Nevski, P; Vaniachine, A; Walker, R; Wong, A

    2010-01-01

    During massive data reprocessing operations an ATLAS Conditions Database application must support concurrent access from numerous ATLAS data processing jobs running on the Grid. By simulating realistic work-flow, ATLAS database scalability tests provided feedback for Conditions Db software optimization and allowed precise determination of required distributed database resources. In distributed data processing one must take into account the chaotic nature of Grid computing characterized by peak loads, which can be much higher than average access rates. To validate database performance at peak loads, we tested database scalability at very high concurrent jobs rates. This has been achieved through coordinated database stress tests performed in series of ATLAS reprocessing exercises at the Tier-1 sites. The goal of database stress tests is to detect scalability limits of the hardware deployed at the Tier-1 sites, so that the server overload conditions can be safely avoided in a production environment. Our analysis of server performance under stress tests indicates that Conditions Db data access is limited by the disk I/O throughput. An unacceptable side-effect of the disk I/O saturation is a degradation of the WLCG 3D Services that update Conditions Db data at all ten ATLAS Tier-1 sites using the technology of Oracle Streams. To avoid such bottlenecks we prototyped and tested a novel approach for database peak load avoidance in Grid computing. Our approach is based upon the proven idea of pilot job submission on the Grid: instead of the actual query, an ATLAS utility library sends to the database server a pilot query first.

  16. Automated Grid Monitoring for LHCb through HammerCloud

    CERN Multimedia

    CERN. Geneva

    2015-01-01

    The HammerCloud system is used by CERN IT to monitor the status of the Worldwide LHC Computing Grid (WLCG). HammerCloud automatically submits jobs to WLCG computing resources, closely replicating the workflow of Grid users (e.g. physicists analyzing data). This allows computation nodes and storage resources to be monitored, software to be tested (somewhat like continuous integration), and new sites to be stress tested with a heavy job load before commissioning. The HammerCloud system has been in use for ATLAS and CMS experiments for about five years. This summer's work involved porting the HammerCloud suite of tools to the LHCb experiment. The HammerCloud software runs functional tests and provides data visualizations. HammerCloud's LHCb variant is written in Python, using the Django web framework and Ganga/DIRAC for job management.

  17. GENECODIS-Grid: An online grid-based tool to predict functional information in gene lists

    International Nuclear Information System (INIS)

    Nogales, R.; Mejia, E.; Vicente, C.; Montes, E.; Delgado, A.; Perez Griffo, F. J.; Tirado, F.; Pascual-Montano, A.

    2007-01-01

    In this work we introduce GeneCodis-Grid, a grid-based alternative to a bioinformatics tool named Genecodis that integrates different sources of biological information to search for biological features (annotations) that frequently co-occur in a set of genes and rank them by statistical significance. GeneCodis-Grid is a web-based application that takes advantage of two independent grid networks and a computer cluster managed by a meta-scheduler and a web server that host the application. The mining of concurrent biological annotations provides significant information for the functional analysis of gene list obtained by high throughput experiments in biology. Due to the large popularity of this tool, that has registered more than 13000 visits since its publication in January 2007, there is a strong need to facilitate users from different sites to access the system simultaneously. In addition, the complexity of some of the statistical tests used in this approach has made this technique a good candidate for its implementation in a Grid opportunistic environment. (Author)

  18. Measurement and simulation of the performance of high energy physics data grids

    Science.gov (United States)

    Crosby, Paul Andrew

    This thesis describes a study of resource brokering in a computational Grid for high energy physics. Such systems are being devised in order to manage the unprecedented workload of the next generation particle physics experiments such as those at the Large Hadron Collider. A simulation of the European Data Grid has been constructed, and calibrated using logging data from a real Grid testbed. This model is then used to explore the Grid's middleware configuration, and suggest improvements to its scheduling policy. The expansion of the simulation to include data analysis of the type conducted by particle physicists is then described. A variety of job and data management policies are explored, in order to determine how well they meet the needs of physicists, as well as how efficiently they make use of CPU and network resources. Appropriate performance indicators are introduced in order to measure how well jobs and resources are managed from different perspectives. The effects of inefficiencies in Grid middleware are explored, as are methods of compensating for them. It is demonstrated that a scheduling algorithm should alter its weighting on load balancing and data distribution, depending on whether data transfer or CPU requirements dominate, and also on the level of job loading. It is also shown that an economic model for data management and replication can improve the efficiency of network use and job processing.

  19. OGC and Grid Interoperability in enviroGRIDS Project

    Science.gov (United States)

    Gorgan, Dorian; Rodila, Denisa; Bacu, Victor; Giuliani, Gregory; Ray, Nicolas

    2010-05-01

    EnviroGRIDS (Black Sea Catchment Observation and Assessment System supporting Sustainable Development) [1] is a 4-years FP7 Project aiming to address the subjects of ecologically unsustainable development and inadequate resource management. The project develops a Spatial Data Infrastructure of the Black Sea Catchment region. The geospatial technologies offer very specialized functionality for Earth Science oriented applications as well as the Grid oriented technology that is able to support distributed and parallel processing. One challenge of the enviroGRIDS project is the interoperability between geospatial and Grid infrastructures by providing the basic and the extended features of the both technologies. The geospatial interoperability technology has been promoted as a way of dealing with large volumes of geospatial data in distributed environments through the development of interoperable Web service specifications proposed by the Open Geospatial Consortium (OGC), with applications spread across multiple fields but especially in Earth observation research. Due to the huge volumes of data available in the geospatial domain and the additional introduced issues (data management, secure data transfer, data distribution and data computation), the need for an infrastructure capable to manage all those problems becomes an important aspect. The Grid promotes and facilitates the secure interoperations of geospatial heterogeneous distributed data within a distributed environment, the creation and management of large distributed computational jobs and assures a security level for communication and transfer of messages based on certificates. This presentation analysis and discusses the most significant use cases for enabling the OGC Web services interoperability with the Grid environment and focuses on the description and implementation of the most promising one. In these use cases we give a special attention to issues such as: the relations between computational grid and

  20. Greedy and metaheuristics for the offline scheduling problem in grid computing

    DEFF Research Database (Denmark)

    Gamst, Mette

    In grid computing a number of geographically distributed resources connected through a wide area network, are utilized as one computations unit. The NP-hard offline scheduling problem in grid computing consists of assigning jobs to resources in advance. In this paper, five greedy heuristics and two....... All heuristics solve instances with up to 2000 jobs and 1000 resources, thus the results are useful both with respect to running times and to solution values....

  1. The LHCb Experience on the Grid from the DIRAC Accounting Data

    CERN Document Server

    Puig, A; Graciani, R; Casajús, A

    2011-01-01

    DIRAC is the software framework developed by LHCb to manage all its computing operations on the Grid. Since 2003 it has been used for large scale Monte Carlo simulation productions and for user analysis of these data. Since the end of 2009, with the start-up of LHC, DIRAC also takes care of the distribution, reconstruction, selection and analysis of the physics data taken by the detector apparatus. During 2009, DIRAC executed almost 5 million jobs for LHCb. In order to execute this workload slightly over 6 million of pilot jobs were submitted, out of which approximately one third were aborted by the Grid infrastructure. In 2010, thanks to their improved efficiency, DIRAC pilots are able, on average, to match and execute between 2 and 3 LHCb jobs during their lifetime, largely reducing the load on the Grid infrastructure. Given the large amount of submitted jobs and used resources, it becomes essential to store detailed information about their execution to track the behaviour of the system. The DIRAC Accountin...

  2. FIFE-Jobsub: a grid submission system for intensity frontier experiments at Fermilab

    International Nuclear Information System (INIS)

    Box, Dennis

    2014-01-01

    The Fermilab Intensity Frontier Experiments use an integrated submission system known as FIFE-jobsub, part of the FIFE (Fabric for Frontier Experiments) initiative, to submit batch jobs to the Open Science Grid. FIFE-jobsub eases the burden on experimenters by integrating data transfer and site selection details in an easy to use and well-documented format. FIFE-jobsub automates tedious details of maintaining grid proxies for the lifetime of the grid job. Data transfer is handled using the Intensity Frontier Data Handling Client (IFDHC) [1] tool suite, which facilitates selecting the appropriate data transfer method from many possibilities while protecting shared resources from overload. Chaining of job dependencies into Directed Acyclic Graphs (Condor DAGS) is well supported and made easier through the use of input flags and parameters.

  3. FIFE-Jobsub: a grid submission system for intensity frontier experiments at Fermilab

    Science.gov (United States)

    Box, Dennis

    2014-06-01

    The Fermilab Intensity Frontier Experiments use an integrated submission system known as FIFE-jobsub, part of the FIFE (Fabric for Frontier Experiments) initiative, to submit batch jobs to the Open Science Grid. FIFE-jobsub eases the burden on experimenters by integrating data transfer and site selection details in an easy to use and well-documented format. FIFE-jobsub automates tedious details of maintaining grid proxies for the lifetime of the grid job. Data transfer is handled using the Intensity Frontier Data Handling Client (IFDHC) [1] tool suite, which facilitates selecting the appropriate data transfer method from many possibilities while protecting shared resources from overload. Chaining of job dependencies into Directed Acyclic Graphs (Condor DAGS) is well supported and made easier through the use of input flags and parameters.

  4. Smart Grid Cybersecurity: Job Performance Model Report

    Energy Technology Data Exchange (ETDEWEB)

    O' Neil, Lori Ross; Assante, Michael; Tobey, David

    2012-08-01

    This is the project report to DOE OE-30 for the completion of Phase 1 of a 3 phase report. This report outlines the work done to develop a smart grid cybersecurity certification. This work is being done with the subcontractor NBISE.

  5. Adaptive Micro-Grid Operation Based on IEC 61850

    Directory of Open Access Journals (Sweden)

    Wei Deng

    2015-05-01

    Full Text Available Automatically identifying the new equipment after its integration and adjusting operation strategy to realize “plug and play” functionality are becoming essential for micro-grid operations. In order to improve and perfect the micro-grid “plug and play” function with the increased amount of equipment with different information protocols and more diverse system applications, this paper presents a solution for adaptive micro-grid operation based on IEC 61850, and proposes the design and specific implementation methods of micro-grid “plug and play” function and system operation mode conversion in detail, by using the established IEC 61850 information model of a micro-grid. Actual operation tests based on the developed IED and micro-grid test platform are performed to verify the feasibility and validity of the proposed solution. The tests results show that the solution can automatically identify the IEC 61850 information model of equipment after its integration, intelligently adjust the operation strategies to adapt to new system states and achieves a reliable system operation mode conversion.

  6. Agent-Mining of Grid Log-Files: A Case Study

    NARCIS (Netherlands)

    Stoter, A.; Dalmolen, Simon; Mulder, .W.

    2013-01-01

    Grid monitoring requires analysis of large amounts of log files across multiple domains. An approach is described for automated extraction of job-flow information from large computer grids, using software agents and genetic computation. A prototype was created as a first step towards communities of

  7. Assessment of job stress factors and organizational personality types for procedure-based jobs in nuclear power plants

    International Nuclear Information System (INIS)

    Kim, Dae-Ho; Lee, Yong-Hee; Lee, Jung-Woon

    2008-01-01

    The purpose of this study is to assess the organizational types and the job stress factors that affect procedure-based job performances in nuclear power plants. We derived 24 organizational factors affecting job stress level in nuclear power plants from the job stress analysis models developed by NIOSH, JDI, and IOR. Considering the safety characteristics in the operating tasks of nuclear power plants, we identified the job contents and characteristics through the analyses of job assignments that appeared in the organizational chart and the results of an activity-based costing. By using questionnaire surveys and structured interviews with the plant personnel and expert panels, we assessed 70 jobs among the 777 jobs managed officially in accordance with the procedures. They consist of the representative jobs of each department and are directly related to safety. We utilized the organizational personality type indicators to characterize the personality types of each organization in nuclear power plants. (author)

  8. ReSS: Resource Selection Service for National and Campus Grid Infrastructure

    International Nuclear Information System (INIS)

    Mhashilkar, Parag; Garzoglio, Gabriele; Levshina, Tanya; Timm, Steve

    2010-01-01

    The Open Science Grid (OSG) offers access to around hundred Compute elements (CE) and storage elements (SE) via standard Grid interfaces. The Resource Selection Service (ReSS) is a push-based workload management system that is integrated with the OSG information systems and resources. ReSS integrates standard Grid tools such as Condor, as a brokering service and the gLite CEMon, for gathering and publishing resource information in GLUE Schema format. ReSS is used in OSG by Virtual Organizations (VO) such as Dark Energy Survey (DES), DZero and Engagement VO. ReSS is also used as a Resource Selection Service for Campus Grids, such as FermiGrid. VOs use ReSS to automate the resource selection in their workload management system to run jobs over the grid. In the past year, the system has been enhanced to enable publication and selection of storage resources and of any special software or software libraries (like MPI libraries) installed at computing resources. In this paper, we discuss the Resource Selection Service, its typical usage on the two scales of a National Cyber Infrastructure Grid, such as OSG, and of a campus Grid, such as FermiGrid.

  9. ReSS: Resource Selection Service for National and Campus Grid Infrastructure

    Energy Technology Data Exchange (ETDEWEB)

    Mhashilkar, Parag; Garzoglio, Gabriele; Levshina, Tanya; Timm, Steve, E-mail: parag@fnal.go, E-mail: garzogli@fnal.go, E-mail: tlevshin@fnal.go, E-mail: timm@fnal.go [Fermi National Accelerator Laboratory, P O Box 500, Batavia, IL - 60510 (United States)

    2010-04-01

    The Open Science Grid (OSG) offers access to around hundred Compute elements (CE) and storage elements (SE) via standard Grid interfaces. The Resource Selection Service (ReSS) is a push-based workload management system that is integrated with the OSG information systems and resources. ReSS integrates standard Grid tools such as Condor, as a brokering service and the gLite CEMon, for gathering and publishing resource information in GLUE Schema format. ReSS is used in OSG by Virtual Organizations (VO) such as Dark Energy Survey (DES), DZero and Engagement VO. ReSS is also used as a Resource Selection Service for Campus Grids, such as FermiGrid. VOs use ReSS to automate the resource selection in their workload management system to run jobs over the grid. In the past year, the system has been enhanced to enable publication and selection of storage resources and of any special software or software libraries (like MPI libraries) installed at computing resources. In this paper, we discuss the Resource Selection Service, its typical usage on the two scales of a National Cyber Infrastructure Grid, such as OSG, and of a campus Grid, such as FermiGrid.

  10. ReSS: Resource Selection Service for National and Campus Grid Infrastructure

    International Nuclear Information System (INIS)

    Mhashilkar, Parag; Garzoglio, Gabriele; Levshina, Tanya; Timm, Steve

    2009-01-01

    The Open Science Grid (OSG) offers access to around hundred Compute elements (CE) and storage elements (SE) via standard Grid interfaces. The Resource Selection Service (ReSS) is a push-based workload management system that is integrated with the OSG information systems and resources. ReSS integrates standard Grid tools such as Condor, as a brokering service and the gLite CEMon, for gathering and publishing resource information in GLUE Schema format. ReSS is used in OSG by Virtual Organizations (VO) such as Dark Energy Survey (DES), DZero and Engagement VO. ReSS is also used as a Resource Selection Service for Campus Grids, such as FermiGrid. VOs use ReSS to automate the resource selection in their workload management system to run jobs over the grid. In the past year, the system has been enhanced to enable publication and selection of storage resources and of any special software or software libraries (like MPI libraries) installed at computing resources. In this paper, we discuss the Resource Selection Service, its typical usage on the two scales of a National Cyber Infrastructure Grid, such as OSG, and of a campus Grid, such as FermiGrid.

  11. A framework for job management in the NorduGrid ARC middleware

    DEFF Research Database (Denmark)

    Jensen, Henrik Thostrup; Kleist, Josva; Ryge Leth, Jesper

    2005-01-01

    a certain job status. Unlike other job management systems, our is run on the client side, under the control of the user. This eliminates the need for the user to share a proxy credential, which is needed to control jobs. Furthermore the system can be extended by the user, as it is designed as a framework...

  12. MCRUNJOB: A High energy physics workflow planner for grid production processing

    International Nuclear Information System (INIS)

    Graham, Gregory E.

    2004-01-01

    McRunjob is a powerful grid workflow manager used to manage the generation of large numbers of production processing jobs in High Energy Physics. In use at both the DZero and CMS experiments, McRunjob has been used to manage large Monte Carlo production processing since 1999 and is being extended to uses in regular production processing for analysis and reconstruction. Described at CHEP 2001, McRunjob converts core metadata into jobs submittable in a variety of environments. The powerful core metadata description language includes methods for converting the metadata into persistent forms, job descriptions, multi-step workflows, and data provenance information. The language features allow for structure in the metadata by including full expressions, namespaces, functional dependencies, site specific parameters in a grid environment, and ontological definitions. It also has simple control structures for parallelization of large jobs. McRunjob features a modular design which allows for easy expansion to new job description languages or new application level tasks

  13. MPI support in the DIRAC Pilot Job Workload Management System

    International Nuclear Information System (INIS)

    Tsaregorodtsev, A; Hamar, V

    2012-01-01

    Parallel job execution in the grid environment using MPI technology presents a number of challenges for the sites providing this support. Multiple flavors of the MPI libraries, shared working directories required by certain applications, special settings for the batch systems make the MPI support difficult for the site managers. On the other hand the workload management systems with Pilot Jobs became ubiquitous although the support for the MPI applications in the Pilot frameworks was not available. This support was recently added in the DIRAC Project in the context of the GISELA Latin American Grid Initiative. Special services for dynamic allocation of virtual computer pools on the grid sites were developed in order to deploy MPI rings corresponding to the requirements of the jobs in the central task queue of the DIRAC Workload Management System. Pilot Jobs using user space file system techniques install the required MPI software automatically. The same technique is used to emulate shared working directories for the parallel MPI processes. This makes it possible to execute MPI jobs even on the sites not supporting them officially. Reusing so constructed MPI rings for execution of a series of parallel jobs increases dramatically their efficiency and turnaround. In this contribution we describe the design and implementation of the DIRAC MPI Service as well as its support for various types of MPI libraries. Advantages of coupling the MPI support with the Pilot frameworks are outlined and examples of usage with real applications are presented.

  14. Intrusion Prevention and Detection in Grid Computing - The ALICE Case

    Science.gov (United States)

    Gomez, Andres; Lara, Camilo; Kebschull, Udo

    2015-12-01

    Grids allow users flexible on-demand usage of computing resources through remote communication networks. A remarkable example of a Grid in High Energy Physics (HEP) research is used in the ALICE experiment at European Organization for Nuclear Research CERN. Physicists can submit jobs used to process the huge amount of particle collision data produced by the Large Hadron Collider (LHC). Grids face complex security challenges. They are interesting targets for attackers seeking for huge computational resources. Since users can execute arbitrary code in the worker nodes on the Grid sites, special care should be put in this environment. Automatic tools to harden and monitor this scenario are required. Currently, there is no integrated solution for such requirement. This paper describes a new security framework to allow execution of job payloads in a sandboxed context. It also allows process behavior monitoring to detect intrusions, even when new attack methods or zero day vulnerabilities are exploited, by a Machine Learning approach. We plan to implement the proposed framework as a software prototype that will be tested as a component of the ALICE Grid middleware.

  15. Intrusion Prevention and Detection in Grid Computing - The ALICE Case

    International Nuclear Information System (INIS)

    Gomez, Andres; Lara, Camilo; Kebschull, Udo

    2015-01-01

    Grids allow users flexible on-demand usage of computing resources through remote communication networks. A remarkable example of a Grid in High Energy Physics (HEP) research is used in the ALICE experiment at European Organization for Nuclear Research CERN. Physicists can submit jobs used to process the huge amount of particle collision data produced by the Large Hadron Collider (LHC). Grids face complex security challenges. They are interesting targets for attackers seeking for huge computational resources. Since users can execute arbitrary code in the worker nodes on the Grid sites, special care should be put in this environment. Automatic tools to harden and monitor this scenario are required. Currently, there is no integrated solution for such requirement. This paper describes a new security framework to allow execution of job payloads in a sandboxed context. It also allows process behavior monitoring to detect intrusions, even when new attack methods or zero day vulnerabilities are exploited, by a Machine Learning approach. We plan to implement the proposed framework as a software prototype that will be tested as a component of the ALICE Grid middleware. (paper)

  16. Constraint-based job shop scheduling with ILOG SCHEDULER

    NARCIS (Netherlands)

    Nuijten, W.P.M.; Le Pape, C.

    1998-01-01

    We introduce constraint-based scheduling and discuss its main principles. An approximation algorithm based on tree search is developed for the job shop scheduling problem using ILOG SCHEDULER. A new way of calculating lower bounds on the makespan of the job shop scheduling problem is presented and

  17. Scaling Off Grid Power In Nigeria

    African Journals Online (AJOL)

    storage, digital revolution and especially, the current job deficit in Nigeria led by ... or is self-employed. .... (RESIP), so from that perspective the legal ... and solar home systems or stand-alone grids ... REAN is also driving self-regulatory.

  18. Personality and Education Mining based Job Advisory System

    Directory of Open Access Journals (Sweden)

    Rajendra S. Choudhary

    2014-09-01

    Full Text Available Every job demands an employee with some specific qualities in addition to the basic educational qualification. For example, an introvert person cannot be a good leader despite of a very good academic qualification. Thinking and logical ability is required for a person to be a successful software engineer. So, the aim of this paper is to present a novel approach for advising an ideal job to the job seeker while considering his personality trait and educational qualification both. Very well-known theories of personality like MBTI indicator and OCEAN theory, are used for personality mining. For education mining, score based system is used. The score based system captures the information from attributes like most scoring subject, dream job etc. After personality mining, the resultant values are coalesced with the information extracted from education mining. And finally, the most suited jobs, in terms of personality and educational qualification are recommended to the job seekers. The experiment is conducted on the students who have earned an engineering degree in the field of computer science, information technology and electronics. Nevertheless, the same architecture can easily be extended to other educational degrees also. To the best of the author’s knowledge, this is a first e-job advisory system that recommends the job best suited as per one’s personality using MBTI and OCEAN theory both.

  19. Intrusion Prevention and Detection in Grid Computing - The ALICE Case

    CERN Document Server

    INSPIRE-00416173; Kebschull, Udo

    2015-01-01

    Grids allow users flexible on-demand usage of computing resources through remote communication networks. A remarkable example of a Grid in High Energy Physics (HEP) research is used in the ALICE experiment at European Organization for Nuclear Research CERN. Physicists can submit jobs used to process the huge amount of particle collision data produced by the Large Hadron Collider (LHC). Grids face complex security challenges. They are interesting targets for attackers seeking for huge computational resources. Since users can execute arbitrary code in the worker nodes on the Grid sites, special care should be put in this environment. Automatic tools to harden and monitor this scenario are required. Currently, there is no integrated solution for such requirement. This paper describes a new security framework to allow execution of job payloads in a sandboxed context. It also allows process behavior monitoring to detect intrusions, even when new attack methods or zero day vulnerabilities are exploited, by a Machin...

  20. Dynamically Authorized Role-Based Access Control for Grid Applications

    Institute of Scientific and Technical Information of China (English)

    YAO Hanbing; HU Heping; LU Zhengding; LI Ruixuan

    2006-01-01

    Grid computing is concerned with the sharing and coordinated use of diverse resources in distributed "virtual organizations". The heterogeneous, dynamic and multi-domain nature of these environments makes challenging security issues that demand new technical approaches. Despite the recent advances in access control approaches applicable to Grid computing, there remain issues that impede the development of effective access control models for Grid applications. Among them there are the lack of context-based models for access control, and reliance on identity or capability-based access control schemes. An access control scheme that resolve these issues is presented, and a dynamically authorized role-based access control (D-RBAC) model extending the RBAC with context constraints is proposed. The D-RABC mechanisms dynamically grant permissions to users based on a set of contextual information collected from the system and user's environments, while retaining the advantages of RBAC model. The implementation architecture of D-RBAC for the Grid application is also described.

  1. Automated Grid Monitoring for the LHCb Experiment Through HammerCloud

    CERN Document Server

    Dice, Bradley

    2015-01-01

    The HammerCloud system is used by CERN IT to monitor the status of the Worldwide LHC Computing Grid (WLCG). HammerCloud automatically submits jobs to WLCG computing resources, closely replicating the workflow of Grid users (e.g. physicists analyzing data). This allows computation nodes and storage resources to be monitored, software to be tested (somewhat like continuous integration), and new sites to be stress tested with a heavy job load before commissioning. The HammerCloud system has been in use for ATLAS and CMS experiments for about five years. This summer's work involved porting the HammerCloud suite of tools to the LHCb experiment. The HammerCloud software runs functional tests and provides data visualizations. HammerCloud's LHCb variant is written in Python, using the Django web framework and Ganga/DIRAC for job management.

  2. Improved delayed signal cancellation-based SRF-PLL for unbalanced grid

    DEFF Research Database (Denmark)

    Messo, Tuomas; Sihvo, Jussi; Yang, Dongsheng

    2017-01-01

    Problems with power quality in the grid have gained a lot of attention recently due to rapid increase in the amount of grid-connected power converters. The converter should produce sinusoidal currents also during abnormal conditions, such as unbalanced grid voltages. Several methods, like...... the delayed signal cancellation-based method (DSC), have been proposed to alleviate the detrimental effect of unbalance. This paper proposes an improvement to a delayed signal cancellation based synchronization algorithm for unbalanced grids. The proposed PLL structure employs only half of the delay required...

  3. GridRun: A lightweight packaging and execution environment forcompact, multi-architecture binaries

    Energy Technology Data Exchange (ETDEWEB)

    Shalf, John; Goodale, Tom

    2004-02-01

    GridRun offers a very simple set of tools for creating and executing multi-platform binary executables. These ''fat-binaries'' archive native machine code into compact packages that are typically a fraction the size of the original binary images they store, enabling efficient staging of executables for heterogeneous parallel jobs. GridRun interoperates with existing distributed job launchers/managers like Condor and the Globus GRAM to greatly simplify the logic required launching native binary applications in distributed heterogeneous environments.

  4. Grid portal-based data management for lattice QCD data

    Energy Technology Data Exchange (ETDEWEB)

    Andronico, G. [Istituto Nazionale di Fisica Nucleare, Sezione di Catania, via S. Sofia 64, 95123 Catania (Italy)]. E-mail: giuseppe.andronico@ct.infn.it; Barbera, R. [Istituto Nazionale di Fisica Nucleare, Sezione di Catania, via S. Sofia 64, 95123 Catania (Italy); Dipartimento di Fisica e Astronomia dell' Universita di Catania, via S. Sofia 64, 95123 Catania (Italy); Falzone, A. [NICE SRL, via Marchesi di Roero 1, 14020 Cortanze (Italy)

    2004-11-21

    We describe here a case of the European Union DataGrid Project data management services by a Lattice Quantum ChromoDynamics (LQCD) application to share the large amount of configurations generated and based on a solution developed from the International Lattice Data Grid Project using a XML dialect called QCDML. In order to allow the user to store, search and browse the lattice configurations described by QCDML in an uniform and transparent way, we have exploited the functionalities of the GENIUS Grid portal, jointly developed by INFN and NICE srl in the context of the Italian INFN Grid and EU DataGrid Projects.

  5. Grid portal-based data management for lattice QCD data

    International Nuclear Information System (INIS)

    Andronico, G.; Barbera, R.; Falzone, A.

    2004-01-01

    We describe here a case of the European Union DataGrid Project data management services by a Lattice Quantum ChromoDynamics (LQCD) application to share the large amount of configurations generated and based on a solution developed from the International Lattice Data Grid Project using a XML dialect called QCDML. In order to allow the user to store, search and browse the lattice configurations described by QCDML in an uniform and transparent way, we have exploited the functionalities of the GENIUS Grid portal, jointly developed by INFN and NICE srl in the context of the Italian INFN Grid and EU DataGrid Projects

  6. McRunjob: A High Energy Physics Workflow Planner for Grid Production Processing

    CERN Document Server

    Graham, G E; Bertram, I; Graham, Gregory E.; Evans, Dave; Bertram, Iain

    2003-01-01

    McRunjob is a powerful grid workflow manager used to manage the generation of large numbers of production processing jobs in High Energy Physics. In use at both the DZero and CMS experiments, McRunjob has been used to manage large Monte Carlo production processing since 1999 and is being extended to uses in regular production processing for analysis and reconstruction. Described at CHEP 2001, McRunjob converts core metadata into jobs submittable in a variety of environments. The powerful core metadata description language includes methods for converting the metadata into persistent forms, job descriptions, multi-step workflows, and data provenance information. The language features allow for structure in the metadata by including full expressions, namespaces, functional dependencies, site specific parameters in a grid environment, and ontological definitions. It also has simple control structures for parallelization of large jobs. McRunjob features a modular design which allows for easy expansion to new job d...

  7. Grist: Grid-based Data Mining for Astronomy

    Science.gov (United States)

    Jacob, J. C.; Katz, D. S.; Miller, C. D.; Walia, H.; Williams, R. D.; Djorgovski, S. G.; Graham, M. J.; Mahabal, A. A.; Babu, G. J.; vanden Berk, D. E.; Nichol, R.

    2005-12-01

    The Grist project is developing a grid-technology based system as a research environment for astronomy with massive and complex datasets. This knowledge extraction system will consist of a library of distributed grid services controlled by a workflow system, compliant with standards emerging from the grid computing, web services, and virtual observatory communities. This new technology is being used to find high redshift quasars, study peculiar variable objects, search for transients in real time, and fit SDSS QSO spectra to measure black hole masses. Grist services are also a component of the ``hyperatlas'' project to serve high-resolution multi-wavelength imagery over the Internet. In support of these science and outreach objectives, the Grist framework will provide the enabling fabric to tie together distributed grid services in the areas of data access, federation, mining, subsetting, source extraction, image mosaicking, statistics, and visualization.

  8. Grist : grid-based data mining for astronomy

    Science.gov (United States)

    Jacob, Joseph C.; Katz, Daniel S.; Miller, Craig D.; Walia, Harshpreet; Williams, Roy; Djorgovski, S. George; Graham, Matthew J.; Mahabal, Ashish; Babu, Jogesh; Berk, Daniel E. Vanden; hide

    2004-01-01

    The Grist project is developing a grid-technology based system as a research environment for astronomy with massive and complex datasets. This knowledge extraction system will consist of a library of distributed grid services controlled by a workflow system, compliant with standards emerging from the grid computing, web services, and virtual observatory communities. This new technology is being used to find high redshift quasars, study peculiar variable objects, search for transients in real time, and fit SDSS QSO spectra to measure black hole masses. Grist services are also a component of the 'hyperatlas' project to serve high-resolution multi-wavelength imagery over the Internet. In support of these science and outreach objectives, the Grist framework will provide the enabling fabric to tie together distributed grid services in the areas of data access, federation, mining, subsetting, source extraction, image mosaicking, statistics, and visualization.

  9. Research on the comparison of extension mechanism of cellular automaton based on hexagon grid and rectangular grid

    Science.gov (United States)

    Zhai, Xiaofang; Zhu, Xinyan; Xiao, Zhifeng; Weng, Jie

    2009-10-01

    Historically, cellular automata (CA) is a discrete dynamical mathematical structure defined on spatial grid. Research on cellular automata system (CAS) has focused on rule sets and initial condition and has not discussed its adjacency. Thus, the main focus of our study is the effect of adjacency on CA behavior. This paper is to compare rectangular grids with hexagonal grids on their characteristics, strengths and weaknesses. They have great influence on modeling effects and other applications including the role of nearest neighborhood in experimental design. Our researches present that rectangular and hexagonal grids have different characteristics. They are adapted to distinct aspects, and the regular rectangular or square grid is used more often than the hexagonal grid. But their relative merits have not been widely discussed. The rectangular grid is generally preferred because of its symmetry, especially in orthogonal co-ordinate system and the frequent use of raster from Geographic Information System (GIS). However, in terms of complex terrain, uncertain and multidirectional region, we have preferred hexagonal grids and methods to facilitate and simplify the problem. Hexagonal grids can overcome directional warp and have some unique characteristics. For example, hexagonal grids have a simpler and more symmetric nearest neighborhood, which avoids the ambiguities of the rectangular grids. Movement paths or connectivity, the most compact arrangement of pixels, make hexagonal appear great dominance in the process of modeling and analysis. The selection of an appropriate grid should be based on the requirements and objectives of the application. We use rectangular and hexagonal grids respectively for developing city model. At the same time we make use of remote sensing images and acquire 2002 and 2005 land state of Wuhan. On the base of city land state in 2002, we make use of CA to simulate reasonable form of city in 2005. Hereby, these results provide a proof of

  10. Two Approaches for the Management of Virtual Machines on Grid Infrastructures

    International Nuclear Information System (INIS)

    Tapiador, D.; Rubio-Montero, A. J.; Juedo, E.; Montero, R. S.; Llorente, I. M.

    2007-01-01

    Virtual machines are a promising technology to overcome some of the problems found in current Grid infrastructures, like heterogeneity, performance partitioning or application isolation. This work shows a comparison between two strategies to manage virtual machines in Globus Grids. The first alternative is a straightforward deployment that does not require additional middle ware to be installed. It is only based on standard Grid services and is not bound to a given virtualization technology. Although this option is fully functional, it is only suitable for single process batch jobs. The second solution makes use of the Virtual Workspace Service which allows a remote client to securely negotiate and manage a virtual resource. This approach better exploits the potential benefits offered by the virtualization technology and provides a wider application range. (Author)

  11. GridPP - Preparing for LHC Run 2 and the Wider Context

    Science.gov (United States)

    Coles, Jeremy

    2015-12-01

    This paper elaborates upon the operational status and directions within the UK Computing for Particle Physics (GridPP) project as it approaches LHC Run 2. It details the pressures that have been gradually reshaping the deployed hardware and middleware environments at GridPP sites - from the increasing adoption of larger multicore nodes to the move towards alternative batch systems and cloud alternatives - as well as changes being driven by funding considerations. The paper highlights work being done with non-LHC communities and describes some of the early outcomes of adopting a generic DIRAC based job submission and management framework. The paper presents results from an analysis of how GridPP effort is distributed across various deployment and operations tasks and how this may be used to target further improvements in efficiency.

  12. GRID-BASED EXPLORATION OF COSMOLOGICAL PARAMETER SPACE WITH SNAKE

    International Nuclear Information System (INIS)

    Mikkelsen, K.; Næss, S. K.; Eriksen, H. K.

    2013-01-01

    We present a fully parallelized grid-based parameter estimation algorithm for investigating multidimensional likelihoods called Snake, and apply it to cosmological parameter estimation. The basic idea is to map out the likelihood grid-cell by grid-cell according to decreasing likelihood, and stop when a certain threshold has been reached. This approach improves vastly on the 'curse of dimensionality' problem plaguing standard grid-based parameter estimation simply by disregarding grid cells with negligible likelihood. The main advantages of this method compared to standard Metropolis-Hastings Markov Chain Monte Carlo methods include (1) trivial extraction of arbitrary conditional distributions; (2) direct access to Bayesian evidences; (3) better sampling of the tails of the distribution; and (4) nearly perfect parallelization scaling. The main disadvantage is, as in the case of brute-force grid-based evaluation, a dependency on the number of parameters, N par . One of the main goals of the present paper is to determine how large N par can be, while still maintaining reasonable computational efficiency; we find that N par = 12 is well within the capabilities of the method. The performance of the code is tested by comparing cosmological parameters estimated using Snake and the WMAP-7 data with those obtained using CosmoMC, the current standard code in the field. We find fully consistent results, with similar computational expenses, but shorter wall time due to the perfect parallelization scheme

  13. Task-and-role-based access-control model for computational grid

    Institute of Scientific and Technical Information of China (English)

    LONG Tao; HONG Fan; WU Chi; SUN Ling-li

    2007-01-01

    Access control in a grid environment is a challenging issue because the heterogeneous nature and independent administration of geographically dispersed resources in grid require access control to use fine-grained policies. We established a task-and-role-based access-control model for computational grid (CG-TRBAC model), integrating the concepts of role-based access control (RBAC) and task-based access control (TBAC). In this model, condition restrictions are defined and concepts specifically tailored to Workflow Management System are simplified or omitted so that role assignment and security administration fit computational grid better than traditional models; permissions are mutable with the task status and system variables, and can be dynamically controlled. The CG-TRBAC model is proved flexible and extendible. It can implement different control policies. It embodies the security principle of least privilege and executes active dynamic authorization. A task attribute can be extended to satisfy different requirements in a real grid system.

  14. Near-Body Grid Adaption for Overset Grids

    Science.gov (United States)

    Buning, Pieter G.; Pulliam, Thomas H.

    2016-01-01

    A solution adaption capability for curvilinear near-body grids has been implemented in the OVERFLOW overset grid computational fluid dynamics code. The approach follows closely that used for the Cartesian off-body grids, but inserts refined grids in the computational space of original near-body grids. Refined curvilinear grids are generated using parametric cubic interpolation, with one-sided biasing based on curvature and stretching ratio of the original grid. Sensor functions, grid marking, and solution interpolation tasks are implemented in the same fashion as for off-body grids. A goal-oriented procedure, based on largest error first, is included for controlling growth rate and maximum size of the adapted grid system. The adaption process is almost entirely parallelized using MPI, resulting in a capability suitable for viscous, moving body simulations. Two- and three-dimensional examples are presented.

  15. Fast and accurate grid representations for atom-based docking with partner flexibility.

    Science.gov (United States)

    de Vries, Sjoerd J; Zacharias, Martin

    2017-06-30

    Macromolecular docking methods can broadly be divided into geometric and atom-based methods. Geometric methods use fast algorithms that operate on simplified, grid-like molecular representations, while atom-based methods are more realistic and flexible, but far less efficient. Here, a hybrid approach of grid-based and atom-based docking is presented, combining precalculated grid potentials with neighbor lists for fast and accurate calculation of atom-based intermolecular energies and forces. The grid representation is compatible with simultaneous multibody docking and can tolerate considerable protein flexibility. When implemented in our docking method ATTRACT, grid-based docking was found to be ∼35x faster. With the OPLSX forcefield instead of the ATTRACT coarse-grained forcefield, the average speed improvement was >100x. Grid-based representations may allow atom-based docking methods to explore large conformational spaces with many degrees of freedom, such as multiple macromolecules including flexibility. This increases the domain of biological problems to which docking methods can be applied. © 2017 Wiley Periodicals, Inc. © 2017 Wiley Periodicals, Inc.

  16. Grid-based electronic structure calculations: The tensor decomposition approach

    Energy Technology Data Exchange (ETDEWEB)

    Rakhuba, M.V., E-mail: rakhuba.m@gmail.com [Skolkovo Institute of Science and Technology, Novaya St. 100, 143025 Skolkovo, Moscow Region (Russian Federation); Oseledets, I.V., E-mail: i.oseledets@skoltech.ru [Skolkovo Institute of Science and Technology, Novaya St. 100, 143025 Skolkovo, Moscow Region (Russian Federation); Institute of Numerical Mathematics, Russian Academy of Sciences, Gubkina St. 8, 119333 Moscow (Russian Federation)

    2016-05-01

    We present a fully grid-based approach for solving Hartree–Fock and all-electron Kohn–Sham equations based on low-rank approximation of three-dimensional electron orbitals. Due to the low-rank structure the total complexity of the algorithm depends linearly with respect to the one-dimensional grid size. Linear complexity allows for the usage of fine grids, e.g. 8192{sup 3} and, thus, cheap extrapolation procedure. We test the proposed approach on closed-shell atoms up to the argon, several molecules and clusters of hydrogen atoms. All tests show systematical convergence with the required accuracy.

  17. Adaptively detecting changes in Autonomic Grid Computing

    KAUST Repository

    Zhang, Xiangliang

    2010-10-01

    Detecting the changes is the common issue in many application fields due to the non-stationary distribution of the applicative data, e.g., sensor network signals, web logs and gridrunning logs. Toward Autonomic Grid Computing, adaptively detecting the changes in a grid system can help to alarm the anomalies, clean the noises, and report the new patterns. In this paper, we proposed an approach of self-adaptive change detection based on the Page-Hinkley statistic test. It handles the non-stationary distribution without the assumption of data distribution and the empirical setting of parameters. We validate the approach on the EGEE streaming jobs, and report its better performance on achieving higher accuracy comparing to the other change detection methods. Meanwhile this change detection process could help to discover the device fault which was not claimed in the system logs. © 2010 IEEE.

  18. ATLAS Job Transforms

    CERN Document Server

    Stewart, G A; The ATLAS collaboration; Maddocks, H J; Harenberg, T; Sandhoff, M; Sarrazin, B

    2013-01-01

    The need to run complex workflows for a high energy physics experiment such as ATLAS has always been present. However, as computing resources have become even more constrained, compared to the wealth of data generated by the LHC, the need to use resources efficiently and manage complex workflows within a single grid job have increased. In ATLAS, a new Job Transform framework has been developed that we describe in this paper. This framework manages the multiple execution steps needed to `transform' one data type into another (e.g., RAW data to ESD to AOD to final ntuple) and also provides a consistent interface for the ATLAS production system. The new framework uses a data driven workflow definition which is both easy to manage and powerful. After a transform is defined, jobs are expressed simply by specifying the input data and the desired output data. The transform infrastructure then executes only the necessary substeps to produce the final data products. The global execution cost of running the job is mini...

  19. ATLAS Job Transforms

    CERN Document Server

    Stewart, G A; The ATLAS collaboration; Maddocks, H J; Harenberg, T; Sandhoff, M; Sarrazin, B

    2013-01-01

    The need to run complex workflows for a high energy physics experiment such as ATLAS has always been present. However, as computing resources have become even more constrained, compared to the wealth of data generated by the LHC, the need to use resources efficiently and manage complex workflows within a single grid job have increased. In ATLAS, a new Job Transform framework has been developed that we describe in this paper. This framework manages the multiple execution steps needed to 'transform' one data type into another (e.g., RAW data to ESD to AOD to final ntuple) and also provides a consistent interface for the ATLAS production system. The new framework uses a data driven workflow definition which is both easy to manage and powerful. After a transform is defined, jobs are expressed simply by specifying the input data and the desired output data. The transform infrastructure then executes only the necessary substeps to produce the final data products. The global execution cost of running the job is mini...

  20. GRID-BASED EXPLORATION OF COSMOLOGICAL PARAMETER SPACE WITH SNAKE

    Energy Technology Data Exchange (ETDEWEB)

    Mikkelsen, K.; Næss, S. K.; Eriksen, H. K., E-mail: kristin.mikkelsen@astro.uio.no [Institute of Theoretical Astrophysics, University of Oslo, P.O. Box 1029, Blindern, NO-0315 Oslo (Norway)

    2013-11-10

    We present a fully parallelized grid-based parameter estimation algorithm for investigating multidimensional likelihoods called Snake, and apply it to cosmological parameter estimation. The basic idea is to map out the likelihood grid-cell by grid-cell according to decreasing likelihood, and stop when a certain threshold has been reached. This approach improves vastly on the 'curse of dimensionality' problem plaguing standard grid-based parameter estimation simply by disregarding grid cells with negligible likelihood. The main advantages of this method compared to standard Metropolis-Hastings Markov Chain Monte Carlo methods include (1) trivial extraction of arbitrary conditional distributions; (2) direct access to Bayesian evidences; (3) better sampling of the tails of the distribution; and (4) nearly perfect parallelization scaling. The main disadvantage is, as in the case of brute-force grid-based evaluation, a dependency on the number of parameters, N{sub par}. One of the main goals of the present paper is to determine how large N{sub par} can be, while still maintaining reasonable computational efficiency; we find that N{sub par} = 12 is well within the capabilities of the method. The performance of the code is tested by comparing cosmological parameters estimated using Snake and the WMAP-7 data with those obtained using CosmoMC, the current standard code in the field. We find fully consistent results, with similar computational expenses, but shorter wall time due to the perfect parallelization scheme.

  1. Scheduling Non-Preemptible Jobs to Minimize Peak Demand

    Directory of Open Access Journals (Sweden)

    Sean Yaw

    2017-10-01

    Full Text Available This paper examines an important problem in smart grid energy scheduling; peaks in power demand are proportionally more expensive to generate and provision for. The issue is exacerbated in local microgrids that do not benefit from the aggregate smoothing experienced by large grids. Demand-side scheduling can reduce these peaks by taking advantage of the fact that there is often flexibility in job start times. We focus attention on the case where the jobs are non-preemptible, meaning once started, they run to completion. The associated optimization problem is called the peak demand minimization problem, and has been previously shown to be NP-hard. Our results include an optimal fixed-parameter tractable algorithm, a polynomial-time approximation algorithm, as well as an effective heuristic that can also be used in an online setting of the problem. Simulation results show that these methods can reduce peak demand by up to 50% versus on-demand scheduling for household power jobs.

  2. Wind Farm Grid Integration Using VSC Based HVDC Transmission - An Overview

    DEFF Research Database (Denmark)

    Chaudhary, Sanjay Kumar; Teodorescu, Remus; Rodriguez, Pedro

    2008-01-01

    The paper gives an overview of HVAC and HVDC connection of wind farm to the grid, with an emphasis on Voltage Source Converter (VSC)-based HVDC for large wind farms requiring long distance cable connection. Flexible control capabilities of a VSC-based HVDC system enables smooth integration of wind...... farm into the power grid network while meeting the Grid Code Requirements (GCR). Operation of a wind farm with VSC-based HVDC connection is described....

  3. Automatic Integration Testbeds validation on Open Science Grid

    International Nuclear Information System (INIS)

    Caballero, J; Potekhin, M; Thapa, S; Gardner, R

    2011-01-01

    A recurring challenge in deploying high quality production middleware is the extent to which realistic testing occurs before release of the software into the production environment. We describe here an automated system for validating releases of the Open Science Grid software stack that leverages the (pilot-based) PanDA job management system developed and used by the ATLAS experiment. The system was motivated by a desire to subject the OSG Integration Testbed to more realistic validation tests. In particular those which resemble to every extent possible actual job workflows used by the experiments thus utilizing job scheduling at the compute element (CE), use of the worker node execution environment, transfer of data to/from the local storage element (SE), etc. The context is that candidate releases of OSG compute and storage elements can be tested by injecting large numbers of synthetic jobs varying in complexity and coverage of services tested. The native capabilities of the PanDA system can thus be used to define jobs, monitor their execution, and archive the resulting run statistics including success and failure modes. A repository of generic workflows and job types to measure various metrics of interest has been created. A command-line toolset has been developed so that testbed managers can quickly submit 'VO-like' jobs into the system when newly deployed services are ready for testing. A system for automatic submission has been crafted to send jobs to integration testbed sites, collecting the results in a central service and generating regular reports for performance and reliability.

  4. Automatic Integration Testbeds validation on Open Science Grid

    Science.gov (United States)

    Caballero, J.; Thapa, S.; Gardner, R.; Potekhin, M.

    2011-12-01

    A recurring challenge in deploying high quality production middleware is the extent to which realistic testing occurs before release of the software into the production environment. We describe here an automated system for validating releases of the Open Science Grid software stack that leverages the (pilot-based) PanDA job management system developed and used by the ATLAS experiment. The system was motivated by a desire to subject the OSG Integration Testbed to more realistic validation tests. In particular those which resemble to every extent possible actual job workflows used by the experiments thus utilizing job scheduling at the compute element (CE), use of the worker node execution environment, transfer of data to/from the local storage element (SE), etc. The context is that candidate releases of OSG compute and storage elements can be tested by injecting large numbers of synthetic jobs varying in complexity and coverage of services tested. The native capabilities of the PanDA system can thus be used to define jobs, monitor their execution, and archive the resulting run statistics including success and failure modes. A repository of generic workflows and job types to measure various metrics of interest has been created. A command-line toolset has been developed so that testbed managers can quickly submit "VO-like" jobs into the system when newly deployed services are ready for testing. A system for automatic submission has been crafted to send jobs to integration testbed sites, collecting the results in a central service and generating regular reports for performance and reliability.

  5. A data grid for imaging-based clinical trials

    Science.gov (United States)

    Zhou, Zheng; Chao, Sander S.; Lee, Jasper; Liu, Brent; Documet, Jorge; Huang, H. K.

    2007-03-01

    Clinical trials play a crucial role in testing new drugs or devices in modern medicine. Medical imaging has also become an important tool in clinical trials because images provide a unique and fast diagnosis with visual observation and quantitative assessment. A typical imaging-based clinical trial consists of: 1) A well-defined rigorous clinical trial protocol, 2) a radiology core that has a quality control mechanism, a biostatistics component, and a server for storing and distributing data and analysis results; and 3) many field sites that generate and send image studies to the radiology core. As the number of clinical trials increases, it becomes a challenge for a radiology core servicing multiple trials to have a server robust enough to administrate and quickly distribute information to participating radiologists/clinicians worldwide. The Data Grid can satisfy the aforementioned requirements of imaging based clinical trials. In this paper, we present a Data Grid architecture for imaging-based clinical trials. A Data Grid prototype has been implemented in the Image Processing and Informatics (IPI) Laboratory at the University of Southern California to test and evaluate performance in storing trial images and analysis results for a clinical trial. The implementation methodology and evaluation protocol of the Data Grid are presented.

  6. State-space-based harmonic stability analysis for paralleled grid-connected inverters

    DEFF Research Database (Denmark)

    Wang, Yanbo; Wang, Xiongfei; Chen, Zhe

    2016-01-01

    This paper addresses a state-space-based harmonic stability analysis of paralleled grid-connected inverters system. A small signal model of individual inverter is developed, where LCL filter, the equivalent delay of control system, and current controller are modeled. Then, the overall small signal...... model of paralleled grid-connected inverters is built. Finally, the state space-based stability analysis approach is developed to explain the harmonic resonance phenomenon. The eigenvalue traces associated with time delay and coupled grid impedance are obtained, which accounts for how the unstable...... inverter produces the harmonic resonance and leads to the instability of whole paralleled system. The proposed approach reveals the contributions of the grid impedance as well as the coupled effect on other grid-connected inverters under different grid conditions. Simulation and experimental results...

  7. Wind power integration in island-based smart grid projects : A comparative study between Jeju Smart Grid Test-bed and Smart Grid Gotland

    OpenAIRE

    Piehl, Hampus

    2014-01-01

    Smart grids seem to be the solution to use energy from renewable and intermittent energy sources in an efficient manner. There are many research projects around the world and two of them are Jeju Smart Grid Test-bed and Smart Grid Gotland. They have in common that they are both island-based projects and connected to the Powergrid on the mainland by HVDC-link. The purpose of this thesis is to compare the two projects and find out what challenges and strategies they have related to wind power i...

  8. Optimising LAN access to grid enabled storage elements

    International Nuclear Information System (INIS)

    Stewart, G A; Dunne, B; Elwell, A; Millar, A P; Cowan, G A

    2008-01-01

    When operational, the Large Hadron Collider experiments at CERN will collect tens of petabytes of physics data per year. The worldwide LHC computing grid (WLCG) will distribute this data to over two hundred Tier-1 and Tier-2 computing centres, enabling particle physicists around the globe to access the data for analysis. Although different middleware solutions exist for effective management of storage systems at collaborating institutes, the patterns of access envisaged for Tier-2s fall into two distinct categories. The first involves bulk transfer of data between different Grid storage elements using protocols such as GridFTP. This data movement will principally involve writing ESD and AOD files into Tier-2 storage. Secondly, once datasets are stored at a Tier-2, physics analysis jobs will read the data from the local SE. Such jobs require a POSIX-like interface to the storage so that individual physics events can be extracted. In this paper we consider the performance of POSIX-like access to files held in Disk Pool Manager (DPM) storage elements, a popular lightweight SRM storage manager from EGEE

  9. Frequency scanning-based stability analysis method for grid-connected inverter system

    DEFF Research Database (Denmark)

    Wang, Yanbo; Wang, Xiongfei; Blaabjerg, Frede

    2017-01-01

    This paper proposes a frequency scanning-based impedance analysis for stability assessment of grid-connected inverter system, which is able to perform stability assessment without using system mathematical models and inherit the superior feature of impedance-based stability criterion with conside......This paper proposes a frequency scanning-based impedance analysis for stability assessment of grid-connected inverter system, which is able to perform stability assessment without using system mathematical models and inherit the superior feature of impedance-based stability criterion...... with consideration of the inverter nonlinearities. Small current disturbance is injected into grid-connected inverter system in a particular frequency range, and the impedance is computed according to the harmonic-frequency response using Fourier analysis, and then the stability is predicted on the basis...... of the impedance stability criterion. The stability issues of grid-connected inverters with grid-current feedback and the converter-current feedback are addressed using the proposed method. The results obtained from simulation and experiments validate the effectiveness of the method. The frequency scanning...

  10. Reinforcement Learning Based Novel Adaptive Learning Framework for Smart Grid Prediction

    Directory of Open Access Journals (Sweden)

    Tian Li

    2017-01-01

    Full Text Available Smart grid is a potential infrastructure to supply electricity demand for end users in a safe and reliable manner. With the rapid increase of the share of renewable energy and controllable loads in smart grid, the operation uncertainty of smart grid has increased briskly during recent years. The forecast is responsible for the safety and economic operation of the smart grid. However, most existing forecast methods cannot account for the smart grid due to the disabilities to adapt to the varying operational conditions. In this paper, reinforcement learning is firstly exploited to develop an online learning framework for the smart grid. With the capability of multitime scale resolution, wavelet neural network has been adopted in the online learning framework to yield reinforcement learning and wavelet neural network (RLWNN based adaptive learning scheme. The simulations on two typical prediction problems in smart grid, including wind power prediction and load forecast, validate the effectiveness and the scalability of the proposed RLWNN based learning framework and algorithm.

  11. AVQS: Attack Route-Based Vulnerability Quantification Scheme for Smart Grid

    Directory of Open Access Journals (Sweden)

    Jongbin Ko

    2014-01-01

    Full Text Available A smart grid is a large, consolidated electrical grid system that includes heterogeneous networks and systems. Based on the data, a smart grid system has a potential security threat in its network connectivity. To solve this problem, we develop and apply a novel scheme to measure the vulnerability in a smart grid domain. Vulnerability quantification can be the first step in security analysis because it can help prioritize the security problems. However, existing vulnerability quantification schemes are not suitable for smart grid because they do not consider network vulnerabilities. We propose a novel attack route-based vulnerability quantification scheme using a network vulnerability score and an end-to-end security score, depending on the specific smart grid network environment to calculate the vulnerability score for a particular attack route. To evaluate the proposed approach, we derive several attack scenarios from the advanced metering infrastructure domain. The experimental results of the proposed approach and the existing common vulnerability scoring system clearly show that we need to consider network connectivity for more optimized vulnerability quantification.

  12. AVQS: attack route-based vulnerability quantification scheme for smart grid.

    Science.gov (United States)

    Ko, Jongbin; Lim, Hyunwoo; Lee, Seokjun; Shon, Taeshik

    2014-01-01

    A smart grid is a large, consolidated electrical grid system that includes heterogeneous networks and systems. Based on the data, a smart grid system has a potential security threat in its network connectivity. To solve this problem, we develop and apply a novel scheme to measure the vulnerability in a smart grid domain. Vulnerability quantification can be the first step in security analysis because it can help prioritize the security problems. However, existing vulnerability quantification schemes are not suitable for smart grid because they do not consider network vulnerabilities. We propose a novel attack route-based vulnerability quantification scheme using a network vulnerability score and an end-to-end security score, depending on the specific smart grid network environment to calculate the vulnerability score for a particular attack route. To evaluate the proposed approach, we derive several attack scenarios from the advanced metering infrastructure domain. The experimental results of the proposed approach and the existing common vulnerability scoring system clearly show that we need to consider network connectivity for more optimized vulnerability quantification.

  13. Schwarz-Christoffel Conformal Mapping based Grid Generation for Global Oceanic Circulation Models

    Science.gov (United States)

    Xu, Shiming

    2015-04-01

    We propose new grid generation algorithms for global ocean general circulation models (OGCMs). Contrary to conventional, analytical forms based dipolar or tripolar grids, the new algorithm are based on Schwarz-Christoffel (SC) conformal mapping with prescribed boundary information. While dealing with the conventional grid design problem of pole relocation, it also addresses more advanced issues of computational efficiency and the new requirements on OGCM grids arisen from the recent trend of high-resolution and multi-scale modeling. The proposed grid generation algorithm could potentially achieve the alignment of grid lines to coastlines, enhanced spatial resolution in coastal regions, and easier computational load balance. Since the generated grids are still orthogonal curvilinear, they can be readily 10 utilized in existing Bryan-Cox-Semtner type ocean models. The proposed methodology can also be applied to the grid generation task for regional ocean modeling when complex land-ocean distribution is present.

  14. Socioeconomic assessment of smart grids. Summary

    International Nuclear Information System (INIS)

    2015-07-01

    In September of 2013, the President of France identified smart grids as an important part of the country's industrial strategy, given the opportunities and advantages they can offer French industry, and asked the Chairman of the RTE Management Board to prepare a road-map outlining ways to support and accelerate smart grid development. This road-map, prepared in cooperation with stakeholders from the power and smart grids industries, identifies ten actions that can be taken in priority to consolidate the smart grids sector and help French firms play a leading role in the segment. These priorities were presented to the President of France on 7 May 2014. Action items 5 and 6 of the road-map on smart grid development relate, respectively, to the quantification of the value of smart grid functions from an economic, environmental and social (impact on employment) standpoint and to the large-scale deployment of some of the functions. Two tasks were set out in the 'Smart Grids' plan for action item 5: - Create a methodological framework that, for all advanced functions, allows the quantification of benefits and costs from an economic, environmental and social (effect on jobs) standpoint; - Quantify, based on this methodological framework, the potential benefits of a set of smart grid functions considered sufficiently mature to be deployed on a large scale in the near future. Having a methodology that can be applied in the same manner to all solutions, taking into account their impacts on the environment and employment in France, will considerably add to and complement the information drawn from demonstration projects. It will notably enable comparisons of benefits provided by smart grid functions and thus help give rise to a French smart grids industry that is competitive. At first, the smart grids industry was organised around demonstration projects testing different advanced functions within specific geographic areas. These projects covered a wide enough

  15. Socioeconomic assessment of smart grids - Summary

    International Nuclear Information System (INIS)

    Janssen, Tanguy

    2015-07-01

    In September of 2013, the President of France identified smart grids as an important part of the country's industrial strategy, given the opportunities and advantages they can offer French industry, and asked the Chairman of the RTE Management Board to prepare a road-map outlining ways to support and accelerate smart grid development. This road-map, prepared in cooperation with stakeholders from the power and smart grids industries, identifies ten actions that can be taken in priority to consolidate the smart grids sector and help French firms play a leading role in the segment. These priorities were presented to the President of France on 7 May 2014. Action items 5 and 6 of the road-map on smart grid development relate, respectively, to the quantification of the value of smart grid functions from an economic, environmental and social (impact on employment) standpoint and to the large-scale deployment of some of the functions. Two tasks were set out in the 'Smart Grids' plan for action item 5: - Create a methodological framework that, for all advanced functions, allows the quantification of benefits and costs from an economic, environmental and social (effect on jobs) standpoint; - Quantify, based on this methodological framework, the potential benefits of a set of smart grid functions considered sufficiently mature to be deployed on a large scale in the near future. Having a methodology that can be applied in the same manner to all solutions, taking into account their impacts on the environment and employment in France, will considerably add to and complement the information drawn from demonstration projects. It will notably enable comparisons of benefits provided by smart grid functions and thus help give rise to a French smart grids industry that is competitive. At first, the smart grids industry was organised around demonstration projects testing different advanced functions within specific geographic areas. These projects covered a wide enough

  16. The relationships among nurses' job characteristics and attitudes toward web-based continuing learning.

    Science.gov (United States)

    Chiu, Yen-Lin; Tsai, Chin-Chung; Fan Chiang, Chih-Yun

    2013-04-01

    The purpose of this study was to explore the relationships between job characteristics (job demands, job control and social support) and nurses' attitudes toward web-based continuing learning. A total of 221 in-service nurses from hospitals in Taiwan were surveyed. The Attitudes toward Web-based Continuing Learning Survey (AWCL) was employed as the outcome variables, and the Chinese version Job Characteristic Questionnaire (C-JCQ) was administered to assess the predictors for explaining the nurses' attitudes toward web-based continuing learning. To examine the relationships among these variables, hierarchical regression was conducted. The results of the regression analysis revealed that job control and social support positively associated with nurses' attitudes toward web-based continuing learning. However, the relationship of job demands to such learning was not significant. Moreover, a significant demands×job control interaction was found, but the job demands×social support interaction had no significant relationships with attitudes toward web-based continuing learning. Copyright © 2013 Elsevier Ltd. All rights reserved.

  17. Reliability Engineering for ATLAS Petascale Data Processing on the Grid

    CERN Document Server

    Golubkov, D V; The ATLAS collaboration; Vaniachine, A V

    2012-01-01

    The ATLAS detector is in its third year of continuous LHC running taking data for physics analysis. A starting point for ATLAS physics analysis is reconstruction of the raw data. First-pass processing takes place shortly after data taking, followed later by reprocessing of the raw data with updated software and calibrations to improve the quality of the reconstructed data for physics analysis. Data reprocessing involves a significant commitment of computing resources and is conducted on the Grid. The reconstruction of one petabyte of ATLAS data with 1B collision events from the LHC takes about three million core-hours. Petascale data processing on the Grid involves millions of data processing jobs. At such scales, the reprocessing must handle a continuous stream of failures. Automatic job resubmission recovers transient failures at the cost of CPU time used by the failed jobs. Orchestrating ATLAS data processing applications to ensure efficient usage of tens of thousands of CPU-cores, reliability engineering ...

  18. dSPACE based adaptive neuro-fuzzy controller of grid interactive inverter

    International Nuclear Information System (INIS)

    Altin, Necmi; Sefa, İbrahim

    2012-01-01

    Highlights: ► We propose a dSPACE based neuro-fuzzy controlled grid interactive inverter. ► The membership functions and rule base of fuzzy logic controller by using ANFIS. ► A LCL output filter is designed. ► A high performance controller is designed. - Abstract: In this study, design, simulation and implementation of a dSPACE based grid interactive voltage source inverter are proposed. This inverter has adaptive neuro-fuzzy controller and capable of importing electrical energy, generated from renewable energy sources such as the wind, the solar and the fuel cells to the grid. A line frequency transformer and a LCL filter are used at the output of the grid interactive inverter which is designed as current controlled to decrease the susceptibility to phase errors. Membership functions and rule base of the fuzzy logic controller, which control the inverter output current, are determined by using artificial neural networks. Both simulation and experimental results show that, the grid interactive inverter operates synchronously with the grid. The inverter output current which is imported to the grid is in sinusoidal waveform and the harmonic level of it meets the international standards (4.3 < 5.0%). In addition, simulation and experimental results of the neuro-fuzzy and the PI controlled inverter are given together and compared in detail. Simulation and experimental results show that the proposed inverter has faster response to the reference variations and lower steady state error than PI controller.

  19. Experience in Grid Site Testing for ATLAS, CMS and LHCb with HammerCloud

    International Nuclear Information System (INIS)

    Elmsheuser, Johannes; Legger, Federica; Llamas, Ramón Medrano; Sciabà, Andrea; García, Mario Úbeda; Ster, Daniel van der; Sciacca, Gianfranco

    2012-01-01

    Frequent validation and stress testing of the network, storage and CPU resources of a grid site is essential to achieve high performance and reliability. HammerCloud was previously introduced with the goals of enabling VO- and site-administrators to run such tests in an automated or on-demand manner. The ATLAS, CMS and LHCb experiments have all developed VO plugins for the service and have successfully integrated it into their grid operations infrastructures. This work will present the experience in running HammerCloud at full scale for more than 3 years and present solutions to the scalability issues faced by the service. First, we will show the particular challenges faced when integrating with CMS and LHCb offline computing, including customized dashboards to show site validation reports for the VOs and a new API to tightly integrate with the LHCbDIRAC Resource Status System. Next, a study of the automatic site exclusion component used by ATLAS will be presented along with results for tuning the exclusion policies. A study of the historical test results for ATLAS, CMS and LHCb will be presented, including comparisons between the experiments’ grid availabilities and a search for site-based or temporal failure correlations. Finally, we will look to future plans that will allow users to gain new insights into the test results; these include developments to allow increased testing concurrency, increased scale in the number of metrics recorded per test job (up to hundreds), and increased scale in the historical job information (up to many millions of jobs per VO).

  20. Experience in Grid Site Testing for ATLAS, CMS and LHCb with HammerCloud

    Science.gov (United States)

    Elmsheuser, Johannes; Medrano Llamas, Ramón; Legger, Federica; Sciabà, Andrea; Sciacca, Gianfranco; Úbeda García, Mario; van der Ster, Daniel

    2012-12-01

    Frequent validation and stress testing of the network, storage and CPU resources of a grid site is essential to achieve high performance and reliability. HammerCloud was previously introduced with the goals of enabling VO- and site-administrators to run such tests in an automated or on-demand manner. The ATLAS, CMS and LHCb experiments have all developed VO plugins for the service and have successfully integrated it into their grid operations infrastructures. This work will present the experience in running HammerCloud at full scale for more than 3 years and present solutions to the scalability issues faced by the service. First, we will show the particular challenges faced when integrating with CMS and LHCb offline computing, including customized dashboards to show site validation reports for the VOs and a new API to tightly integrate with the LHCbDIRAC Resource Status System. Next, a study of the automatic site exclusion component used by ATLAS will be presented along with results for tuning the exclusion policies. A study of the historical test results for ATLAS, CMS and LHCb will be presented, including comparisons between the experiments’ grid availabilities and a search for site-based or temporal failure correlations. Finally, we will look to future plans that will allow users to gain new insights into the test results; these include developments to allow increased testing concurrency, increased scale in the number of metrics recorded per test job (up to hundreds), and increased scale in the historical job information (up to many millions of jobs per VO).

  1. High performance workflow implementation for protein surface characterization using grid technology

    Directory of Open Access Journals (Sweden)

    Clematis Andrea

    2005-12-01

    Full Text Available Abstract Background This study concerns the development of a high performance workflow that, using grid technology, correlates different kinds of Bioinformatics data, starting from the base pairs of the nucleotide sequence to the exposed residues of the protein surface. The implementation of this workflow is based on the Italian Grid.it project infrastructure, that is a network of several computational resources and storage facilities distributed at different grid sites. Methods Workflows are very common in Bioinformatics because they allow to process large quantities of data by delegating the management of resources to the information streaming. Grid technology optimizes the computational load during the different workflow steps, dividing the more expensive tasks into a set of small jobs. Results Grid technology allows efficient database management, a crucial problem for obtaining good results in Bioinformatics applications. The proposed workflow is implemented to integrate huge amounts of data and the results themselves must be stored into a relational database, which results as the added value to the global knowledge. Conclusion A web interface has been developed to make this technology accessible to grid users. Once the workflow has started, by means of the simplified interface, it is possible to follow all the different steps throughout the data processing. Eventually, when the workflow has been terminated, the different features of the protein, like the amino acids exposed on the protein surface, can be compared with the data present in the output database.

  2. Hybrid method based on embedded coupled simulation of vortex particles in grid based solution

    Science.gov (United States)

    Kornev, Nikolai

    2017-09-01

    The paper presents a novel hybrid approach developed to improve the resolution of concentrated vortices in computational fluid mechanics. The method is based on combination of a grid based and the grid free computational vortex (CVM) methods. The large scale flow structures are simulated on the grid whereas the concentrated structures are modeled using CVM. Due to this combination the advantages of both methods are strengthened whereas the disadvantages are diminished. The procedure of the separation of small concentrated vortices from the large scale ones is based on LES filtering idea. The flow dynamics is governed by two coupled transport equations taking two-way interaction between large and fine structures into account. The fine structures are mapped back to the grid if their size grows due to diffusion. Algorithmic aspects of the hybrid method are discussed. Advantages of the new approach are illustrated on some simple two dimensional canonical flows containing concentrated vortices.

  3. A scalable architecture for online anomaly detection of WLCG batch jobs

    Science.gov (United States)

    Kuehn, E.; Fischer, M.; Giffels, M.; Jung, C.; Petzold, A.

    2016-10-01

    For data centres it is increasingly important to monitor the network usage, and learn from network usage patterns. Especially configuration issues or misbehaving batch jobs preventing a smooth operation need to be detected as early as possible. At the GridKa data and computing centre we therefore operate a tool BPNetMon for monitoring traffic data and characteristics of WLCG batch jobs and pilots locally on different worker nodes. On the one hand local information itself are not sufficient to detect anomalies for several reasons, e.g. the underlying job distribution on a single worker node might change or there might be a local misconfiguration. On the other hand a centralised anomaly detection approach does not scale regarding network communication as well as computational costs. We therefore propose a scalable architecture based on concepts of a super-peer network.

  4. Agent based approach for engineering and control of micro-grids

    International Nuclear Information System (INIS)

    Basso, Gillian

    2013-01-01

    Energy management is, nowadays, a subject of uttermost importance. Indeed, we are facing growing concerns such as petroleum reserve depletion, earth global warming or power quality (e.g. avoiding blackouts during peak times). Smart grids is an attempt to solve such problems, by adding to power grids bidirectional communications and ICT capabilities in order to provide an intelligent autonomic management for the grid. This thesis focuses on the management of micro-grids thanks to multi-agent systems (MAS). Micro-grids are low-power networks, composed of small and decentralized energy producers (possibly renewable) and consumers. These networks can be connected to the main grid or islanded, this make them more complex. Due to their complexity and their geographical distribution, smart grids and micro-grids can not be easily managed by a centralized system. Distributed artificial intelligences especially MAS appear to be a solution to resolve problems related to smart grids. Firstly we defined an approach implementing feedback loops. These feedback loops exist in complex systems which can be defined with several abstraction levels. Two levels are interacting. The micro-level contains a set of agents owning behaviours that can be combined. The result of the combination impact the state of the system. The macro-level processes these influences to define a new state of the system which will impact the agents behaviours at the micro-level. This feedback loop separates behaviours on several levels. This approach is used to defined a demand and supply matching problem in micro-grid. This problem afford to manage a set of goals which currently are independently processed. Finally, an application is developed using MAS that ensures grid stability thanks to storage systems. This application was thought to be integrated to the approach detailed above. Secondly, a grid simulator id developed. This simulator allows dynamic control of devices. It is based on three main principles

  5. The Rise of Market-Based Job Search Institutions and Job Niches for Low-Skilled Chinese Immigrants

    Directory of Open Access Journals (Sweden)

    Zai Liang

    2018-01-01

    Full Text Available Increasingly, market-based job search institutions, such as employment agencies and ethnic media, are playing a more important role than migrant networks for low-skilled Chinese immigrants searching for jobs. We argue that two major factors are driving this trend: the diversification of Chinese immigrants’ provinces of origin, and the spatial diffusion of businesses in the United States owned by Chinese immigrants. We also identify some new niche jobs for Chinese immigrants and assess the extent to which this development is driven by China’s growing prosperity. We use data from multiple sources, including a survey of employment agencies in Manhattan’s Chinatown, job advertisements in Chinese-language newspapers, and information on Chinese immigrant hometown associations in the United States.

  6. Optimisation of LHCb Applications for Multi- and Manycore Job Submission

    CERN Document Server

    Rauschmayr, Nathalie; Graciani Diaz, Ricardo; Charpentier, Philippe

    The Worldwide LHC Computing Grid (WLCG) is the largest Computing Grid and is used by all Large Hadron Collider experiments in order to process their recorded data. It provides approximately 400k cores and storages. Nowadays, most of the resources consist of multi- and manycore processors. Conditions at the Large Hadron Collider experiments will change and much larger workloads and jobs consuming more memory are expected in future. This has lead to a shift of paradigm which focuses on executing jobs as multiprocessor tasks in order to use multi- and manycore processors more efficiently. All experiments at CERN are currently investigating how such computing resources can be used more efficiently in terms of memory requirements and handling of concurrency. Until now, there are still many unsolved issues regarding software, scheduling, CPU accounting, task queues, which need to be solved by grid sites and experiments. This thesis develops a systematic approach to optimise the software of the LHCb experiment fo...

  7. What could make 2017 the banner year for smart grids?

    International Nuclear Information System (INIS)

    Ortega, Florian

    2015-01-01

    With billings slated to reach 6 euros billion per year by 2020, intelligent networks, known as smart grids are an attractive proposition for many companies and will generate up to 25 000 jobs directly in France. While it seems, in light of all the commitments that have been made, that 2017 can considered as 'the year of the smart grids', there remain a number of uncertainties. (author)

  8. Soil Erosion Estimation Using Grid-based Computation

    Directory of Open Access Journals (Sweden)

    Josef Vlasák

    2005-06-01

    Full Text Available Soil erosion estimation is an important part of a land consolidation process. Universal soil loss equation (USLE was presented by Wischmeier and Smith. USLE computation uses several factors, namely R – rainfall factor, K – soil erodability, L – slope length factor, S – slope gradient factor, C – cropping management factor, and P – erosion control management factor. L and S factors are usually combined to one LS factor – Topographic factor. The single factors are determined from several sources, such as DTM (Digital Terrain Model, BPEJ – soil type map, aerial and satellite images, etc. A conventional approach to the USLE computation, which is widely used in the Czech Republic, is based on the selection of characteristic profiles for which all above-mentioned factors must be determined. The result (G – annual soil loss of such computation is then applied for a whole area (slope of interest. Another approach to the USLE computation uses grids as a main data-structure. A prerequisite for a grid-based USLE computation is that each of the above-mentioned factors exists as a separate grid layer. The crucial step in this computation is a selection of appropriate grid resolution (grid cell size. A large cell size can cause an undesirable precision degradation. Too small cell size can noticeably slow down the whole computation. Provided that the cell size is derived from the source’s precision, the appropriate cell size for the Czech Republic varies from 30m to 50m. In some cases, especially when new surveying was done, grid computations can be performed with higher accuracy, i.e. with a smaller grid cell size. In such case, we have proposed a new method using the two-step computation. The first step computation uses a bigger cell size and is designed to identify higher erosion spots. The second step then uses a smaller cell size but it make the computation only the area identified in the previous step. This decomposition allows a

  9. Storage element performance optimization for CMS analysis jobs

    International Nuclear Information System (INIS)

    Behrmann, G; Dahlblom, J; Guldmyr, J; Happonen, K; Lindén, T

    2012-01-01

    Tier-2 computing sites in the Worldwide Large Hadron Collider Computing Grid (WLCG) host CPU-resources (Compute Element, CE) and storage resources (Storage Element, SE). The vast amount of data that needs to processed from the Large Hadron Collider (LHC) experiments requires good and efficient use of the available resources. Having a good CPU efficiency for the end users analysis jobs requires that the performance of the storage system is able to scale with I/O requests from hundreds or even thousands of simultaneous jobs. In this presentation we report on the work on improving the SE performance at the Helsinki Institute of Physics (HIP) Tier-2 used for the Compact Muon Experiment (CMS) at the LHC. Statistics from CMS grid jobs are collected and stored in the CMS Dashboard for further analysis, which allows for easy performance monitoring by the sites and by the CMS collaboration. As part of the monitoring framework CMS uses the JobRobot which sends every four hours 100 analysis jobs to each site. CMS also uses the HammerCloud tool for site monitoring and stress testing and it has replaced the JobRobot. The performance of the analysis workflow submitted with JobRobot or HammerCloud can be used to track the performance due to site configuration changes, since the analysis workflow is kept the same for all sites and for months in time. The CPU efficiency of the JobRobot jobs at HIP was increased approximately by 50 % to more than 90 %, by tuning the SE and by improvements in the CMSSW and dCache software. The performance of the CMS analysis jobs improved significantly too. Similar work has been done on other CMS Tier-sites, since on average the CPU efficiency for CMSSW jobs has increased during 2011. Better monitoring of the SE allows faster detection of problems, so that the performance level can be kept high. The next storage upgrade at HIP consists of SAS disk enclosures which can be stress tested on demand with HammerCloud workflows, to make sure that the I

  10. A population-based job exposure matrix for power-frequency magnetic fields.

    Science.gov (United States)

    Bowman, Joseph D; Touchstone, Jennifer A; Yost, Michael G

    2007-09-01

    A population-based job exposure matrix (JEM) was developed to assess personal exposures to power-frequency magnetic fields (MF) for epidemiologic studies. The JEM compiled 2,317 MF measurements taken on or near workers by 10 studies in the United States, Sweden, New Zealand, Finland, and Italy. A database was assembled from the original data for six studies plus summary statistics grouped by occupation from four other published studies. The job descriptions were coded into the 1980 Standard Occupational Classification system (SOC) and then translated to the 1980 job categories of the U.S. Bureau of the Census (BOC). For each job category, the JEM database calculated the arithmetic mean, standard deviation, geometric mean, and geometric standard deviation of the workday-average MF magnitude from the combined data. Analysis of variance demonstrated that the combining of MF data from the different sources was justified, and that the homogeneity of MF exposures in the SOC occupations was comparable to JEMs for solvents and particulates. BOC occupation accounted for 30% of the MF variance (p job variance to the total of within- and between-job variances) was 88%. Jobs lacking data had their exposures inferred from measurements on similar occupations. The JEM provided MF exposures for 97% of the person-months in a population-based case-control study and 95% of the jobs on death certificates in a registry study covering 22 states. Therefore, we expect this JEM to be useful in other population-based epidemiologic studies.

  11. Real Time Load Optimisation of Cable Based Transmission Grids

    DEFF Research Database (Denmark)

    Olsen, Rasmus Schmidt; Holbøll, Joachim; Guðmundsdottir, Unnur Stella

    2011-01-01

    Energinet.dk has launched an investigation of dynamic current ratings of cable based transmission grids, where both internal and external parameters are variables. The first topic was to investigate state of the art within calculating the current carrying capacity (ampacity or loadability......) of cables embedded in larger cable systems. Some recently published research has been concerned with dynamic loadability, but such researches are based on many assumptions. It is shown in the paper, that only limited research has been concerned with larger cable grids, and no remarkable work could been...

  12. MICROARRAY IMAGE GRIDDING USING GRID LINE REFINEMENT TECHNIQUE

    Directory of Open Access Journals (Sweden)

    V.G. Biju

    2015-05-01

    Full Text Available An important stage in microarray image analysis is gridding. Microarray image gridding is done to locate sub arrays in a microarray image and find co-ordinates of spots within each sub array. For accurate identification of spots, most of the proposed gridding methods require human intervention. In this paper a fully automatic gridding method which enhances spot intensity in the preprocessing step as per a histogram based threshold method is used. The gridding step finds co-ordinates of spots from horizontal and vertical profile of the image. To correct errors due to the grid line placement, a grid line refinement technique is proposed. The algorithm is applied on different image databases and results are compared based on spot detection accuracy and time. An average spot detection accuracy of 95.06% depicts the proposed method’s flexibility and accuracy in finding the spot co-ordinates for different database images.

  13. Evolution of user analysis on the grid in ATLAS

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00218990; The ATLAS collaboration; Dewhurst, Alastair

    2017-01-01

    More than one thousand physicists analyse data collected by the ATLAS experiment at the Large Hadron Collider (LHC) at CERN through 150 computing facilities around the world. Efficient distributed analysis requires optimal resource usage and the interplay of several factors: robust grid and software infrastructures, and system capability to adapt to different workloads. The continuous automatic validation of grid sites and the user support provided by a dedicated team of expert shifters have been proven to provide a solid distributed analysis system for ATLAS users. Typical user workflows on the grid, and their associated metrics, are discussed. Measurements of user job performance and typical requirements are also shown.

  14. The more the merrier: grid based modelling of Kepler dwarfs with 5-dimensional stellar grids

    Directory of Open Access Journals (Sweden)

    Serenelli Aldo

    2017-01-01

    Full Text Available We present preliminary results of our grid based modelling (GBM of the dwarf/subgiant sample of stars observed with Kepler including global asteroseismic parameters. GBM analysis in this work is based on a large grid of stellar models that is characterized by five independent parameters: model mass and age, initial metallicity (Zini, initial helium (Yini, and mixing length parameter (αMLT. Using this grid relaxes assumptions used in all previous GBM work where the initial composition is determined by a single parameter and that αMLT is fixed to a solar-calibrated value. The new grid allows us to study, for example, the impact of different galactic chemical enrichment models on the determination of stellar parameters such as mass radius and age. Also, it allows to include new results from stellar atmosphere models on αMLT in the GBM analysis in a simple manner. Alternatively, it can be tested if global asteroseismology is a useful tool to constraint our ignorance on quantities such as Yini and αMLT. Initial findings show that mass determination is robust with respect to freedom in the latter quantities, with a 4.4% maximum deviation for extreme assumptions regarding prior information on Yini – Zini relations and aMLT. On the other hand, tests carried out so far seem to indicate that global seismology does not have much power to constrain Yini – Zni relations of αMLT values without resourcing to additional information.

  15. Disturbance estimator based predictive current control of grid-connected inverters

    OpenAIRE

    Al-Khafaji, Ahmed Samawi Ghthwan

    2013-01-01

    ABSTRACT: The work presented in my thesis considers one of the modern discrete-time control approaches based on digital signal processing methods, that have been developed to improve the performance control of grid-connected three-phase inverters. Disturbance estimator based predictive current control of grid-connected inverters is proposed. For inverter modeling with respect to the design of current controllers, we choose the d-q synchronous reference frame to make it easier to understand an...

  16. CMS on the GRID: Toward a fully distributed computing architecture

    International Nuclear Information System (INIS)

    Innocente, Vincenzo

    2003-01-01

    The computing systems required to collect, analyse and store the physics data at LHC would need to be distributed and global in scope. CMS is actively involved in several grid-related projects to develop and deploy a fully distributed computing architecture. We present here recent developments of tools for automating job submission and for serving data to remote analysis stations. Plans for further test and deployment of a production grid are also described

  17. The impact of job crafting on job demands, job resources, and well-being

    NARCIS (Netherlands)

    Tims, M.; Bakker, A.B.; Derks, D.

    2013-01-01

    This longitudinal study examined whether employees can impact their own well-being by crafting their job demands and resources. Based on the Job Demands-Resources model, we hypothesized that employee job crafting would have an impact on work engagement, job satisfaction, and burnout through changes

  18. Reputation, Princing and the E-Science Grid

    Science.gov (United States)

    Anandasivam, Arun; Neumann, Dirk

    One of the fundamental aspects for an efficient Grid usage is the optimization of resource allocation among the participants. However, this has not yet materialized. Each user is a self-interested participant trying to maximize his utility whereas the utility is not only determined by the fastest completion time, but on the prices as well. Future revenues are influenced by users' reputation. Reputation mechanisms help to build trust between loosely coupled and geographically distributed participants. Providers need an incentive to reduce selfish cancellation of jobs and privilege own jobs. In this chapter we present first an offline scheduling mechanism with a fixed price. Jobs are collected by a broker and scheduled to machines. The goal of the broker is to balance the load and to maximize the revenue in the network. Consumers can submit their jobs according to their preferences, but taking the incentives of the broker into account. This mechanism does not consider reputation. In a second step a reputation-based pricing mechanism for a simple, but fair pricing of resources is analyzed. In e-Science researchers do not appreciate idiosyncratic pricing strategies and policies. Their interest lies in doing research in an efficient manner. Consequently, in our mechanism the price is tightly coupled to the reputation of a site to guarantee fairness of pricing and facilitate price determination. Furthermore, the price is not the only parameter as completion time plays an important role, when deadlines have to be met. We provide a flexible utility and decision model for every participant and analyze the outcome of our reputation-based pricing system via simulation.

  19. Development and deployment of a Desktop and Mobile application on grid for GPS studie

    Science.gov (United States)

    Ntumba, Patient; Lotoy, Vianney; Djungu, Saint Jean; Fleury, Rolland; Petitdidier, Monique; Gemünd, André; Schwichtenberg, Horst

    2013-04-01

    GPS networks for scientific studies are developed all other the world and large databases, regularly updated, like IGS are also available. Many GPS have been installed in West and Central Africa during AMMA (African Monsoon Multiplidisciplinary Analysis), IHY (International heliophysical Year)and many other projects since 2005. African scientists have been educated to use those data especially for meteorological and ionospheric studies. The annual variations of ionospheric parameters for a given station or map of a given region are very intensive computing. Then grid or cloud computing may be a solution to obtain results in a relatively short time. Real time At the University of Kinshasa the chosen solution is a grid of several PCs. It has been deployed by using Globus Toolkit on a Condor pool in order to support the processing of GPS data for ionospheric studies. To be user-friendly, graphical user interfaces(GUI) have been developed to help the user to prepare and submit jobs. One is a java GUI for desktop client, the other is an Android GUI for mobile client. The interest of a grid is the possibility to send a bunch of jobs with an adequate agent control in order to survey the job execution and result storage. After the feasibility study the grid will be extended to a larger number of PCs. Other solutions will be in parallel explored.

  20. Grid-friendly wind power systems based on the synchronverter technology

    International Nuclear Information System (INIS)

    Zhong, Qing-Chang; Ma, Zhenyu; Ming, Wen-Long; Konstantopoulos, George C.

    2015-01-01

    Highlights: • A grid-friendly wind power system that uses the synchronverter technology is proposed. • Both the rotor-side and the grid-side converters act as synchronverters. • The complete generator–motor–generator system improves the performance under grid faults. • Real-time digital simulation results verify the effectiveness of the proposed method. - Abstract: Back-to-back PWM converters are becoming a realistic alternative to conventional converters in high-power wind power applications. In this paper, a control strategy based on the synchronverter technology is proposed for back-to-back PWM converters. Both converters are run as synchronverters, which are mathematically equivalent to the conventional synchronous generators. The rotor-side converter is responsible for maintaining the DC link voltage and the grid-side converter is responsible for the maximum power point tracking (MPPT). As the two converters are operated using the synchronverter technology, the formed wind power system becomes more friendly to the grid. Extensive real-time digital simulation results are presented to verify the effectiveness of the proposed method under normal operation and grid-fault scenarios

  1. The extended RBAC model based on grid computing

    Institute of Scientific and Technical Information of China (English)

    CHEN Jian-gang; WANG Ru-chuan; WANG Hai-yan

    2006-01-01

    This article proposes the extended role-based access control (RBAC) model for solving dynamic and multidomain problems in grid computing, The formulated description of the model has been provided. The introduction of context and the mapping relations of context-to-role and context-to-permission help the model adapt to dynamic property in grid environment.The multidomain role inheritance relation by the authorization agent service realizes the multidomain authorization amongst the autonomy domain. A function has been proposed for solving the role inheritance conflict during the establishment of the multidomain role inheritance relation.

  2. Smart Energy Management and Control for Fuel Cell Based Micro-Grid Connected Neighborhoods

    Energy Technology Data Exchange (ETDEWEB)

    Dr. Mohammad S. Alam

    2006-03-15

    Fuel cell power generation promises to be an efficient, pollution-free, reliable power source in both large scale and small scale, remote applications. DOE formed the Solid State Energy Conversion Alliance with the intention of breaking one of the last barriers remaining for cost effective fuel cell power generation. The Alliance’s goal is to produce a core solid-state fuel cell module at a cost of no more than $400 per kilowatt and ready for commercial application by 2010. With their inherently high, 60-70% conversion efficiencies, significantly reduced carbon dioxide emissions, and negligible emissions of other pollutants, fuel cells will be the obvious choice for a broad variety of commercial and residential applications when their cost effectiveness is improved. In a research program funded by the Department of Energy, the research team has been investigating smart fuel cell-operated residential micro-grid communities. This research has focused on using smart control systems in conjunction with fuel cell power plants, with the goal to reduce energy consumption, reduce demand peaks and still meet the energy requirements of any household in a micro-grid community environment. In Phases I and II, a SEMaC was developed and extended to a micro-grid community. In addition, an optimal configuration was determined for a single fuel cell power plant supplying power to a ten-home micro-grid community. In Phase III, the plan is to expand this work to fuel cell based micro-grid connected neighborhoods (mini-grid). The economic implications of hydrogen cogeneration will be investigated. These efforts are consistent with DOE’s mission to decentralize domestic electric power generation and to accelerate the onset of the hydrogen economy. A major challenge facing the routine implementation and use of a fuel cell based mini-grid is the varying electrical demand of the individual micro-grids, and, therefore, analyzing these issues is vital. Efforts are needed to determine

  3. SQoS based Planning using 4-regular Grid for Optical Fiber Metworks

    DEFF Research Database (Denmark)

    Riaz, Muhammad Tahir; Pedersen, Jens Myrup; Madsen, Ole Brun

    optical fiber based network infrastructures. In the first step of SQoS based planning, this paper describes how 4-regular Grid structures can be implemented in the physical level of optical fiber network infrastructures. A systematic approach for implementing the Grid structure is presented. We used...

  4. SQoS based Planning using 4-regular Grid for Optical Fiber Networks

    DEFF Research Database (Denmark)

    Riaz, Muhammad Tahir; Pedersen, Jens Myrup; Madsen, Ole Brun

    2005-01-01

    optical fiber based network infrastructures. In the first step of SQoS based planning, this paper describes how 4-regular Grid structures can be implemented in the physical level of optical fiber network infrastructures. A systematic approach for implementing the Grid structure is presented. We used...

  5. A Grid-Based Cyber Infrastructure for High Performance Chemical Dynamics Simulations

    Directory of Open Access Journals (Sweden)

    Khadka Prashant

    2008-10-01

    Full Text Available Chemical dynamics simulation is an effective means to study atomic level motions of molecules, collections of molecules, liquids, surfaces, interfaces of materials, and chemical reactions. To make chemical dynamics simulations globally accessible to a broad range of users, recently a cyber infrastructure was developed that provides an online portal to VENUS, a popular chemical dynamics simulation program package, to allow people to submit simulation jobs that will be executed on the web server machine. In this paper, we report new developments of the cyber infrastructure for the improvement of its quality of service by dispatching the submitted simulations jobs from the web server machine onto a cluster of workstations for execution, and by adding an animation tool, which is optimized for animating the simulation results. The separation of the server machine from the simulation-running machine improves the service quality by increasing the capacity to serve more requests simultaneously with even reduced web response time, and allows the execution of large scale, time-consuming simulation jobs on the powerful workstation cluster. With the addition of an animation tool, the cyber infrastructure automatically converts, upon the selection of the user, some simulation results into an animation file that can be viewed on usual web browsers without requiring installation of any special software on the user computer. Since animation is essential for understanding the results of chemical dynamics simulations, this animation capacity provides a better way for understanding simulation details of the chemical dynamics. By combining computing resources at locations under different administrative controls, this cyber infrastructure constitutes a grid environment providing physically and administratively distributed functionalities through a single easy-to-use online portal

  6. Efficient identification of opportunities for Distributed Generation based on Smart Grid Technology

    DEFF Research Database (Denmark)

    Mutule, Anna; Obushevs, Artjoms; Lvov, Aleksandr

    2013-01-01

    The paper presents the main goals and achievements of the Smart Grids ERA-NET project named “Efficient identification of opportunities for Distributed Generation based on Smart Grid Technology (SmartGen)” during the second stage of project implementation. A description of Smart Grid Technology (S......) models developed within the framework of the project is given. The performed study cases where the SGT-models were implemented to analyze the impact of the electrical grid are discussed....

  7. Job optimization in ATLAS TAG-based distributed analysis

    Science.gov (United States)

    Mambelli, M.; Cranshaw, J.; Gardner, R.; Maeno, T.; Malon, D.; Novak, M.

    2010-04-01

    The ATLAS experiment is projected to collect over one billion events/year during the first few years of operation. The efficient selection of events for various physics analyses across all appropriate samples presents a significant technical challenge. ATLAS computing infrastructure leverages the Grid to tackle the analysis across large samples by organizing data into a hierarchical structure and exploiting distributed computing to churn through the computations. This includes events at different stages of processing: RAW, ESD (Event Summary Data), AOD (Analysis Object Data), DPD (Derived Physics Data). Event Level Metadata Tags (TAGs) contain information about each event stored using multiple technologies accessible by POOL and various web services. This allows users to apply selection cuts on quantities of interest across the entire sample to compile a subset of events that are appropriate for their analysis. This paper describes new methods for organizing jobs using the TAGs criteria to analyze ATLAS data. It further compares different access patterns to the event data and explores ways to partition the workload for event selection and analysis. Here analysis is defined as a broader set of event processing tasks including event selection and reduction operations ("skimming", "slimming" and "thinning") as well as DPD making. Specifically it compares analysis with direct access to the events (AOD and ESD data) to access mediated by different TAG-based event selections. We then compare different ways of splitting the processing to maximize performance.

  8. Personnel Selection Method Based on Personnel-Job Matching

    OpenAIRE

    Li Wang; Xilin Hou; Lili Zhang

    2013-01-01

    The existing personnel selection decisions in practice are based on the evaluation of job seeker's human capital, and it may be difficult to make personnel-job matching and make each party satisfy. Therefore, this paper puts forward a new personnel selection method by consideration of bilateral matching. Starting from the employment thoughts of ¡°satisfy¡±, the satisfaction evaluation indicator system of each party are constructed. The multi-objective optimization model is given according to ...

  9. Scheduling strategies for cycle scavenging in multicluster grid systems

    NARCIS (Netherlands)

    Sonmez, O.O.; Grundeken, B.; Mohamed, H.H.; Iosup, A.; Epema, D.H.J.

    2009-01-01

    The use of today's multicluster grids exhibits periods of submission bursts with periods of normal use and even of idleness. To avoid resource contention, many users employ observational scheduling, that is, they postpone the submission of relatively low-priority jobs until a cluster becomes

  10. A methodology toward manufacturing grid-based virtual enterprise operation platform

    Science.gov (United States)

    Tan, Wenan; Xu, Yicheng; Xu, Wei; Xu, Lida; Zhao, Xianhua; Wang, Li; Fu, Liuliu

    2010-08-01

    Virtual enterprises (VEs) have become one of main types of organisations in the manufacturing sector through which the consortium companies organise their manufacturing activities. To be competitive, a VE relies on the complementary core competences among members through resource sharing and agile manufacturing capacity. Manufacturing grid (M-Grid) is a platform in which the production resources can be shared. In this article, an M-Grid-based VE operation platform (MGVEOP) is presented as it enables the sharing of production resources among geographically distributed enterprises. The performance management system of the MGVEOP is based on the balanced scorecard and has the capacity of self-learning. The study shows that a MGVEOP can make a semi-automated process possible for a VE, and the proposed MGVEOP is efficient and agile.

  11. Experimental evaluation of job provenance in ATLAS environment

    International Nuclear Information System (INIS)

    Krenek, A; Sitera, J; Chudoba, J; Dvorak, F; Filipovic, J; KmunIcek, J; Matyska, L; Mulas, M; Ruda, M; Sustr, Z; Campana, S; Molinari, E; Rebatto, D

    2008-01-01

    Grid middleware stacks, including gLite, matured into the state of being able to process up to millions of jobs per day. Logging and Bookkeeping, the gLite job-tracking service, keeps pace with this rate; however, it is not designed to provide a long-term archive of information on executed jobs. ATLAS - representative of a large user community - addresses this issue with its own job catalogue (ProdDB). Development of such a customized service, not easily reusable, took considerable effort which is not affordable by smaller communities. On the contrary, Job Provenance (JP), a generic gLite service designed for long-term archiving of information on executed jobs focusing on scalability, extensibility, uniform data view, and configurability, allows more specialized catalogues to be easily built. We present the first results of an experimental JP deployment for the ATLAS production infrastructure where a JP installation was fed with a part of ATLAS jobs, and also stress tested with real production data. The main outcome of this work is a demonstration that JP can complement large-scale application-specific job catalogue services, while serving a similar purpose where there are none available

  12. Deploying web-based visual exploration tools on the grid

    Energy Technology Data Exchange (ETDEWEB)

    Jankun-Kelly, T.J.; Kreylos, Oliver; Shalf, John; Ma, Kwan-Liu; Hamann, Bernd; Joy, Kenneth; Bethel, E. Wes

    2002-02-01

    We discuss a web-based portal for the exploration, encapsulation, and dissemination of visualization results over the Grid. This portal integrates three components: an interface client for structured visualization exploration, a visualization web application to manage the generation and capture of the visualization results, and a centralized portal application server to access and manage grid resources. We demonstrate the usefulness of the developed system using an example for Adaptive Mesh Refinement (AMR) data visualization.

  13. OPNET/Simulink Based Testbed for Disturbance Detection in the Smart Grid

    Energy Technology Data Exchange (ETDEWEB)

    Sadi, Mohammad A. H. [University of Memphis; Dasgupta, Dipankar [ORNL; Ali, Mohammad Hassan [University of Memphis; Abercrombie, Robert K [ORNL

    2015-01-01

    The important backbone of the smart grid is the cyber/information infrastructure, which is primarily used to communicate with different grid components. A smart grid is a complex cyber physical system containing a numerous and variety number of sources, devices, controllers and loads. Therefore, the smart grid is vulnerable to grid related disturbances. For such dynamic system, disturbance and intrusion detection is a paramount issue. This paper presents a Simulink and Opnet based co-simulated platform to carry out a cyber-intrusion in cyber network for modern power systems and the smart grid. The IEEE 30 bus power system model is used to demonstrate the effectiveness of the simulated testbed. The experiments were performed by disturbing the circuit breakers reclosing time through a cyber-attack. Different disturbance situations in the considered test system are considered and the results indicate the effectiveness of the proposed co-simulated scheme.

  14. Mini-Grids for the Base of the Pyramid Market: A Critical Review

    Directory of Open Access Journals (Sweden)

    Subhes C. Bhattacharyya

    2018-04-01

    Full Text Available The lack of access to electricity of more than 1.1 billion people around the world remains a major developmental challenge and Goal 7 of the Sustainable Development Goals (SDG as well as Sustainable Energy for All (SE4ALL have set a target of universal electrification by 2030. Various studies have identified mini-grid-based electrification as a possible solution. There is a growing body of literature available now that has explored the feasibility, practical application and policy interventions required to support mini-grids. Through a review of available literature, this paper explores whether mini-grids can be a solution for the base of the pyramid (BoP market and the challenges faced in deploying mini-grids in such markets. Interventions to support the mini-grid deployment are also discussed. The paper finds that the mini-grids are targeting the BoP market but the business is not attractive in profitability terms and requires financial support. Lack of regulatory clarity and non-coordinated policies affect the financial viability of projects, which requires careful support. Mini-grid electrification has hardly been embedded in rural development agenda and hence they have not contributed significantly to livelihood generation. Careful realignment of policies, regulatory frameworks and support systems can better support mini-grid deployment in developing countries.

  15. A genetic algorithm-based job scheduling model for big data analytics.

    Science.gov (United States)

    Lu, Qinghua; Li, Shanshan; Zhang, Weishan; Zhang, Lei

    Big data analytics (BDA) applications are a new category of software applications that process large amounts of data using scalable parallel processing infrastructure to obtain hidden value. Hadoop is the most mature open-source big data analytics framework, which implements the MapReduce programming model to process big data with MapReduce jobs. Big data analytics jobs are often continuous and not mutually separated. The existing work mainly focuses on executing jobs in sequence, which are often inefficient and consume high energy. In this paper, we propose a genetic algorithm-based job scheduling model for big data analytics applications to improve the efficiency of big data analytics. To implement the job scheduling model, we leverage an estimation module to predict the performance of clusters when executing analytics jobs. We have evaluated the proposed job scheduling model in terms of feasibility and accuracy.

  16. Active Job Monitoring in Pilots

    Science.gov (United States)

    Kuehn, Eileen; Fischer, Max; Giffels, Manuel; Jung, Christopher; Petzold, Andreas

    2015-12-01

    Recent developments in high energy physics (HEP) including multi-core jobs and multi-core pilots require data centres to gain a deep understanding of the system to monitor, design, and upgrade computing clusters. Networking is a critical component. Especially the increased usage of data federations, for example in diskless computing centres or as a fallback solution, relies on WAN connectivity and availability. The specific demands of different experiments and communities, but also the need for identification of misbehaving batch jobs, requires an active monitoring. Existing monitoring tools are not capable of measuring fine-grained information at batch job level. This complicates network-aware scheduling and optimisations. In addition, pilots add another layer of abstraction. They behave like batch systems themselves by managing and executing payloads of jobs internally. The number of real jobs being executed is unknown, as the original batch system has no access to internal information about the scheduling process inside the pilots. Therefore, the comparability of jobs and pilots for predicting run-time behaviour or network performance cannot be ensured. Hence, identifying the actual payload is important. At the GridKa Tier 1 centre a specific tool is in use that allows the monitoring of network traffic information at batch job level. This contribution presents the current monitoring approach and discusses recent efforts and importance to identify pilots and their substructures inside the batch system. It will also show how to determine monitoring data of specific jobs from identified pilots. Finally, the approach is evaluated.

  17. ETICS: the international software engineering service for the grid

    Energy Technology Data Exchange (ETDEWEB)

    Meglio, A D; Begin, M-E [CERN (Switzerland); Couvares, P [University of Wisconsin-Madison (United States); Ronchieri, E [INFN CNAF (Italy); Takacs, E [4D SOFT Ltd (Hungary)], E-mail: alberto.di.meglio@cern.ch

    2008-07-15

    The ETICS system is a distributed software configuration, build and test system designed to fulfil the needs of improving the quality, reliability and interoperability of distributed software in general and grid software in particular. The ETICS project is a consortium of five partners (CERN, INFN, Engineering Ingegneria Informatica, 4D Soft and the University of Wisconsin-Madison). The ETICS service consists of a build and test job execution system based on the Metronome software and an integrated set of web services and software engineering tools to design, maintain and control build and test scenarios. The ETICS system allows taking into account complex dependencies among applications and middleware components and provides a rich environment to perform static and dynamic analysis of the software and execute deployment, system and interoperability tests. This paper gives an overview of the system architecture and functionality set and then describes how the EC-funded EGEE, DILIGENT and OMII-Europe projects are using the software engineering services to build, validate and distribute their software. Finally a number of significant use and test cases will be described to show how ETICS can be used in particular to perform interoperability tests of grid middleware using the grid itself.

  18. ETICS: the international software engineering service for the grid

    Science.gov (United States)

    Meglio, A. D.; Bégin, M.-E.; Couvares, P.; Ronchieri, E.; Takacs, E.

    2008-07-01

    The ETICS system is a distributed software configuration, build and test system designed to fulfil the needs of improving the quality, reliability and interoperability of distributed software in general and grid software in particular. The ETICS project is a consortium of five partners (CERN, INFN, Engineering Ingegneria Informatica, 4D Soft and the University of Wisconsin-Madison). The ETICS service consists of a build and test job execution system based on the Metronome software and an integrated set of web services and software engineering tools to design, maintain and control build and test scenarios. The ETICS system allows taking into account complex dependencies among applications and middleware components and provides a rich environment to perform static and dynamic analysis of the software and execute deployment, system and interoperability tests. This paper gives an overview of the system architecture and functionality set and then describes how the EC-funded EGEE, DILIGENT and OMII-Europe projects are using the software engineering services to build, validate and distribute their software. Finally a number of significant use and test cases will be described to show how ETICS can be used in particular to perform interoperability tests of grid middleware using the grid itself.

  19. ETICS: the international software engineering service for the grid

    International Nuclear Information System (INIS)

    Meglio, A D; Begin, M-E; Couvares, P; Ronchieri, E; Takacs, E

    2008-01-01

    The ETICS system is a distributed software configuration, build and test system designed to fulfil the needs of improving the quality, reliability and interoperability of distributed software in general and grid software in particular. The ETICS project is a consortium of five partners (CERN, INFN, Engineering Ingegneria Informatica, 4D Soft and the University of Wisconsin-Madison). The ETICS service consists of a build and test job execution system based on the Metronome software and an integrated set of web services and software engineering tools to design, maintain and control build and test scenarios. The ETICS system allows taking into account complex dependencies among applications and middleware components and provides a rich environment to perform static and dynamic analysis of the software and execute deployment, system and interoperability tests. This paper gives an overview of the system architecture and functionality set and then describes how the EC-funded EGEE, DILIGENT and OMII-Europe projects are using the software engineering services to build, validate and distribute their software. Finally a number of significant use and test cases will be described to show how ETICS can be used in particular to perform interoperability tests of grid middleware using the grid itself

  20. Performance-based training: from job and task analysis to training materials

    International Nuclear Information System (INIS)

    Davis, L.T.; Spinney, R.W.

    1983-01-01

    Historically, the smoke filled room approach has been used to revise training programs: instructors would sit down and design a program based on existing training materials and any federal requirements that applied. This failure to reflect a systematic definition of required job functions, responsibilities and performance standards in training programs has resulted in generic program deficiencies: they do not provide complete training of required skills and knowledge. Recognition of this need for change, coupled with a decrease in experienced industry personnel inputs and long training pipelines, has heightened the need for efficient performance-based training programs which are derived from and referenced to job performance criteria. This paper presents the process for developing performance-based training materials based on job and task analysis products

  1. Application of Job Demands-Resources model in research on relationships between job satisfaction, job resources, individual resources and job demands

    OpenAIRE

    Adrianna Potocka; Małgorzata Waszkowska

    2013-01-01

    Background: The aim of this study was to explore the relationships between job demands, job resourses, personal resourses and job satisfaction and to assess the usefulness of the Job Demands-Resources (JD-R) model in the explanation of these phenomena. Materials and Methods: The research was based on a sample of 500 social workers. The "Psychosocial Factors" and "Job satisfaction" questionnaires were used to test the hypothesis. Results: The results showed that job satisfaction increased with...

  2. Using Hadoop as a grid storage element

    International Nuclear Information System (INIS)

    Bockelman, Brian

    2009-01-01

    Hadoop is an open-source data processing framework that includes a scalable, fault-tolerant distributed file system, HDFS. Although HDFS was designed to work in conjunction with Hadoop's job scheduler, we have re-purposed it to serve as a grid storage element by adding GridFTP and SRM servers. We have tested the system thoroughly in order to understand its scalability and fault tolerance. The turn-on of the Large Hadron Collider (LHC) in 2009 poses a significant data management and storage challenge; we have been working to introduce HDFS as a solution for data storage for one LHC experiment, the Compact Muon Solenoid (CMS).

  3. Agent-based Decentralization of Applications in Distributed Smart Grid Systems

    DEFF Research Database (Denmark)

    Kienesberger, Georg; Xypolytou, Evangelia; Marchgraber, Jurgen

    2015-01-01

    systems (DMACS) and aims to give an overview on the different requirements and challenges on the way from current centralized control systems to DMACS. Therefore, different ICT scenarios and MAS topologies are employed to discuss the decentralization of three exemplary smart grid applications: voltage......Smart grid technology promises to prepare today’s power systems for the challenges of the future by extensive integration of information and communication technology (ICT). One key aspect is the control paradigm which will have to be shifted from completely centralized control systems to more...... dezentralized concepts in order to adapt to the distributed nature of smart grids. Multi-agent systems (MAS) are a very promising approach for designing distributed, decentralized systems, naturally also in the field of smart grids. This work introduces the notion of decentralized multi-agent-based control...

  4. AMP: a science-driven web-based application for the TeraGrid

    Science.gov (United States)

    Woitaszek, M.; Metcalfe, T.; Shorrock, I.

    The Asteroseismic Modeling Portal (AMP) provides a web-based interface for astronomers to run and view simulations that derive the properties of Sun-like stars from observations of their pulsation frequencies. In this paper, we describe the architecture and implementation of AMP, highlighting the lightweight design principles and tools used to produce a functional fully-custom web-based science application in less than a year. Targeted as a TeraGrid science gateway, AMP's architecture and implementation are intended to simplify its orchestration of TeraGrid computational resources. AMP's web-based interface was developed as a traditional standalone database-backed web application using the Python-based Django web development framework, allowing us to leverage the Django framework's capabilities while cleanly separating the user interface development from the grid interface development. We have found this combination of tools flexible and effective for rapid gateway development and deployment.

  5. Professional confidence and job satisfaction: an examination of counselors' perceptions in faith-based and non-faith-based drug treatment programs.

    Science.gov (United States)

    Chu, Doris C; Sung, Hung-En

    2014-08-01

    Understanding substance abuse counselors' professional confidence and job satisfaction is important since such confidence and satisfaction can affect the way counselors go about their jobs. Analyzing data derived from a random sample of 110 counselors from faith-based and non-faith-based treatment programs, this study examines counselors' professional confidence and job satisfaction in both faith-based and non-faith-based programs. The multivariate analyses indicate years of experience and being a certified counselor were the only significant predictors of professional confidence. There was no significant difference in perceived job satisfaction and confidence between counselors in faith-based and non-faith-based programs. A majority of counselors in both groups expressed a high level of satisfaction with their job. Job experience in drug counseling and prior experience as an abuser were perceived by counselors as important components to facilitate counseling skills. Policy implications are discussed. © The Author(s) 2013.

  6. Deployment of a Grid-based Medical Imaging Application

    CERN Document Server

    Amendolia, S R; Frate, C; Gálvez, J; Hassan, W; Hauer, T; Manset, D; McClatchey, R; Odeh, M; Rogulin, D; Solomonides, T; Warren, R

    2005-01-01

    The MammoGrid project has deployed its Service-Oriented Architecture (SOA)-based Grid application in a real environment comprising actual participating hospitals. The resultant setup is currently being exploited to conduct rigorous in-house tests in the first phase before handing over the setup to the actual clinicians to get their feedback. This paper elaborates the deployment details and the experiences acquired during this phase of the project. Finally the strategy regarding migration to an upcoming middleware from EGEE project will be described. This paper concludes by highlighting some of the potential areas of future work.

  7. Single-Phase LLCL-Filter-based Grid-Tied Inverter with Low-Pass Filter Based Capacitor Current Feedback Active damper

    DEFF Research Database (Denmark)

    Liu, Yuan; Wu, Weimin; Li, Yun

    2016-01-01

    The capacitor-current-feedback active damping method is attractive for high-order-filter-based high power grid-tied inverter when the grid impedance varies within a wide range. In order to improve the system control bandwidth and attenuate the high order grid background harmonics by using the quasi....... In this paper, a low pass filter is proposed to be inserted in the capacitor current feedback loop op LLCL-filter based grid-tied inverter together with a digital proportional and differential compensator. The detailed theoretical analysis is given. For verification, simulations on a 2kW/220V/10kHz LLCL...

  8. The GridSite Web/Grid security system

    International Nuclear Information System (INIS)

    McNab, Andrew; Li Yibiao

    2010-01-01

    We present an overview of the current status of the GridSite toolkit, describing the security model for interactive and programmatic uses introduced in the last year. We discuss our experiences of implementing these internal changes and how they and previous rounds of improvements have been prompted by requirements from users and wider security trends in Grids (such as CSRF). Finally, we explain how these have improved the user experience of GridSite-based websites, and wider implications for portals and similar web/grid sites.

  9. Software-Based Challenges of Developing the Future Distribution Grid

    Energy Technology Data Exchange (ETDEWEB)

    Stewart, Emma; Kiliccote, Sila; McParland, Charles

    2014-06-01

    distribution grid modeling, and measured data sources are a key missing element . Modeling tools need to be calibrated based on measured grid data to validate their output in varied conditions such as high renewables penetration and rapidly changing topology. In addition, establishing a standardized data modeling format would enable users to transfer data among tools to take advantage of different analysis features. ?

  10. A Costing Analysis for Decision Making Grid Model in Failure-Based Maintenance

    Directory of Open Access Journals (Sweden)

    Burhanuddin M. A.

    2011-01-01

    Full Text Available Background. In current economic downturn, industries have to set good control on production cost, to maintain their profit margin. Maintenance department as an imperative unit in industries should attain all maintenance data, process information instantaneously, and subsequently transform it into a useful decision. Then act on the alternative to reduce production cost. Decision Making Grid model is used to identify strategies for maintenance decision. However, the model has limitation as it consider two factors only, that is, downtime and frequency of failures. We consider third factor, cost, in this study for failure-based maintenance. The objective of this paper is to introduce the formulae to estimate maintenance cost. Methods. Fish bone analysis conducted with Ishikawa model and Decision Making Grid methods are used in this study to reveal some underlying risk factors that delay failure-based maintenance. The goal of the study is to estimate the risk factor that is, repair cost to fit in the Decision Making Grid model. Decision Making grid model consider two variables, frequency of failure and downtime in the analysis. This paper introduces third variable, repair cost for Decision Making Grid model. This approaches give better result to categorize the machines, reduce cost, and boost the earning for the manufacturing plant. Results. We collected data from one of the food processing factories in Malaysia. From our empirical result, Machine C, Machine D, Machine F, and Machine I must be in the Decision Making Grid model even though their frequency of failures and downtime are less than Machine B and Machine N, based on the costing analysis. The case study and experimental results show that the cost analysis in Decision Making Grid model gives more promising strategies in failure-based maintenance. Conclusions. The improvement of Decision Making Grid model for decision analysis with costing analysis is our contribution in this paper for

  11. An Analysis of Delay-based and Integrator-based Sequence Detectors for Grid-Connected Converters

    DEFF Research Database (Denmark)

    Khazraj, Hesam; Silva, Filipe Miguel Faria da; Bak, Claus Leth

    2017-01-01

    -signal cancellation operators are the main members of the delay-based sequence detectors. The aim of this paper is to provide a theoretical and experimental comparative study between integrator and delay based sequence detectors. The theoretical analysis is conducted based on the small-signal modelling......Detecting and separating positive and negative sequence components of the grid voltage or current is of vital importance in the control of grid-connected power converters, HVDC systems, etc. To this end, several techniques have been proposed in recent years. These techniques can be broadly...... classified into two main classes: The integrator-based techniques and Delay-based techniques. The complex-coefficient filter-based technique, dual second-order generalized integrator-based method, multiple reference frame approach are the main members of the integrator-based sequence detector and the delay...

  12. JOB SHOP METHODOLOGY BASED ON AN ANT COLONY

    Directory of Open Access Journals (Sweden)

    OMAR CASTRILLON

    2009-01-01

    Full Text Available The purpose of this study is to reduce the total process time (Makespan and to increase the machines working time, in a job shop environment, using a heuristic based on ant colony optimization. This work is developed in two phases: The first stage describes the identification and definition of heuristics for the sequential processes in the job shop. The second stage shows the effectiveness of the system in the traditional programming of production. A good solution, with 99% efficiency is found using this technique.

  13. GANGA powerful job submission and management tool

    CERN Document Server

    Maier, Andrew; Mendez Lorenzo, Patricia; Moscicki, Jakub; Lamanna, Massimo; Muraru, Adrian

    2008-01-01

    The computational and storage capability of the Grid are attracting several research communities, also beyond HEP. Ganga is a lightweight Grid job management tool developed at CERN. It is a key component in the distributed Data Analysis for ATLAS and LHCb. Ganga`s open and general framework allows to plug-in applications, which has attracted users from other domains outside HEP. In addition, Ganga interfaces to a variety of Grid and non-Grid backends using the same, simple end-user interface Ganga has already gained widespread use, the incomplete list of applications using Ganga include: Imaging processing and classification (developed by Cambridge Ontology Ltd.), Theoretical physics (Lattice QCD, Feynman-loop evaluation), Bio-informatics (Avian Flu Data Challenge), Geant4 (Monte Carlo package), HEP data analysis (ATLAS, LHCb). All these communities have different goals and requirements and the main challenge is the creation of a standard and general software infrastructure for the immersion of these communit...

  14. The impact of job crafting on job demands, job resources, and well-being.

    Science.gov (United States)

    Tims, Maria; Bakker, Arnold B; Derks, Daantje

    2013-04-01

    This longitudinal study examined whether employees can impact their own well-being by crafting their job demands and resources. Based on the job demands-resources model, we hypothesized that employee job crafting would have an impact on work engagement, job satisfaction, and burnout through changes in job demands and job resources. Data was collected in a chemical plant at three time points with one month in between the measurement waves (N = 288). The results of structural equation modeling showed that employees who crafted their job resources in the first month of the study showed an increase in their structural and social resources over the course of the study (2 months). This increase in job resources was positively related to employee well-being (increased engagement and job satisfaction, and decreased burnout). Crafting job demands did not result in a change in job demands, but results revealed direct effects of crafting challenging demands on increases in well-being. We conclude that employee job crafting has a positive impact on well-being and that employees therefore should be offered opportunities to craft their own jobs.

  15. caGrid 1.0: a Grid enterprise architecture for cancer research.

    Science.gov (United States)

    Oster, Scott; Langella, Stephen; Hastings, Shannon; Ervin, David; Madduri, Ravi; Kurc, Tahsin; Siebenlist, Frank; Covitz, Peter; Shanbhag, Krishnakant; Foster, Ian; Saltz, Joel

    2007-10-11

    caGrid is the core Grid architecture of the NCI-sponsored cancer Biomedical Informatics Grid (caBIG) program. The current release, caGrid version 1.0, is developed as the production Grid software infrastructure of caBIG. Based on feedback from adopters of the previous version (caGrid 0.5), it has been significantly enhanced with new features and improvements to existing components. This paper presents an overview of caGrid 1.0, its main components, and enhancements over caGrid 0.5.

  16. Designing the organizational creativity based on job characteristics and job involvement among physical education units’ employees of Technology University

    Directory of Open Access Journals (Sweden)

    Seyed Hossein Alavi

    2017-10-01

    Full Text Available The aim of the present study was designing the organizational creativity based on job characteristics and job involvement among physical education employees of Technology University. The statistical population of the present study consisted of all employees, experts, and physical education units’ managers. The statistical samples were 389 people reduced to 338 employees after eliminating the incomplete and useless questionnaires. The research method of the present study was descriptive-correlative. Three questionnaires of job characteristics, job involvement, and organizational creativity were used to evaluate the research variables which were distributed and gathered among research samples after translation and comparison with research samples via face validity (views of 12 sport management experts and construct validity. The research results were analyzed by Partial Least Square (PLS and showed that the best predictor of job involvement among physical education units’ employees of Technology University was problem solving (ES= 0.41. However, the job characteristics of autonomy (0.26 and job feedback (.024 had more effects on employees’ creativity. Other results showed that managers reported lower scores of job characteristics and creativity than sport employees. According to research findings, in order to achieve creativity in the workplace within physical education units of Technology University, the managers should stimulate the autonomy and feedback characteristics. Of course, it should not be neglected that this mechanism would be facilitated by creating the problem solving in job.

  17. Grid Portal for Image and Video Processing

    International Nuclear Information System (INIS)

    Dinitrovski, I.; Kakasevski, G.; Buckovska, A.; Loskovska, S.

    2007-01-01

    Users are typically best served by G rid Portals . G rid Portals a re web servers that allow the user to configure or run a class of applications. The server is then given the task of authentication of the user with the Grid and invocation of the required grid services to launch the user's application. PHP is a widely-used general-purpose scripting language that is especially suited for Web development and can be embedded into HTML. PHP is powerful and modern server-side scripting language producing HTML or XML output which easily can be accessed by everyone via web interface (with the browser of your choice) and can execute shell scripts on the server side. The aim of our work is development of Grid portal for image and video processing. The shell scripts contains gLite and globus commands for obtaining proxy certificate, job submission, data management etc. Using this technique we can easily create web interface to the Grid infrastructure. The image and video processing algorithms are implemented in C++ language using various image processing libraries. (Author)

  18. Development of a Cartesian grid based CFD solver (CARBS)

    International Nuclear Information System (INIS)

    Vaidya, A.M.; Maheshwari, N.K.; Vijayan, P.K.

    2013-12-01

    Formulation for 3D transient incompressible CFD solver is developed. The solution of variable property, laminar/turbulent, steady/unsteady, single/multi specie, incompressible with heat transfer in complex geometry will be obtained. The formulation can handle a flow system in which any number of arbitrarily shaped solid and fluid regions are present. The solver is based on the use of Cartesian grids. A method is proposed to handle complex shaped objects and boundaries on Cartesian grids. Implementation of multi-material, different types of boundary conditions, thermo physical properties is also considered. The proposed method is validated by solving two test cases. 1 st test case is that of lid driven flow in inclined cavity. 2 nd test case is the flow over cylinder. The 1 st test case involved steady internal flow subjected to WALL boundaries. The 2 nd test case involved unsteady external flow subjected to INLET, OUTLET and FREE-SLIP boundary types. In both the test cases, non-orthogonal geometry was involved. It was found that, under such a wide conditions, the Cartesian grid based code was found to give results which were matching well with benchmark data. Convergence characteristics are excellent. In all cases, the mass residue was converged to 1E-8. Based on this, development of 3D general purpose code based on the proposed approach can be taken up. (author)

  19. Deployment of job priority mechanisms in the Italian Cloud of the ATLAS experiment

    International Nuclear Information System (INIS)

    Doria, Alessandra; Carlino, Gianpaolo; Salvo, Alessandro De; Musto, Elisa; Barchiesi, Alex; Campana, Simone; Ciocca, Claudia; Italiano, Alessandro; Rinaldi, Lorenzo; Salomoni, Davide; Perini, Laura; Pistolese, Massimo; Vaccarossa, Luca; Vilucchi, Elisabetta

    2010-01-01

    An optimized use of the Grid computing resources in the ATLAS experiment requires the enforcement of a mechanism of job priorities and of resource sharing among the different activities inside the ATLAS VO. This mechanism has been implemented through the VOViews publication in the information system and the fair share implementation per UNIX group in the batch system. The VOView concept consists of publishing resource information, such as running and waiting jobs, as a function of VO groups and roles. The ATLAS Italian Cloud is composed of the CNAF Tier1 and Roma Tier2, with farms based on the LSF batch system, and the Tier2s of Frascati, Milano and Napoli based on PBS/Torque. In this paper we describe how test and deployment of the job priorities has been performed in the cloud, where the VOMS-based regional group /atlas/it has been created. We show that the VOViews are published and correctly managed by the WMS and that the resources allocated to generic VO users, users with production role and users of the /atlas/it group correspond to the defined share.

  20. One-fifth of nonelderly Californians do not have access to job-based health insurance coverage.

    Science.gov (United States)

    Lavarreda, Shana Alex; Cabezas, Livier

    2010-11-01

    Lack of job-based health insurance does not affect just workers, but entire families who depend on job-based coverage for their health care. This policy brief shows that in 2007 one-fifth of all Californians ages 0-64 who lived in households where at least one family member was employed did not have access to job-based coverage. Among adults with no access to job-based coverage through their own or a spouse's job, nearly two-thirds remained uninsured. In contrast, the majority of children with no access to health insurance through a parent obtained public health insurance, highlighting the importance of such programs. Low-income, Latino and small business employees were more likely to have no access to job-based insurance. Provisions enacted under national health care reform (the Patient Protection and Affordable Care Act of 2010) will aid some of these populations in accessing health insurance coverage.

  1. The Geographic Information Grid System Based on Mobile Agent

    Institute of Scientific and Technical Information of China (English)

    2006-01-01

    We analyze the deficiencies of current application systems, and discuss the key requirements of distributed Geographic Information service (GIS). We construct the distributed GIS on grid platform. Considering the flexibility and efficiency, we integrate the mobile agent technology into the system. We propose a new prototype system, the Geographic Information Grid System (GIGS) based on mobile agent. This system has flexible services and high performance, and improves the sharing of distributed resources. The service strategy of the system and the examples are also presented.

  2. Public storage for the Open Science Grid

    International Nuclear Information System (INIS)

    Levshina, T; Guru, A

    2014-01-01

    The Open Science Grid infrastructure doesn't provide efficient means to manage public storage offered by participating sites. A Virtual Organization that relies on opportunistic storage has difficulties finding appropriate storage, verifying its availability, and monitoring its utilization. The involvement of the production manager, site administrators and VO support personnel is required to allocate or rescind storage space. One of the main requirements for Public Storage implementation is that it should use SRM or GridFTP protocols to access the Storage Elements provided by the OSG Sites and not put any additional burden on sites. By policy, no new services related to Public Storage can be installed and run on OSG sites. Opportunistic users also have difficulties in accessing the OSG Storage Elements during the execution of jobs. A typical users' data management workflow includes pre-staging common data on sites before a job's execution, then storing for a subsequent download to a local institution the output data produced by a job on a worker node. When the amount of data is significant, the only means to temporarily store the data is to upload it to one of the Storage Elements. In order to do that, a user's job should be aware of the storage location, availability, and free space. After a successful data upload, users must somehow keep track of the data's location for future access. In this presentation we propose solutions for storage management and data handling issues in the OSG. We are investigating the feasibility of using the integrated Rule-Oriented Data System developed at RENCI as a front-end service to the OSG SEs. The current architecture, state of deployment and performance test results will be discussed. We will also provide examples of current usage of the system by beta-users.

  3. Public storage for the Open Science Grid

    Science.gov (United States)

    Levshina, T.; Guru, A.

    2014-06-01

    The Open Science Grid infrastructure doesn't provide efficient means to manage public storage offered by participating sites. A Virtual Organization that relies on opportunistic storage has difficulties finding appropriate storage, verifying its availability, and monitoring its utilization. The involvement of the production manager, site administrators and VO support personnel is required to allocate or rescind storage space. One of the main requirements for Public Storage implementation is that it should use SRM or GridFTP protocols to access the Storage Elements provided by the OSG Sites and not put any additional burden on sites. By policy, no new services related to Public Storage can be installed and run on OSG sites. Opportunistic users also have difficulties in accessing the OSG Storage Elements during the execution of jobs. A typical users' data management workflow includes pre-staging common data on sites before a job's execution, then storing for a subsequent download to a local institution the output data produced by a job on a worker node. When the amount of data is significant, the only means to temporarily store the data is to upload it to one of the Storage Elements. In order to do that, a user's job should be aware of the storage location, availability, and free space. After a successful data upload, users must somehow keep track of the data's location for future access. In this presentation we propose solutions for storage management and data handling issues in the OSG. We are investigating the feasibility of using the integrated Rule-Oriented Data System developed at RENCI as a front-end service to the OSG SEs. The current architecture, state of deployment and performance test results will be discussed. We will also provide examples of current usage of the system by beta-users.

  4. [Application of job demands-resources model in research on relationships between job satisfaction, job resources, individual resources and job demands].

    Science.gov (United States)

    Potocka, Adrianna; Waszkowska, Małgorzata

    2013-01-01

    The aim of this study was to explore the relationships between job demands, job resourses, personal resourses and job satisfaction and to assess the usefulness of the Job Demands-Resources (JD-R) model in the explanation of these phenomena. The research was based on a sample of 500 social workers. The "Psychosocial Factors" and "Job satisfaction" questionnaires were used to test the hypothesis. The results showed that job satisfaction increased with increasing job accessibility and personal resources (r = 0.44; r = 0.31; p job resources and job demands [F(1.474) = 4.004; F(1.474) = 4.166; p job satisfaction. Moreover, interactions between job demands and job resources [F(3,474) = 2.748; p job demands and personal resources [F(3.474) = 3.021; p job satisfaction. The post hoc tests showed that 1) in low job demands, but high job resources employees declared higher job satisfaction, than those who perceived them as medium (p = 0.0001) or low (p = 0.0157); 2) when the level of job demands was perceived as medium, employees with high personal resources declared significantly higher job satisfaction than those with low personal resources (p = 0.0001). The JD-R model can be used to investigate job satisfaction. Taking into account fundamental factors of this model, in organizational management there are possibilities of shaping job satisfaction among employees.

  5. Grid impedance estimation based hybrid islanding detection method for AC microgrids

    DEFF Research Database (Denmark)

    Ghzaiel, Walid; Jebali-Ben Ghorbal, Manel; Slama-Belkhodja, Ilhem

    2017-01-01

    This paper focuses on a hybrid islanding detection algorithm for parallel-inverters-based microgrids. The proposed algorithm is implemented on the unit ensuring the control of the intelligent bypass switch connecting or disconnecting the microgrid from the utility. This method employs a grid...... to avoid interactions with other units. The selected inverter will be the one closest to the controllable distributed generation system or to a healthy grid side in case of meshed microgrid with multiple-grid connections. The detection algorithm is applied to quickly detect the resonance phenomena, so...

  6. Improving mobile robot localization: grid-based approach

    Science.gov (United States)

    Yan, Junchi

    2012-02-01

    Autonomous mobile robots have been widely studied not only as advanced facilities for industrial and daily life automation, but also as a testbed in robotics competitions for extending the frontier of current artificial intelligence. In many of such contests, the robot is supposed to navigate on the ground with a grid layout. Based on this observation, we present a localization error correction method by exploring the geometric feature of the tile patterns. On top of the classical inertia-based positioning, our approach employs three fiber-optic sensors that are assembled under the bottom of the robot, presenting an equilateral triangle layout. The sensor apparatus, together with the proposed supporting algorithm, are designed to detect a line's direction (vertical or horizontal) by monitoring the grid crossing events. As a result, the line coordinate information can be fused to rectify the cumulative localization deviation from inertia positioning. The proposed method is analyzed theoretically in terms of its error bound and also has been implemented and tested on a customary developed two-wheel autonomous mobile robot.

  7. Work stress and job satisfaction in hospital-based home care.

    Science.gov (United States)

    Beck-Friis, B; Strang, P; Sjödén, P O

    1991-01-01

    The entire staff of the hospital-based home care (HBHC) at Motala (n = 35) participated in a study concerning work stress and job satisfaction. A significant number of the patients in the HBHC have advanced malignancies and most of them are terminally ill. A total of 219 questions about stress and job satisfaction were asked in a self-administered questionnaire. Only 3%-17% of the staff often or very often experienced stress factors such as high expectations, confusing orders, or lack of information. Instead, a majority stated that they often/very often experienced different aspects of job satisfaction, such as meaningfulness, security, and stimulation. Staff members stating that they often were proud/very proud of their jobs, members feeling that their skill and experience were needed, as well as staff members who often received praise from their superiors, were less prone to look for other jobs (p less than 0.01, p less than 0.05, and p less than 0.05, respectively). Those who often/very often were allowed to take initiatives of their own more often regarded their jobs as non-monotonous (p less than 0.05) and stimulating to their personal development (p less than 0.001). Despite demanding jobs with severely ill patients, most of the staff gave high ratings for different aspects of job satisfaction. This positive spirit was also reflected in the exceptionally low job turnover among them. Possible explanations may be a careful selection of personnel and an organization which both stimulates the staff's own initiatives and provides support when necessary.

  8. Grid-based Simulation of Industrial Thin Film Production

    NARCIS (Netherlands)

    Krzhizhanovskaya, V.V.; Sloot, P.M.A.; Gorbachev, Y.E.

    2005-01-01

    In this article, the authors introduce a Grid-based virtual reactor, a High Level Architecture (HLA)-supported problem-solving environment that allows for detailed numerical study of industrial thin-film production in plasma-enhanced chemical vapor deposition (PECVD) reactors. They briefly describe

  9. Hierarchical Controlled Grid-Connected Microgrid based on a Novel Autonomous Current Sharing Controller

    DEFF Research Database (Denmark)

    Guan, Yajuan; Quintero, Juan Carlos Vasquez; Guerrero, Josep M.

    2015-01-01

    In this paper, a hierarchical control system based on a novel autonomous current sharing controller for grid-connected microgrids (MGs) is presented. A three-level hierarchical control system is implemented to guarantee the power sharing performance among voltage controlled parallel inverters......, while providing the required active and reactive power to the utility grid. A communication link is used to transmit the control signal from the tertiary and secondary control levels to the primary control. Simulation results from a MG based on two grid-connected parallel inverters are shown in order...

  10. Grid based calibration of SWAT hydrological models

    Directory of Open Access Journals (Sweden)

    D. Gorgan

    2012-07-01

    Full Text Available The calibration and execution of large hydrological models, such as SWAT (soil and water assessment tool, developed for large areas, high resolution, and huge input data, need not only quite a long execution time but also high computation resources. SWAT hydrological model supports studies and predictions of the impact of land management practices on water, sediment, and agricultural chemical yields in complex watersheds. The paper presents the gSWAT application as a web practical solution for environmental specialists to calibrate extensive hydrological models and to run scenarios, by hiding the complex control of processes and heterogeneous resources across the grid based high computation infrastructure. The paper highlights the basic functionalities of the gSWAT platform, and the features of the graphical user interface. The presentation is concerned with the development of working sessions, interactive control of calibration, direct and basic editing of parameters, process monitoring, and graphical and interactive visualization of the results. The experiments performed on different SWAT models and the obtained results argue the benefits brought by the grid parallel and distributed environment as a solution for the processing platform. All the instances of SWAT models used in the reported experiments have been developed through the enviroGRIDS project, targeting the Black Sea catchment area.

  11. Grid computing in pakistan and: opening to large hadron collider experiments

    International Nuclear Information System (INIS)

    Batool, N.; Osman, A.; Mahmood, A.; Rana, M.A.

    2009-01-01

    A grid computing facility was developed at sister institutes Pakistan Institute of Nuclear Science and Technology (PINSTECH) and Pakistan Institute of Engineering and Applied Sciences (PIEAS) in collaboration with Large Hadron Collider (LHC) Computing Grid during early years of the present decade. The Grid facility PAKGRID-LCG2 as one of the grid node in Pakistan was developed employing mainly local means and is capable of supporting local and international research and computational tasks in the domain of LHC Computing Grid. Functional status of the facility is presented in terms of number of jobs performed. The facility developed provides a forum to local researchers in the field of high energy physics to participate in the LHC experiments and related activities at European particle physics research laboratory (CERN), which is one of the best physics laboratories in the world. It also provides a platform of an emerging computing technology (CT). (author)

  12. Adaptation of Powerline Communications-Based Smart Metering Deployments to the Requirements of Smart Grids

    Directory of Open Access Journals (Sweden)

    Alberto Sendin

    2015-11-01

    Full Text Available Powerline communications (PLC-based smart meter deployments are now a reality in many regions of the world. Although PLC elements are generally incorporated in smart meters and data concentrators, the underlying PLC network allows the integration of other smart grid services directly over it. The remote control capabilities that automation programs need and are today deployed over their medium voltage (MV grid, can be extended to the low voltage (LV grid through these existing PLC networks. This paper demonstrates the capabilities of narrowband high data rate (NB HDR PLC technologies deployed over LV grids for smart metering purposes to support internet protocol internet protocol (IP communications in the LV grid. The paper demonstrates these possibilities with the presentation of the simulation and laboratory results of IP communications over international telecommunication union: ITU-T G.9904 PLC technology, and the definition of a PLC Network Management System based on a simple network management protocol (SNMP management information base (MIB definition and applicable use cases.

  13. The Knowledge Base Interface for Parametric Grid Information

    International Nuclear Information System (INIS)

    Hipp, James R.; Simons, Randall W.; Young, Chris J.

    1999-01-01

    The parametric grid capability of the Knowledge Base (KBase) provides an efficient robust way to store and access interpolatable information that is needed to monitor the Comprehensive Nuclear Test Ban Treaty. To meet both the accuracy and performance requirements of operational monitoring systems, we use an approach which combines the error estimation of kriging with the speed and robustness of Natural Neighbor Interpolation. The method involves three basic steps: data preparation, data storage, and data access. In past presentations we have discussed in detail the first step. In this paper we focus on the latter two, describing in detail the type of information which must be stored and the interface used to retrieve parametric grid data from the Knowledge Base. Once data have been properly prepared, the information (tessellation and associated value surfaces) needed to support the interface functionality, can be entered into the KBase. The primary types of parametric grid data that must be stored include (1) generic header information; (2) base model, station, and phase names and associated ID's used to construct surface identifiers; (3) surface accounting information; (4) tessellation accounting information; (5) mesh data for each tessellation; (6) correction data defined for each surface at each node of the surfaces owning tessellation (7) mesh refinement calculation set-up and flag information; and (8) kriging calculation set-up and flag information. The eight data components not only represent the results of the data preparation process but also include all required input information for several population tools that would enable the complete regeneration of the data results if that should be necessary

  14. Synchronization method for grid integrated battery storage systems during asymmetrical grid faults

    Directory of Open Access Journals (Sweden)

    Popadić Bane

    2017-01-01

    Full Text Available This paper aims at presenting a robust and reliable synchronization method for battery storage systems during asymmetrical grid faults. For this purpose, a Matlab/Simulink based model for testing of the power electronic interface between the grid and the battery storage systems has been developed. The synchronization method proposed in the paper is based on the proportional integral resonant controller with the delay signal cancellation. The validity of the synchronization method has been verified using the advanced laboratory station for the control of grid connected distributed energy sources. The proposed synchronization method has eliminated unfavourable components from the estimated grid angular frequency, leading to the more accurate and reliable tracking of the grid voltage vector positive sequence during both the normal operation and the operation during asymmetrical grid faults. [Project of the Serbian Ministry of Education, Science and Technological Development, Grant no. III 042004: Integrated and Interdisciplinary Research entitled: Smart Electricity Distribution Grids Based on Distribution Management System and Distributed Generation

  15. Advanced technologies for scalable ATLAS conditions database access on the grid

    CERN Document Server

    Basset, R; Dimitrov, G; Girone, M; Hawkings, R; Nevski, P; Valassi, A; Vaniachine, A; Viegas, F; Walker, R; Wong, A

    2010-01-01

    During massive data reprocessing operations an ATLAS Conditions Database application must support concurrent access from numerous ATLAS data processing jobs running on the Grid. By simulating realistic work-flow, ATLAS database scalability tests provided feedback for Conditions Db software optimization and allowed precise determination of required distributed database resources. In distributed data processing one must take into account the chaotic nature of Grid computing characterized by peak loads, which can be much higher than average access rates. To validate database performance at peak loads, we tested database scalability at very high concurrent jobs rates. This has been achieved through coordinated database stress tests performed in series of ATLAS reprocessing exercises at the Tier-1 sites. The goal of database stress tests is to detect scalability limits of the hardware deployed at the Tier-1 sites, so that the server overload conditions can be safely avoided in a production environment. Our analysi...

  16. Enhancing adaptive sparse grid approximations and improving refinement strategies using adjoint-based a posteriori error estimates

    Science.gov (United States)

    Jakeman, J. D.; Wildey, T.

    2015-01-01

    In this paper we present an algorithm for adaptive sparse grid approximations of quantities of interest computed from discretized partial differential equations. We use adjoint-based a posteriori error estimates of the physical discretization error and the interpolation error in the sparse grid to enhance the sparse grid approximation and to drive adaptivity of the sparse grid. Utilizing these error estimates provides significantly more accurate functional values for random samples of the sparse grid approximation. We also demonstrate that alternative refinement strategies based upon a posteriori error estimates can lead to further increases in accuracy in the approximation over traditional hierarchical surplus based strategies. Throughout this paper we also provide and test a framework for balancing the physical discretization error with the stochastic interpolation error of the enhanced sparse grid approximation.

  17. Enhancing adaptive sparse grid approximations and improving refinement strategies using adjoint-based a posteriori error estimates

    International Nuclear Information System (INIS)

    Jakeman, J.D.; Wildey, T.

    2015-01-01

    In this paper we present an algorithm for adaptive sparse grid approximations of quantities of interest computed from discretized partial differential equations. We use adjoint-based a posteriori error estimates of the physical discretization error and the interpolation error in the sparse grid to enhance the sparse grid approximation and to drive adaptivity of the sparse grid. Utilizing these error estimates provides significantly more accurate functional values for random samples of the sparse grid approximation. We also demonstrate that alternative refinement strategies based upon a posteriori error estimates can lead to further increases in accuracy in the approximation over traditional hierarchical surplus based strategies. Throughout this paper we also provide and test a framework for balancing the physical discretization error with the stochastic interpolation error of the enhanced sparse grid approximation

  18. Control strategy based on SPWM switching patterns for grid connected photovoltaic inverter

    Science.gov (United States)

    Hassaine, L.; Mraoui, A.

    2017-02-01

    Generally, for lower installation of photovoltaic systems connected to the grid, pulse width modulation (PWM) is a widely used technique for controlling the voltage source inverters injects currents into the grid. The current injected must be sinusoidal with reduced harmonic distortion. In this paper, a digital implementation of a control strategy based on PWM switching patterns for an inverter for photovoltaic system connected to the grid is presented. This strategy synchronize a sinusoidal inverter output current with a grid voltage The digital implementation of the proposed PWM switching pattern when is compared with the conventional one exhibit the advantage: Simplicity, reduction of the memory requirements and power calculation for the control

  19. Integration of the Chinese HPC Grid in ATLAS Distributed Computing

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00081160; The ATLAS collaboration

    2016-01-01

    Fifteen Chinese High Performance Computing sites, many of them on the TOP500 list of most powerful supercomputers, are integrated into a common infrastructure providing coherent access to a user through an interface based on a RESTful interface called SCEAPI. These resources have been integrated into the ATLAS Grid production system using a bridge between ATLAS and SCEAPI which translates the authorization and job submission protocols between the two environments. The ARC Computing Element (ARC CE) forms the bridge using an extended batch system interface to allow job submission to SCEAPI. The ARC CE was setup at the Institute for High Energy Physics, Beijing, in order to be as close as possible to the SCEAPI front-end interface at the Computing Network Information Center, also in Beijing. This paper describes the technical details of the integration between ARC CE and SCEAPI and presents results so far with two supercomputer centers, Tianhe-IA and ERA. These two centers have been the pilots for ATLAS Monte C...

  20. Integration of the Chinese HPC Grid in ATLAS Distributed Computing

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00081160

    2017-01-01

    Fifteen Chinese High-Performance Computing sites, many of them on the TOP500 list of most powerful supercomputers, are integrated into a common infrastructure providing coherent access to a user through an interface based on a RESTful interface called SCEAPI. These resources have been integrated into the ATLAS Grid production system using a bridge between ATLAS and SCEAPI which translates the authorization and job submission protocols between the two environments. The ARC Computing Element (ARC-CE) forms the bridge using an extended batch system interface to allow job submission to SCEAPI. The ARC-CE was setup at the Institute for High Energy Physics, Beijing, in order to be as close as possible to the SCEAPI front-end interface at the Computing Network Information Center, also in Beijing. This paper describes the technical details of the integration between ARC-CE and SCEAPI and presents results so far with two supercomputer centers, Tianhe-IA and ERA. These two centers have been the pilots for ATLAS Monte C...

  1. The HEPiX Virtualisation Working Group: Towards a “Grid of Clouds”

    CERN Multimedia

    CERN. Geneva

    2012-01-01

    The HEPiX Virtualisation Working Group has sponsored the development of policies and technologies that permit Grid sites to safely instantiate remotely generated virtual machine images confident in the knowledge that they will be able to meet their obligations, most notably in terms of guaranteeing the accountability and traceability of any Grid Job activity at their site. We will present the current status of the HEPiX Virtualisation Working Group technology and or links to related projects, notably StratusLab. We will also comment on the utility of our work in enabling a move from a Grid environment to a “Grid of Clouds” to provide a more responsive service to end users and reduce the service management load at participating sites.

  2. Does low job satisfaction lead to job mobility?

    DEFF Research Database (Denmark)

    Kristensen, Nicolai; Westergård-Nielsen, Niels Chr.

    This paper seeks to analyse the role of job satisfaction and actual job change behaviour. The analysis is based on the European Community Household Panel (ECHP) data for Danish families 1994-2000. The results show that inclusion of job satisfaction, which is a subjective measure, does improve...... the ability to predict actual quit behaviour: Low overall job satisfaction significantly increases the probability of quit. Various job satisfaction domains are ranked according to their ability to predict quits. Satisfaction with Type of Work is found to be the most important job characteristic while...... satisfaction with Job Security is found to be insignificant. These results hold across age, gender and education sub-groups and are opposed to results for UK, where job security is found to be the most important job domain. This discrepancy between UK and Denmark might be due to differences in unemployment...

  3. An Efficient Topology-Based Algorithm for Transient Analysis of Power Grid

    KAUST Repository

    Yang, Lan

    2015-08-10

    In the design flow of integrated circuits, chip-level verification is an important step that sanity checks the performance is as expected. Power grid verification is one of the most expensive and time-consuming steps of chip-level verification, due to its extremely large size. Efficient power grid analysis technology is highly demanded as it saves computing resources and enables faster iteration. In this paper, a topology-base power grid transient analysis algorithm is proposed. Nodal analysis is adopted to analyze the topology which is mathematically equivalent to iteratively solving a positive semi-definite linear equation. The convergence of the method is proved.

  4. Power Grid Construction Project Portfolio Optimization Based on Bi-level programming model

    Science.gov (United States)

    Zhao, Erdong; Li, Shangqi

    2017-08-01

    As the main body of power grid operation, county-level power supply enterprises undertake an important emission to guarantee the security of power grid operation and safeguard social power using order. The optimization of grid construction projects has been a key issue of power supply capacity and service level of grid enterprises. According to the actual situation of power grid construction project optimization of county-level power enterprises, on the basis of qualitative analysis of the projects, this paper builds a Bi-level programming model based on quantitative analysis. The upper layer of the model is the target restriction of the optimal portfolio; the lower layer of the model is enterprises’ financial restrictions on the size of the enterprise project portfolio. Finally, using a real example to illustrate operation proceeding and the optimization result of the model. Through qualitative analysis and quantitative analysis, the bi-level programming model improves the accuracy and normative standardization of power grid enterprises projects.

  5. Distributed computing grid experiences in CMS

    CERN Document Server

    Andreeva, Julia; Barrass, T; Bonacorsi, D; Bunn, Julian; Capiluppi, P; Corvo, M; Darmenov, N; De Filippis, N; Donno, F; Donvito, G; Eulisse, G; Fanfani, A; Fanzago, F; Filine, A; Grandi, C; Hernández, J M; Innocente, V; Jan, A; Lacaprara, S; Legrand, I; Metson, S; Newbold, D; Newman, H; Pierro, A; Silvestris, L; Steenberg, C; Stockinger, H; Taylor, Lucas; Thomas, M; Tuura, L; Van Lingen, F; Wildish, Tony

    2005-01-01

    The CMS experiment is currently developing a computing system capable of serving, processing and archiving the large number of events that will be generated when the CMS detector starts taking data. During 2004 CMS undertook a large scale data challenge to demonstrate the ability of the CMS computing system to cope with a sustained data- taking rate equivalent to 25% of startup rate. Its goals were: to run CMS event reconstruction at CERN for a sustained period at 25 Hz input rate; to distribute the data to several regional centers; and enable data access at those centers for analysis. Grid middleware was utilized to help complete all aspects of the challenge. To continue to provide scalable access from anywhere in the world to the data, CMS is developing a layer of software that uses Grid tools to gain access to data and resources, and that aims to provide physicists with a user friendly interface for submitting their analysis jobs. This paper describes the data challenge experience with Grid infrastructure ...

  6. 75 FR 24990 - Proposed Information Collection for the Evaluation of the Community-Based Job Training Grants...

    Science.gov (United States)

    2010-05-06

    ... Evaluation of the Community-Based Job Training Grants; Comment Request AGENCY: Employment and Training...- Based Job Training Grants. A copy of the proposed information collection request can be obtained by...-Based Job Training Grants (CBJTG) program is sponsored by ETA as an investment in building the capacity...

  7. The eGo grid model: An open-source and open-data based synthetic medium-voltage grid model for distribution power supply systems

    Science.gov (United States)

    Amme, J.; Pleßmann, G.; Bühler, J.; Hülk, L.; Kötter, E.; Schwaegerl, P.

    2018-02-01

    The increasing integration of renewable energy into the electricity supply system creates new challenges for distribution grids. The planning and operation of distribution systems requires appropriate grid models that consider the heterogeneity of existing grids. In this paper, we describe a novel method to generate synthetic medium-voltage (MV) grids, which we applied in our DIstribution Network GeneratOr (DINGO). DINGO is open-source software and uses freely available data. Medium-voltage grid topologies are synthesized based on location and electricity demand in defined demand areas. For this purpose, we use GIS data containing demand areas with high-resolution spatial data on physical properties, land use, energy, and demography. The grid topology is treated as a capacitated vehicle routing problem (CVRP) combined with a local search metaheuristics. We also consider the current planning principles for MV distribution networks, paying special attention to line congestion and voltage limit violations. In the modelling process, we included power flow calculations for validation. The resulting grid model datasets contain 3608 synthetic MV grids in high resolution, covering all of Germany and taking local characteristics into account. We compared the modelled networks with real network data. In terms of number of transformers and total cable length, we conclude that the method presented in this paper generates realistic grids that could be used to implement a cost-optimised electrical energy system.

  8. Operation of an InGrid based X-ray detector at the CAST experiment

    Science.gov (United States)

    Krieger, Christoph; Desch, Klaus; Kaminski, Jochen; Lupberger, Michael

    2018-02-01

    The CERN Axion Solar Telescope (CAST) is searching for axions and other particles which could be candidates for DarkMatter and even Dark Energy. These particles could be produced in the Sun and detected by a conversion into soft X-ray photons inside a strong magnetic field. In order to increase the sensitivity for physics beyond the Standard Model, detectors with a threshold below 1 keV as well as efficient background rejection methods are required to compensate for low energies and weak couplings resulting in very low detection rates. Those criteria are fulfilled by a detector utilizing the combination of a pixelized readout chip with an integrated Micromegas stage. These InGrid (Integrated Grid) devices can be build by photolithographic postprocessing techniques, resulting in a close to perfect match of grid and pixels facilitating the detection of single electrons on the chip surface. The high spatial resolution allows for energy determination by simple electron counting as well as for an event-shape based analysis as background rejection method. Tests at an X-ray generator revealed the energy threshold of an InGrid based X-ray detector to be well below the carbon Kα line at 277 eV. After the successful demonstration of the detectors key features, the detector was mounted at one of CAST's four detector stations behind an X-ray telescope in 2014. After several months of successful operation without any detector related interruptions, the InGrid based X-ray detector continues data taking at CAST in 2015. During operation at the experiment, background rates in the order of 10-5 keV-1 cm-2 s-1 have been achieved by application of a likelihood based method discriminating the non-photon background originating mostly from cosmic rays. For continued operation in 2016, an upgraded InGrid based detector is to be installed among other improvements including decoupling and sampling of the signal induced on the grid as well as a veto scintillator to further lower the

  9. Application of Job Demands-Resources model in research on relationships between job satisfaction, job resources, individual resources and job demands

    Directory of Open Access Journals (Sweden)

    Adrianna Potocka

    2013-04-01

    Full Text Available Background: The aim of this study was to explore the relationships between job demands, job resourses, personal resourses and job satisfaction and to assess the usefulness of the Job Demands-Resources (JD-R model in the explanation of these phenomena. Materials and Methods: The research was based on a sample of 500 social workers. The "Psychosocial Factors" and "Job satisfaction" questionnaires were used to test the hypothesis. Results: The results showed that job satisfaction increased with increasing job accessibility and personal resources (r = 0.44; r = 0.31; p < 0.05. The analysis of variance (ANOVA indicated that job resources and job demands [F(1.474 = 4.004; F(1.474 = 4.166; p < 0.05] were statistically significant sources of variation in job satisfaction. Moreover, interactions between job demands and job resources [F(3,474 = 2.748; p < 0.05], as well as between job demands and personal resources [F(3.474 = 3.021; p < 0.05] had a significant impact on job satisfaction. The post hoc tests showed that 1 in low job demands, but high job resources employees declared higher job satisfaction, than those who perceived them as medium (p = 0.0001 or low (p = 0.0157; 2 when the level of job demands was perceived as medium, employees with high personal resources declared significantly higher job satisfaction than those with low personal resources (p = 0.0001. Conclusion: The JD-R model can be used to investigate job satisfaction. Taking into account fundamental factors of this model, in organizational management there are possibilities of shaping job satisfaction among employees. Med Pr 2013;64(2:217–225

  10. Design of Energy Storage Management System Based on FPGA in Micro-Grid

    Science.gov (United States)

    Liang, Yafeng; Wang, Yanping; Han, Dexiao

    2018-01-01

    Energy storage system is the core to maintain the stable operation of smart micro-grid. Aiming at the existing problems of the energy storage management system in the micro-grid such as Low fault tolerance, easy to cause fluctuations in micro-grid, a new intelligent battery management system based on field programmable gate array is proposed : taking advantage of FPGA to combine the battery management system with the intelligent micro-grid control strategy. Finally, aiming at the problem that during estimation of battery charge State by neural network, initialization of weights and thresholds are not accurate leading to large errors in prediction results, the genetic algorithm is proposed to optimize the neural network method, and the experimental simulation is carried out. The experimental results show that the algorithm has high precision and provides guarantee for the stable operation of micro-grid.

  11. Development and verification of remote research environment based on 'Fusion research grid'

    International Nuclear Information System (INIS)

    Iba, Katsuyuki; Ozeki, Takahisa; Totsuka, Toshiyuki; Suzuki, Yoshio; Oshima, Takayuki; Sakata, Shinya; Sato, Minoru; Suzuki, Mitsuhiro; Hamamatsu, Kiyotaka; Kiyono, Kimihiro

    2008-01-01

    'Fusion research grid' is a concept that unites scientists and let them collaborate effectively against their difference in time zone and location in a nuclear fusion research. Fundamental technologies of 'Fusion research grid' have been developed at JAEA in the VizGrid project under the e-Japan project at the Ministry of Education, Culture, Sports, Science and Technology (MEXT). We are conscious of needs to create new systems that assist researchers with their research activities because remote collaborations have been increasing in international projects. Therefore we have developed prototype remote research environments for experiments, diagnostics, analyses and communications based on 'Fusion research grid'. All users can access these environments from anywhere because 'Fusion research grid' does not require a closed network like Super SINET to maintain security. The prototype systems were verified in experiments at JT-60U and their availability was confirmed

  12. Micro-grid platform based on NODE.JS architecture, implemented in electrical network instrumentation

    Science.gov (United States)

    Duque, M.; Cando, E.; Aguinaga, A.; Llulluna, F.; Jara, N.; Moreno, T.

    2016-05-01

    In this document, I propose a theory about the impact of systems based on microgrids in non-industrialized countries that have the goal to improve energy exploitation through alternatives methods of a clean and renewable energy generation and the creation of the app to manage the behavior of the micro-grids based on the NodeJS, Django and IOJS technologies. The micro-grids allow the optimal way to manage energy flow by electric injection directly in electric network small urban's cells in a low cost and available way. In difference from conventional systems, micro-grids can communicate between them to carry energy to places that have higher demand in accurate moments. This system does not require energy storage, so, costs are lower than conventional systems like fuel cells, solar panels or else; even though micro-grids are independent systems, they are not isolated. The impact that this analysis will generate, is the improvement of the electrical network without having greater control than an intelligent network (SMART-GRID); this leads to move to a 20% increase in energy use in a specified network; that suggest there are others sources of energy generation; but for today's needs, we need to standardize methods and remain in place to support all future technologies and the best option are the Smart Grids and Micro-Grids.

  13. iSERVO: Implementing the International Solid Earth Research Virtual Observatory by Integrating Computational Grid and Geographical Information Web Services

    Science.gov (United States)

    Aktas, Mehmet; Aydin, Galip; Donnellan, Andrea; Fox, Geoffrey; Granat, Robert; Grant, Lisa; Lyzenga, Greg; McLeod, Dennis; Pallickara, Shrideep; Parker, Jay; Pierce, Marlon; Rundle, John; Sayar, Ahmet; Tullis, Terry

    2006-12-01

    We describe the goals and initial implementation of the International Solid Earth Virtual Observatory (iSERVO). This system is built using a Web Services approach to Grid computing infrastructure and is accessed via a component-based Web portal user interface. We describe our implementations of services used by this system, including Geographical Information System (GIS)-based data grid services for accessing remote data repositories and job management services for controlling multiple execution steps. iSERVO is an example of a larger trend to build globally scalable scientific computing infrastructures using the Service Oriented Architecture approach. Adoption of this approach raises a number of research challenges in millisecond-latency message systems suitable for internet-enabled scientific applications. We review our research in these areas.

  14. Use of DAGMan in CRAB3 to Improve the Splitting of CMS User Jobs

    Energy Technology Data Exchange (ETDEWEB)

    Wolf, M. [Notre Dame U.; Mascheroni, M. [Fermilab; Woodard, A. [Notre Dame U.; Belforte, S. [INFN, Trieste; Bockelman, B. [Nebraska U.; Hernandez, J. M. [Madrid, CIEMAT; Vaandering, E. [Fermilab

    2017-11-22

    CRAB3 is a workload management tool used by CMS physicists to analyze data acquired by the Compact Muon Solenoid (CMS) detector at the CERN Large Hadron Collider (LHC). Research in high energy physics often requires the analysis of large collections of files, referred to as datasets. The task is divided into jobs that are distributed among a large collection of worker nodes throughout the Worldwide LHC Computing Grid (WLCG). Splitting a large analysis task into optimally sized jobs is critical to efficient use of distributed computing resources. Jobs that are too big will have excessive runtimes and will not distribute the work across all of the available nodes. However, splitting the project into a large number of very small jobs is also inefficient, as each job creates additional overhead which increases load on infrastructure resources. Currently this splitting is done manually, using parameters provided by the user. However the resources needed for each job are difficult to predict because of frequent variations in the performance of the user code and the content of the input dataset. As a result, dividing a task into jobs by hand is difficult and often suboptimal. In this work we present a new feature called “automatic splitting” which removes the need for users to manually specify job splitting parameters. We discuss how HTCondor DAGMan can be used to build dynamic Directed Acyclic Graphs (DAGs) to optimize the performance of large CMS analysis jobs on the Grid. We use DAGMan to dynamically generate interconnected DAGs that estimate the processing time the user code will require to analyze each event. This is used to calculate an estimate of the total processing time per job, and a set of analysis jobs are run using this estimate as a specified time limit. Some jobs may not finish within the alloted time; they are terminated at the time limit, and the unfinished data is regrouped into smaller jobs and resubmitted.

  15. Use of DAGMan in CRAB3 to improve the splitting of CMS user jobs

    Science.gov (United States)

    Wolf, M.; Mascheroni, M.; Woodard, A.; Belforte, S.; Bockelman, B.; Hernandez, J. M.; Vaandering, E.

    2017-10-01

    CRAB3 is a workload management tool used by CMS physicists to analyze data acquired by the Compact Muon Solenoid (CMS) detector at the CERN Large Hadron Collider (LHC). Research in high energy physics often requires the analysis of large collections of files, referred to as datasets. The task is divided into jobs that are distributed among a large collection of worker nodes throughout the Worldwide LHC Computing Grid (WLCG). Splitting a large analysis task into optimally sized jobs is critical to efficient use of distributed computing resources. Jobs that are too big will have excessive runtimes and will not distribute the work across all of the available nodes. However, splitting the project into a large number of very small jobs is also inefficient, as each job creates additional overhead which increases load on infrastructure resources. Currently this splitting is done manually, using parameters provided by the user. However the resources needed for each job are difficult to predict because of frequent variations in the performance of the user code and the content of the input dataset. As a result, dividing a task into jobs by hand is difficult and often suboptimal. In this work we present a new feature called “automatic splitting” which removes the need for users to manually specify job splitting parameters. We discuss how HTCondor DAGMan can be used to build dynamic Directed Acyclic Graphs (DAGs) to optimize the performance of large CMS analysis jobs on the Grid. We use DAGMan to dynamically generate interconnected DAGs that estimate the processing time the user code will require to analyze each event. This is used to calculate an estimate of the total processing time per job, and a set of analysis jobs are run using this estimate as a specified time limit. Some jobs may not finish within the alloted time; they are terminated at the time limit, and the unfinished data is regrouped into smaller jobs and resubmitted.

  16. AliEn: ALICE environment on the GRID

    International Nuclear Information System (INIS)

    Bagnasco, S; Betev, L; Buncic, P; Carminati, F; Cirstoiu, C; Grigoras, C; Hayrapetyan, A; Harutyunyan, A; Peters, A J; Saiz, P

    2008-01-01

    Starting from mid-2008, the ALICE detector at CERN LHC will collect data at a rate of 4PB per year. ALICE will use exclusively distributed Grid resources to store, process and analyse this data. The top-level management of the Grid resources is done through the AliEn (ALICE Environment) system, which is in continuous development since year 2000. AliEn presents several original solutions, which have shown their viability in a number of large exercises of increasing complexity called Data Challenges. This paper describes the AliEn architecture: Job Management, Data Management and UI. The current status of AliEn will be illustrated, as well as the performance of the system during the data challenges. The paper also describes the future AliEn development roadmap

  17. AliEn: ALICE Environment on the GRID

    CERN Multimedia

    Bagnasco, S; Buncic, P; Carminati, F; Cirstoiu, C; Grigoras, C; Hayrapetyan, A; Harutyunyan, A; Peters, A J; Saiz, P

    2007-01-01

    Starting from mid-2008, the ALICE detector at CERN LHC will collect data at a rate of 4PB per year. ALICE will use exclusively distributed Grid resources to store, process and analyse this data. The top-level management of the Grid resources is done through the AliEn (ALICE Environment) system, which is in continuous development since year 2000. AliEn presents several original solutions, which have shown their viability in a number of large exercises of increasing complexity called Data Challenges. This paper describes the AliEn architecture: Job Management, Data Management and UI. The current status of AliEn will be illustrated, as well as the performance of the system during the data challenges. The paper also describes the future AliEn development roadmap.

  18. Simulation-Based Approach for Studying the Balancing of Local Smart Grids with Electric Vehicle Batteries

    Directory of Open Access Journals (Sweden)

    Juhani Latvakoski

    2015-07-01

    Full Text Available Modern society is facing great challenges due to pollution and increased carbon dioxide (CO2 emissions. As part of solving these challenges, the use of renewable energy sources and electric vehicles (EVs is rapidly increasing. However, increased dynamics have triggered problems in balancing energy supply and consumption demand in the power systems. The resulting uncertainty and unpredictability of energy production, consumption, and management of peak loads has caused an increase in costs for energy market actors. Therefore, the means for studying the balancing of local smart grids with EVs is a starting point for this paper. The main contribution is a simulation-based approach which was developed to enable the study of the balancing of local distribution grids with EV batteries in a cost-efficient manner. The simulation-based approach is applied to enable the execution of a distributed system with the simulation of a local distribution grid, including a number of charging stations and EVs. A simulation system has been constructed to support the simulation-based approach. The evaluation has been carried out by executing the scenario related to balancing local distribution grids with EV batteries in a step-by-step manner. The evaluation results indicate that the simulation-based approach is able to facilitate the evaluation of smart grid– and EV-related communication protocols, control algorithms for charging, and functionalities of local distribution grids as part of a complex, critical cyber-physical system. In addition, the simulation system is able to incorporate advanced methods for monitoring, controlling, tracking, and modeling behavior. The simulation model of the local distribution grid can be executed with the smart control of charging and discharging powers of the EVs according to the load situation in the local distribution grid. The resulting simulation system can be applied to the study of balancing local smart grids with EV

  19. Croatia's rural areas - renewable energy based electricity generation for isolated grids

    Directory of Open Access Journals (Sweden)

    Protic Sonja Maria

    2014-01-01

    Full Text Available Several Western Balkan states face the consequences of the Yugoslavian war, which left hometowns with dilapidated electricity grid connections, a high average age of power plant capacities and low integration of renewable energy sources, grid bottlenecks and a lack of competition. In order to supply all households with electricity, UNDP Croatia did a research on decentralized supply systems based on renewable energy sources. Decentralized supply systems offer cheaper electricity connections and provide faster support to rural development. This paper proposes a developed methodology to financially compare isolated grid solutions that primarily use renewable energies to an extension of the public electricity network to small regions in Croatia. Isolated grid supply proves to be very often a preferable option. Furthermore, it points out the lack of a reliable evaluation of non-monetizable aspects and promotes a new interdisciplinary approach.

  20. Jobs to Manufacturing Careers: Work-Based Courses. Work-Based Learning in Action

    Science.gov (United States)

    Kobes, Deborah

    2016-01-01

    This case study, one of a series of publications exploring effective and inclusive models of work-based learning, finds that work-based courses bring college to the production line by using the job as a learning lab. Work-based courses are an innovative way to give incumbent workers access to community college credits and degrees. They are…

  1. Camera Coverage Estimation Based on Multistage Grid Subdivision

    Directory of Open Access Journals (Sweden)

    Meizhen Wang

    2017-04-01

    Full Text Available Visual coverage is one of the most important quality indexes for depicting the usability of an individual camera or camera network. It is the basis for camera network deployment, placement, coverage-enhancement, planning, etc. Precision and efficiency are critical influences on applications, especially those involving several cameras. This paper proposes a new method to efficiently estimate superior camera coverage. First, the geographic area that is covered by the camera and its minimum bounding rectangle (MBR without considering obstacles is computed using the camera parameters. Second, the MBR is divided into grids using the initial grid size. The status of the four corners of each grid is estimated by a line of sight (LOS algorithm. If the camera, considering obstacles, covers a corner, the status is represented by 1, otherwise by 0. Consequently, the status of a grid can be represented by a code that is a combination of 0s or 1s. If the code is not homogeneous (not four 0s or four 1s, the grid will be divided into four sub-grids until the sub-grids are divided into a specific maximum level or their codes are homogeneous. Finally, after performing the process above, total camera coverage is estimated according to the size and status of all grids. Experimental results illustrate that the proposed method’s accuracy is determined by the method that divided the coverage area into the smallest grids at the maximum level, while its efficacy is closer to the method that divided the coverage area into the initial grids. It considers both efficiency and accuracy. The initial grid size and maximum level are two critical influences on the proposed method, which can be determined by weighing efficiency and accuracy.

  2. Geospatial Information Service System Based on GeoSOT Grid & Encoding

    Directory of Open Access Journals (Sweden)

    LI Shizhong

    2016-12-01

    Full Text Available With the rapid development of the space and earth observation technology, it is important to establish a multi-source, multi-scale and unified cross-platform reference for global data. In practice, the production and maintenance of geospatial data are scattered in different units, and the standard of the data grid varies between departments and systems. All these bring out the disunity of standards among different historical periods or orgnizations. Aiming at geospatial information security library for the national high resolution earth observation, there are some demands for global display, associated retrieval and template applications and other integrated services for geospatial data. Based on GeoSOT grid and encoding theory system, "geospatial information security library information of globally unified grid encoding management" data subdivision organization solutions have been proposed; system-level analyses, researches and designs have been carried out. The experimental results show that the data organization and management method based on GeoSOT can significantly improve the overall efficiency of the geospatial information security service system.

  3. Predicting Job Stress Based on Elements of Coping Styles in Nurses

    Directory of Open Access Journals (Sweden)

    Mansoureh Nezari Sedeh

    2016-07-01

    Full Text Available Using coping methods can help to dominate on physical, mental, and social relationships, individual contradiction problems, and can be considered as one of effective factors in general and mental health of nurses. The objective of the present research is predicting job stress based on elements of coping styles in nurses. By correlative methodology for this research, 120 female20-45 years old nurses in Tehran city were selected by simple random sampling method based on Cochran formula. The research instrument includes job stress questionnaire and coping style questionnaire of Lazzarus & Folkman; the Pearson correlation coefficient test, and linear regression were used to test hypotheses and generalize the obtained information from tests. Findings showed that participants’ scores were near normal range and Cronbach’s alpha coefficient was 0.58 which indicated scores internal consistency. The obtained results showed that coping elements in 0.05 significant level with f-value of 12.403 significantly predicted job stress. In addition, regression coefficient among support, responsibility, and managerial solution elements was negative and positive with two other relationships including job stress and escape-avoidance. Therefore, it can be concluded that elements of support, responsibility, escape-avoidance, and managerial solution significantly predict nurses’ job stress among coping elements.

  4. Project Scheduling Heuristics-Based Standard PSO for Task-Resource Assignment in Heterogeneous Grid

    OpenAIRE

    Chen, Ruey-Maw; Wang, Chuin-Mu

    2011-01-01

    The task scheduling problem has been widely studied for assigning resources to tasks in heterogeneous grid environment. Effective task scheduling is an important issue for the performance of grid computing. Meanwhile, the task scheduling problem is an NP-complete problem. Hence, this investigation introduces a named “standard“ particle swarm optimization (PSO) metaheuristic approach to efficiently solve the task scheduling problems in grid. Meanwhile, two promising heuristics based on multimo...

  5. Intelligent Control of Micro Grid: A Big Data-Based Control Center

    Science.gov (United States)

    Liu, Lu; Wang, Yanping; Liu, Li; Wang, Zhiseng

    2018-01-01

    In this paper, a structure of micro grid system with big data-based control center is introduced. Energy data from distributed generation, storage and load are analized through the control center, and from the results new trends will be predicted and applied as a feedback to optimize the control. Therefore, each step proceeded in micro grid can be adjusted and orgnized in a form of comprehensive management. A framework of real-time data collection, data processing and data analysis will be proposed by employing big data technology. Consequently, a integrated distributed generation and a optimized energy storage and transmission process can be implemented in the micro grid system.

  6. McRunjob: A High Energy Physics Workflow Planner for Grid Production Processing

    OpenAIRE

    Graham, G E; Evans, D; Bertram, I

    2003-01-01

    McRunjob is a powerful grid workflow manager used to manage the generation of large numbers of production processing jobs in High Energy Physics. In use at both the DZero and CMS experiments, McRunjob has been used to manage large Monte Carlo production processing since 1999 and is being extended to uses in regular production processing for analysis and reconstruction. Described at CHEP 2001, McRunjob converts core metadata into jobs submittable in a variety of environments. The powerful core...

  7. A roadmap for caGrid, an enterprise Grid architecture for biomedical research.

    Science.gov (United States)

    Saltz, Joel; Hastings, Shannon; Langella, Stephen; Oster, Scott; Kurc, Tahsin; Payne, Philip; Ferreira, Renato; Plale, Beth; Goble, Carole; Ervin, David; Sharma, Ashish; Pan, Tony; Permar, Justin; Brezany, Peter; Siebenlist, Frank; Madduri, Ravi; Foster, Ian; Shanbhag, Krishnakant; Mead, Charlie; Chue Hong, Neil

    2008-01-01

    caGrid is a middleware system which combines the Grid computing, the service oriented architecture, and the model driven architecture paradigms to support development of interoperable data and analytical resources and federation of such resources in a Grid environment. The functionality provided by caGrid is an essential and integral component of the cancer Biomedical Informatics Grid (caBIG) program. This program is established by the National Cancer Institute as a nationwide effort to develop enabling informatics technologies for collaborative, multi-institutional biomedical research with the overarching goal of accelerating translational cancer research. Although the main application domain for caGrid is cancer research, the infrastructure provides a generic framework that can be employed in other biomedical research and healthcare domains. The development of caGrid is an ongoing effort, adding new functionality and improvements based on feedback and use cases from the community. This paper provides an overview of potential future architecture and tooling directions and areas of improvement for caGrid and caGrid-like systems. This summary is based on discussions at a roadmap workshop held in February with participants from biomedical research, Grid computing, and high performance computing communities.

  8. Asymmetrical Grid Fault Ride-Through Strategy of Three-phase Grid-connected Inverter Considering Network Impedance Impact in Low Voltage Grid

    DEFF Research Database (Denmark)

    Guo, Xiaoqiang; Zhang, Xue; Wang, Baocheng

    2014-01-01

    This letter presents a new control strategy of threephase grid-connected inverter for the positive sequence voltage recovery and negative sequence voltage reduction under asymmetrical grid faults. Unlike the conventional control strategy based on an assumption that the network impedance is mainly...... of the proposed solution for the flexible voltage support in a low-voltage grid, where thenetwork impedance is mainly resistive.......This letter presents a new control strategy of threephase grid-connected inverter for the positive sequence voltage recovery and negative sequence voltage reduction under asymmetrical grid faults. Unlike the conventional control strategy based on an assumption that the network impedance is mainly...... inductive, the proposed control strategy is more flexible and effective by considering the network impedance impact, which is of great importance for the high penetration of grid-connected renewable energy systems into low-voltage grids. The experimental tests are carried out to validate the effectiveness...

  9. Photovoltaic Grid-Connected Modeling and Characterization Based on Experimental Results.

    Science.gov (United States)

    Humada, Ali M; Hojabri, Mojgan; Sulaiman, Mohd Herwan Bin; Hamada, Hussein M; Ahmed, Mushtaq N

    2016-01-01

    A grid-connected photovoltaic (PV) system operates under fluctuated weather condition has been modeled and characterized based on specific test bed. A mathematical model of a small-scale PV system has been developed mainly for residential usage, and the potential results have been simulated. The proposed PV model based on three PV parameters, which are the photocurrent, IL, the reverse diode saturation current, Io, the ideality factor of diode, n. Accuracy of the proposed model and its parameters evaluated based on different benchmarks. The results showed that the proposed model fitting the experimental results with high accuracy compare to the other models, as well as the I-V characteristic curve. The results of this study can be considered valuable in terms of the installation of a grid-connected PV system in fluctuated climatic conditions.

  10. Combining Market-Based Control with Distribution Grid Constraints when Coordinating Electric Vehicle Charging

    Directory of Open Access Journals (Sweden)

    Geert Deconinck

    2015-12-01

    Full Text Available The charging of electric vehicles (EVs impacts the distribution grid, and its cost depends on the price of electricity when charging. An aggregator that is responsible for a large fleet of EVs can use a market-based control algorithm to coordinate the charging of these vehicles, in order to minimize the costs. In such an optimization, the operational parameters of the distribution grid, to which the EVs are connected, are not considered. This can lead to violations of the technical constraints of the grid (e.g., under-voltage, phase unbalances; for example, because many vehicles start charging simultaneously when the price is low. An optimization that simultaneously takes the economic and technical aspects into account is complex, because it has to combine time-driven control at the market level with event-driven control at the operational level. Different case studies investigate under which circumstances the market-based control, which coordinates EV charging, conflicts with the operational constraints of the distribution grid. Especially in weak grids, phase unbalance and voltage issues arise with a high share of EVs. A low-level voltage droop controller at the charging point of the EV can be used to avoid many grid constraint violations, by reducing the charge power if the local voltage is too low. While this action implies a deviation from the cost-optimal operating point, it is shown that this has a very limited impact on the business case of an aggregator, and is able to comply with the technical distribution grid constraints, even in weak distribution grids with many EVs.

  11. Migration of Monte Carlo simulation of high energy atmospheric showers to GRID infrastructure

    Energy Technology Data Exchange (ETDEWEB)

    Vazquez, Adolfo; Contreras, Jose Luis [Grupo de Altas EnergIas Departamento de Fisica Atomica, Molecular y Nuclear Universidad Complutense de Madrid Avenida Complutense s/n, 28040 Madrid - Spain (Spain); Calle, Ignacio de la; Ibarra, Aitor; Tapiador, Daniel, E-mail: avazquez@gae.ucm.e [INSA. IngenierIa y Servicios Aeroespaciales S.A. Paseo Pintor Rosales 34, 28008 Madrid - Spain (Spain)

    2010-04-01

    A system to run Monte Carlo simulations on a Grid environment is presented. The architectural design proposed uses the current resources of the MAGIC Virtual Organization on EGEE and can be easily generalized to support the simulation of any similar experiment, such as that of the future European planned project, the Cherenkov Telescope Array. The proposed system is based on a Client/Server architecture, and provides the user with a single access point to the simulation environment through a remote graphical user interface, the Client. The Client can be accessed via web browser, using web service technology, with no additional software installation on the user side required. The Server processes the user request and uses a database for both data catalogue and job management inside the Grid. The design, first production tests and lessons learned from the system will be discussed here.

  12. Migration of Monte Carlo simulation of high energy atmospheric showers to GRID infrastructure

    International Nuclear Information System (INIS)

    Vazquez, Adolfo; Contreras, Jose Luis; Calle, Ignacio de la; Ibarra, Aitor; Tapiador, Daniel

    2010-01-01

    A system to run Monte Carlo simulations on a Grid environment is presented. The architectural design proposed uses the current resources of the MAGIC Virtual Organization on EGEE and can be easily generalized to support the simulation of any similar experiment, such as that of the future European planned project, the Cherenkov Telescope Array. The proposed system is based on a Client/Server architecture, and provides the user with a single access point to the simulation environment through a remote graphical user interface, the Client. The Client can be accessed via web browser, using web service technology, with no additional software installation on the user side required. The Server processes the user request and uses a database for both data catalogue and job management inside the Grid. The design, first production tests and lessons learned from the system will be discussed here.

  13. An adaptive multi-agent-based approach to smart grids control and optimization

    Energy Technology Data Exchange (ETDEWEB)

    Carvalho, Marco [Florida Institute of Technology, Melbourne, FL (United States); Perez, Carlos; Granados, Adrian [Institute for Human and Machine Cognition, Ocala, FL (United States)

    2012-03-15

    In this paper, we describe a reinforcement learning-based approach to power management in smart grids. The scenarios we consider are smart grid settings where renewable power sources (e.g. Photovoltaic panels) have unpredictable variations in power output due, for example, to weather or cloud transient effects. Our approach builds on a multi-agent system (MAS)-based infrastructure for the monitoring and coordination of smart grid environments with renewable power sources and configurable energy storage devices (battery banks). Software agents are responsible for tracking and reporting power flow variations at different points in the grid, and to optimally coordinate the engagement of battery banks (i.e. charge/idle/discharge modes) to maintain energy requirements to end-users. Agents are able to share information and coordinate control actions through a parallel communications infrastructure, and are also capable of learning, from experience, how to improve their response strategies for different operational conditions. In this paper we describe our approach and address some of the challenges associated with the communications infrastructure for distributed coordination. We also present some preliminary results of our first simulations using the GridLAB-D simulation environment, created by the US Department of Energy (DoE) at Pacific Northwest National Laboratory (PNNL). (orig.)

  14. Job-Preference and Job-Matching Assessment Results and Their Association with Job Performance and Satisfaction among Young Adults with Developmental Disabilities

    Science.gov (United States)

    Hall, Julie; Morgan, Robert L.; Salzberg, Charles L.

    2014-01-01

    We investigated the effects of preference and degree of match on job performance of four 19 to 20-year-old young adults with developmental disabilities placed in community-based job conditions. We identified high-preference, high-matched and low-preference, low-matched job tasks using a video web-based assessment program. The job matching…

  15. Wireless Sensor Network Based Smart Grid Communications: Cyber Attacks, Intrusion Detection System and Topology Control

    Directory of Open Access Journals (Sweden)

    Lipi Chhaya

    2017-01-01

    Full Text Available The existing power grid is going through a massive transformation. Smart grid technology is a radical approach for improvisation in prevailing power grid. Integration of electrical and communication infrastructure is inevitable for the deployment of Smart grid network. Smart grid technology is characterized by full duplex communication, automatic metering infrastructure, renewable energy integration, distribution automation and complete monitoring and control of entire power grid. Wireless sensor networks (WSNs are small micro electrical mechanical systems that are deployed to collect and communicate the data from surroundings. WSNs can be used for monitoring and control of smart grid assets. Security of wireless sensor based communication network is a major concern for researchers and developers. The limited processing capabilities of wireless sensor networks make them more vulnerable to cyber-attacks. The countermeasures against cyber-attacks must be less complex with an ability to offer confidentiality, data readiness and integrity. The address oriented design and development approach for usual communication network requires a paradigm shift to design data oriented WSN architecture. WSN security is an inevitable part of smart grid cyber security. This paper is expected to serve as a comprehensive assessment and analysis of communication standards, cyber security issues and solutions for WSN based smart grid infrastructure.

  16. High density grids

    Energy Technology Data Exchange (ETDEWEB)

    Cohen, Aina E.; Baxter, Elizabeth L.

    2018-01-16

    An X-ray data collection grid device is provided that includes a magnetic base that is compatible with robotic sample mounting systems used at synchrotron beamlines, a grid element fixedly attached to the magnetic base, where the grid element includes at least one sealable sample window disposed through a planar synchrotron-compatible material, where the planar synchrotron-compatible material includes at least one automated X-ray positioning and fluid handling robot fiducial mark.

  17. ARC Cache: A solution for lightweight Grid sites in ATLAS

    CERN Document Server

    Garonne, Vincent; The ATLAS collaboration

    2016-01-01

    Many Grid sites have the need to reduce operational manpower, and running a storage element consumes a large amount of effort. In addition, setting up a new Grid site including a storage element involves a steep learning curve and large investment of time. For these reasons so-called storage-less sites are becoming more popular as a way to provide Grid computing resources with less operational overhead. ARC CE is a widely-used and mature Grid middleware which was designed from the start to be used on sites with no persistent storage element. Instead, it maintains a local self-managing cache of data which retains popular data for future jobs. As the cache is simply an area on a local posix shared filesystem with no external-facing service, it requires no extra maintenance. The cache can be scaled up as required by increasing the size of the filesystem or adding new filesystems. This paper describes how ARC CE and its cache are an ideal solution for lightweight Grid sites in the ATLAS experiment, and the integr...

  18. Dynamic workload balancing of parallel applications with user-level scheduling on the Grid

    CERN Document Server

    Korkhov, Vladimir V; Krzhizhanovskaya, Valeria V

    2009-01-01

    This paper suggests a hybrid resource management approach for efficient parallel distributed computing on the Grid. It operates on both application and system levels, combining user-level job scheduling with dynamic workload balancing algorithm that automatically adapts a parallel application to the heterogeneous resources, based on the actual resource parameters and estimated requirements of the application. The hybrid environment and the algorithm for automated load balancing are described, the influence of resource heterogeneity level is measured, and the speedup achieved with this technique is demonstrated for different types of applications and resources.

  19. GSIMF: a web service based software and database management system for the next generation grids

    International Nuclear Information System (INIS)

    Wang, N; Ananthan, B; Gieraltowski, G; May, E; Vaniachine, A

    2008-01-01

    To process the vast amount of data from high energy physics experiments, physicists rely on Computational and Data Grids; yet, the distribution, installation, and updating of a myriad of different versions of different programs over the Grid environment is complicated, time-consuming, and error-prone. Our Grid Software Installation Management Framework (GSIMF) is a set of Grid Services that has been developed for managing versioned and interdependent software applications and file-based databases over the Grid infrastructure. This set of Grid services provide a mechanism to install software packages on distributed Grid computing elements, thus automating the software and database installation management process on behalf of the users. This enables users to remotely install programs and tap into the computing power provided by Grids

  20. A Hierarchical and Distributed Approach for Mapping Large Applications to Heterogeneous Grids using Genetic Algorithms

    Science.gov (United States)

    Sanyal, Soumya; Jain, Amit; Das, Sajal K.; Biswas, Rupak

    2003-01-01

    In this paper, we propose a distributed approach for mapping a single large application to a heterogeneous grid environment. To minimize the execution time of the parallel application, we distribute the mapping overhead to the available nodes of the grid. This approach not only provides a fast mapping of tasks to resources but is also scalable. We adopt a hierarchical grid model and accomplish the job of mapping tasks to this topology using a scheduler tree. Results show that our three-phase algorithm provides high quality mappings, and is fast and scalable.

  1. Mini-grid based off-grid electrification to enhance electricity access in developing countries: What policies may be required?

    International Nuclear Information System (INIS)

    Bhattacharyya, Subhes C.; Palit, Debajit

    2016-01-01

    With 1.2 billion people still lacking electricity access by 2013, electricity access remains a major global challenge. Although mini-grid based electrification has received attention in recent times, their full exploitation requires policy support covering a range of areas. Distilling the experience from a five year research project, OASYS South Asia, this paper presents the summary of research findings and shares the experience from four demonstration activities. It suggests that cost-effective universal electricity service remains a challenge and reaching the universal electrification target by 2030 will remain a challenge for the less developed countries. The financial, organisational and governance weaknesses hinder successful implementation of projects in many countries. The paper then provides 10 policy recommendations to promote mini-grids as a complementary route to grid extension to promote electricity access for successful outcomes. - Highlights: •The academic and action research activities undertaken through OASYS South Asia Project are reported. •Evidence produced through a multi-dimensional participatory framework supplemented by four demonstration projects. •Funding and regulatory challenges militate against universal electrification objectives by 2030. •Innovative business approaches linking local mini-grids and livelihood opportunities exist. •Enabling policies are suggested to exploit such options.

  2. Trusted data management for Grid-based medical applications

    NARCIS (Netherlands)

    van 't Noordende, G.J.; Olabarriaga, S.D.; Koot, M.R.; de Laat, C.T.A.M.; Udoh, E.

    2011-01-01

    Existing Grid technology has been foremost designed with performance and scalability in mind. When using Grid infrastructure for medical applications, privacy and security considerations become paramount. Privacy aspects require a re-thinking of the design and implementation of common Grid

  3. A Stationary Reference Frame Grid Synchronization System for Three-Phase Grid-Connected Power Converters Under Adverse Grid Conditions

    DEFF Research Database (Denmark)

    Rodríguez, P.; Luna, A.; Muñoz-Aguilar, R. S.

    2012-01-01

    synchronization method for three-phase three-wire networks, namely dual second-order generalized integrator (SOGI) frequency-locked loop. The method is based on two adaptive filters, implemented by using a SOGI on the stationary αβ reference frame, and it is able to perform an excellent estimation......Grid synchronization algorithms are of great importance in the control of grid-connected power converters, as fast and accurate detection of the grid voltage parameters is crucial in order to implement stable control strategies under generic grid conditions. This paper presents a new grid...

  4. Optimal Control of Micro Grid Operation Mode Seamless Switching Based on Radau Allocation Method

    Science.gov (United States)

    Chen, Xiaomin; Wang, Gang

    2017-05-01

    The seamless switching process of micro grid operation mode directly affects the safety and stability of its operation. According to the switching process from island mode to grid-connected mode of micro grid, we establish a dynamic optimization model based on two grid-connected inverters. We use Radau allocation method to discretize the model, and use Newton iteration method to obtain the optimal solution. Finally, we implement the optimization mode in MATLAB and get the optimal control trajectory of the inverters.

  5. Interior point algorithm-based power flow optimisation of a combined AC and DC multi-terminal grid

    Directory of Open Access Journals (Sweden)

    Farhan Beg

    2015-01-01

    Full Text Available The high cost of power electronic equipment, lower reliability and poor power handling capacity of the semiconductor devices had stalled the deployment of systems based on DC (multi-terminal direct current system (MTDC networks. The introduction of voltage source converters (VSCs for transmission has renewed the interest in the development of large interconnected grids based on both alternate current (AC and DC transmission networks. Such a grid platform also realises the added advantage of integrating the renewable energy sources into the grid. Thus a grid based on DC MTDC network is a possible solution to improve energy security and check the increasing supply demand gap. An optimal power solution for combined AC and DC grids obtained by the solution of the interior point algorithm is proposed in this study. Multi-terminal HVDC grids lie at the heart of various suggested transmission capacity increases. A significant difference is observed when MTDC grids are solved for power flows in place of conventional AC grids. This study deals with the power flow problem of a combined MTDC and an AC grid. The AC side is modelled with the full power flow equations and the VSCs are modelled using a connecting line, two generators and an AC node. The VSC and the DC losses are also considered. The optimisation focuses on several different goals. Three different scenarios are presented in an arbitrary grid network with ten AC nodes and five converter stations.

  6. BESIII and SuperB: distributed job management with Ganga

    International Nuclear Information System (INIS)

    Antoniev, I; Kenyon, M; Moscicki, J; Deng, Z; Han, Y; Zhang, X; Ebke, J; Egede, U; Richards, A; Fella, A; Galvani, A; Lin, L; Luppi, E; Manzali, M; Tomassetti, L; Nicholson, C; Slater, M; Spinoso, V

    2012-01-01

    A job submission and management tool is one of the necessary components in any distributed computing system. Such a tool should provide a user-friendly interface for physics production groups and ordinary analysis users to access heterogeneous computing resources, without requiring knowledge of the underlying grid middleware. Ganga, with its common framework and customizable plug-in structure is such a tool. This paper will describe how experiment-specific job management tools for BESIII and SuperB were developed as Ganga plug-ins to meet their own unique requirements, discuss and contrast their challenges met and lessons learned.

  7. Modular Multilevel Converters Based Variable Speed Wind Turbines for Grid Faults

    DEFF Research Database (Denmark)

    Deng, Fujin; Liu, Dong; Wang, Yanbo

    2016-01-01

    in the dc-link of the power converter to improve system performance, but also ensure the grid-side current balancing to increase the generated power of the wind turbine under the unbalanced grid fault, in comparison with the conventional VSWT based on BTB three-level NPC converters. The simulation studies......The modular multilevel converter (MMC) becomes attractive in the medium- and high-power application with high modularity. In this paper, the MMC is proposed to be applied in the variable speed wind turbine (VSWT) based on the full-scale back-to-back (BTB) power converter, where the generator...

  8. Grid regulation services for energy storage devices based on grid frequency

    Energy Technology Data Exchange (ETDEWEB)

    Pratt, Richard M.; Hammerstrom, Donald J.; Kintner-Meyer, Michael C. W.; Tuffner, Francis K.

    2017-09-05

    Disclosed herein are representative embodiments of methods, apparatus, and systems for charging and discharging an energy storage device connected to an electrical power distribution system. In one exemplary embodiment, a controller monitors electrical characteristics of an electrical power distribution system and provides an output to a bi-directional charger causing the charger to charge or discharge an energy storage device (e.g., a battery in a plug-in hybrid electric vehicle (PHEV)). The controller can help stabilize the electrical power distribution system by increasing the charging rate when there is excess power in the electrical power distribution system (e.g., when the frequency of an AC power grid exceeds an average value), or by discharging power from the energy storage device to stabilize the grid when there is a shortage of power in the electrical power distribution system (e.g., when the frequency of an AC power grid is below an average value).

  9. Grid regulation services for energy storage devices based on grid frequency

    Science.gov (United States)

    Pratt, Richard M; Hammerstrom, Donald J; Kintner-Meyer, Michael C.W.; Tuffner, Francis K

    2013-07-02

    Disclosed herein are representative embodiments of methods, apparatus, and systems for charging and discharging an energy storage device connected to an electrical power distribution system. In one exemplary embodiment, a controller monitors electrical characteristics of an electrical power distribution system and provides an output to a bi-directional charger causing the charger to charge or discharge an energy storage device (e.g., a battery in a plug-in hybrid electric vehicle (PHEV)). The controller can help stabilize the electrical power distribution system by increasing the charging rate when there is excess power in the electrical power distribution system (e.g., when the frequency of an AC power grid exceeds an average value), or by discharging power from the energy storage device to stabilize the grid when there is a shortage of power in the electrical power distribution system (e.g., when the frequency of an AC power grid is below an average value).

  10. Bayesian grid matching

    DEFF Research Database (Denmark)

    Hartelius, Karsten; Carstensen, Jens Michael

    2003-01-01

    A method for locating distorted grid structures in images is presented. The method is based on the theories of template matching and Bayesian image restoration. The grid is modeled as a deformable template. Prior knowledge of the grid is described through a Markov random field (MRF) model which r...

  11. The biometric-based module of smart grid system

    Science.gov (United States)

    Engel, E.; Kovalev, I. V.; Ermoshkina, A.

    2015-10-01

    Within Smart Grid concept the flexible biometric-based module base on Principal Component Analysis (PCA) and selective Neural Network is developed. The formation of the selective Neural Network the biometric-based module uses the method which includes three main stages: preliminary processing of the image, face localization and face recognition. Experiments on the Yale face database show that (i) selective Neural Network exhibits promising classification capability for face detection, recognition problems; and (ii) the proposed biometric-based module achieves near real-time face detection, recognition speed and the competitive performance, as compared to some existing subspaces-based methods.

  12. Degenerate Quadtree Latitude/Longitude Grid Based on WGS-84 Ellipsoidal Facet

    Directory of Open Access Journals (Sweden)

    HU Bailin

    2016-12-01

    Full Text Available For the needs of digital earth development and solving many global problems, a new discrete global grid system-DQLLG (degenerate quadtree latitude/longitude grid was put forward, which was based on WGS-84 ellipsoidal facet. The hierarchical subdivision method, characteristics and grid column/row coordinate system were detailed. The Latitude/Longitude coordinate, area and side length of multi-resolution meshes on different subdivision levels were calculated. Then the changes of mesh areas and side lengths were analyzed and compared that with spherical DQLLG. The research indicates that the DQLLG had many excellent features:uniformity, hierarchy, consistency of direction, extensive data compatibility and so on. It has certain practicality for Global GIS in the future.

  13. Towards Agent-Based Model Specification in Smart Grid: A Cognitive Agent-based Computing Approach

    OpenAIRE

    Akram, Waseem; Niazi, Muaz A.; Iantovics, Laszlo Barna

    2017-01-01

    A smart grid can be considered as a complex network where each node represents a generation unit or a consumer. Whereas links can be used to represent transmission lines. One way to study complex systems is by using the agent-based modeling (ABM) paradigm. An ABM is a way of representing a complex system of autonomous agents interacting with each other. Previously, a number of studies have been presented in the smart grid domain making use of the ABM paradigm. However, to the best of our know...

  14. Photovoltaic Grid-Connected Modeling and Characterization Based on Experimental Results

    Science.gov (United States)

    Humada, Ali M.; Hojabri, Mojgan; Sulaiman, Mohd Herwan Bin; Hamada, Hussein M.; Ahmed, Mushtaq N.

    2016-01-01

    A grid-connected photovoltaic (PV) system operates under fluctuated weather condition has been modeled and characterized based on specific test bed. A mathematical model of a small-scale PV system has been developed mainly for residential usage, and the potential results have been simulated. The proposed PV model based on three PV parameters, which are the photocurrent, IL, the reverse diode saturation current, Io, the ideality factor of diode, n. Accuracy of the proposed model and its parameters evaluated based on different benchmarks. The results showed that the proposed model fitting the experimental results with high accuracy compare to the other models, as well as the I-V characteristic curve. The results of this study can be considered valuable in terms of the installation of a grid-connected PV system in fluctuated climatic conditions. PMID:27035575

  15. A multi-agent based distributed energy management scheme for smart grid applications

    International Nuclear Information System (INIS)

    Radhakrishnan, Bharat Menon; Srinivasan, Dipti

    2016-01-01

    A multi-agent system based distributed EMS (energy management system) is proposed in this paper to perform optimal energy allocation and management for grids comprising of renewables, storage and distributed generation. The reliable and efficient operation of smart grids is slackened due to the presence of intermittent renewables. As the load demand and renewables are uncertain throughout the day, an energy management system is essential to ensure grid stability and achieve reductions in operation costs and CO_2 emissions. The main objectives of the proposed algorithm is to maintain power balance in the system and to ensure long cycle life for storage units by controlling their SOC (state of charge). The proposed EMS scheme is tested and validated on a practical test system, which replicates a small-scale smart grid with a variety of distributed sources, storage devices, loads, power electronic converters, and SCADA (supervisory control and data acquisition) system. This system is also connected to the utility grid and the power exchange is controlled with the help of a battery system through a fuzzy based decision-making framework. The proposed algorithm is also extensively verified and tested using a series of sensitivity analyses and benchmarking with existing algorithms. - Highlights: • An agent-based decentralized algorithm is proposed to perform energy management. • The multi-agent system approach eliminates the possibility of single point failures. • Adaptive fuzzy systems make the decision making more reliable, flexible and robust. • The algorithm is extensively tested and validated using sensitivity and verification analyses.

  16. Elliptic Curve Cryptography-Based Authentication with Identity Protection for Smart Grids.

    Directory of Open Access Journals (Sweden)

    Liping Zhang

    Full Text Available In a smart grid, the power service provider enables the expected power generation amount to be measured according to current power consumption, thus stabilizing the power system. However, the data transmitted over smart grids are not protected, and then suffer from several types of security threats and attacks. Thus, a robust and efficient authentication protocol should be provided to strength the security of smart grid networks. As the Supervisory Control and Data Acquisition system provides the security protection between the control center and substations in most smart grid environments, we focus on how to secure the communications between the substations and smart appliances. Existing security approaches fail to address the performance-security balance. In this study, we suggest a mitigation authentication protocol based on Elliptic Curve Cryptography with privacy protection by using a tamper-resistant device at the smart appliance side to achieve a delicate balance between performance and security of smart grids. The proposed protocol provides some attractive features such as identity protection, mutual authentication and key agreement. Finally, we demonstrate the completeness of the proposed protocol using the Gong-Needham-Yahalom logic.

  17. Elliptic Curve Cryptography-Based Authentication with Identity Protection for Smart Grids.

    Science.gov (United States)

    Zhang, Liping; Tang, Shanyu; Luo, He

    2016-01-01

    In a smart grid, the power service provider enables the expected power generation amount to be measured according to current power consumption, thus stabilizing the power system. However, the data transmitted over smart grids are not protected, and then suffer from several types of security threats and attacks. Thus, a robust and efficient authentication protocol should be provided to strength the security of smart grid networks. As the Supervisory Control and Data Acquisition system provides the security protection between the control center and substations in most smart grid environments, we focus on how to secure the communications between the substations and smart appliances. Existing security approaches fail to address the performance-security balance. In this study, we suggest a mitigation authentication protocol based on Elliptic Curve Cryptography with privacy protection by using a tamper-resistant device at the smart appliance side to achieve a delicate balance between performance and security of smart grids. The proposed protocol provides some attractive features such as identity protection, mutual authentication and key agreement. Finally, we demonstrate the completeness of the proposed protocol using the Gong-Needham-Yahalom logic.

  18. Air Pollution Monitoring and Mining Based on Sensor Grid in London

    OpenAIRE

    Ma, Yajie; Richards, Mark; Ghanem, Moustafa; Guo, Yike; Hassard, John

    2008-01-01

    In this paper, we present a distributed infrastructure based on wireless sensors network and Grid computing technology for air pollution monitoring and mining, which aims to develop low-cost and ubiquitous sensor networks to collect real-time, large scale and comprehensive environmental data from road traffic emissions for air pollution monitoring in urban environment. The main informatics challenges in respect to constructing the high-throughput sensor Grid are discussed in this paper. We pr...

  19. AliEn - ALICE environment on the GRID

    International Nuclear Information System (INIS)

    Saiz, P.; Aphecetche, L.; Buncic, P.; Piskac, R.; Revsbech, J.-E.; Sego, V.

    2003-01-01

    AliEn (http://alien.cern.ch) (ALICE Environment) is a Grid framework built on top of the latest Internet standards for information exchange and authentication (SOAP, PKI) and common Open Source components. AliEn provides a virtual file catalogue that allows transparent access to distributed datasets and a number of collaborating Web services which implement the authentication, job execution, file transport, performance monitor and event logging. In the paper we will present the architecture and components of the system

  20. AliEn - ALICE environment on the GRID

    CERN Document Server

    Saiz, P; Buncic, P; Piskac, R; Revsbech, J E; Sego, V

    2003-01-01

    AliEn (http://alien.cern.ch) (ALICE Environment) is a Grid framework built on top of the latest Internet standards for information exchange and authentication (SOAP, PKI) and common Open Source components. AliEn provides a virtual file catalogue that allows transparent access to distributed datasets and a number of collaborating Web services which implement the authentication, job execution, file transport, performance monitor and event logging. In the paper we will present the architecture and components of the system.

  1. Disaster Monitoring using Grid Based Data Fusion Algorithms

    Directory of Open Access Journals (Sweden)

    Cătălin NAE

    2010-12-01

    Full Text Available This is a study of the application of Grid technology and high performance parallelcomputing to a candidate algorithm for jointly accomplishing data fusion from different sensors. Thisincludes applications for both image analysis and/or data processing for simultaneously trackingmultiple targets in real-time. The emphasis is on comparing the architectures of the serial andparallel algorithms, and characterizing the performance benefits achieved by the parallel algorithmwith both on-ground and in-space hardware implementations. The improved performance levelsachieved by the use of Grid technology (middleware for Parallel Data Fusion are presented for themain metrics of interest in near real-time applications, namely latency, total computation load, andtotal sustainable throughput. The objective of this analysis is, therefore, to demonstrate animplementation of multi-sensor data fusion and/or multi-target tracking functions within an integratedmulti-node portable HPC architecture based on emerging Grid technology. The key metrics to bedetermined in support of ongoing system analyses includes: required computational throughput inMFLOPS; latency between receipt of input data and resulting outputs; and scalability, processorutilization and memory requirements. Furthermore, the standard MPI functions are considered to beused for inter-node communications in order to promote code portability across multiple HPCcomputer platforms, both in space and on-ground.

  2. Grid occupancy estimation for environment perception based on belief functions and PCR6

    Science.gov (United States)

    Moras, Julien; Dezert, Jean; Pannetier, Benjamin

    2015-05-01

    In this contribution, we propose to improve the grid map occupancy estimation method developed so far based on belief function modeling and the classical Dempster's rule of combination. Grid map offers a useful representation of the perceived world for mobile robotics navigation. It will play a major role for the security (obstacle avoidance) of next generations of terrestrial vehicles, as well as for future autonomous navigation systems. In a grid map, the occupancy of each cell representing a small piece of the surrounding area of the robot must be estimated at first from sensors measurements (typically LIDAR, or camera), and then it must also be classified into different classes in order to get a complete and precise perception of the dynamic environment where the robot moves. So far, the estimation and the grid map updating have been done using fusion techniques based on the probabilistic framework, or on the classical belief function framework thanks to an inverse model of the sensors. Mainly because the latter offers an interesting management of uncertainties when the quality of available information is low, and when the sources of information appear as conflicting. To improve the performances of the grid map estimation, we propose in this paper to replace Dempster's rule of combination by the PCR6 rule (Proportional Conflict Redistribution rule #6) proposed in DSmT (Dezert-Smarandache) Theory. As an illustrating scenario, we consider a platform moving in dynamic area and we compare our new realistic simulation results (based on a LIDAR sensor) with those obtained by the probabilistic and the classical belief-based approaches.

  3. Grid generation methods

    CERN Document Server

    Liseikin, Vladimir D

    2010-01-01

    This book is an introduction to structured and unstructured grid methods in scientific computing, addressing graduate students, scientists as well as practitioners. Basic local and integral grid quality measures are formulated and new approaches to mesh generation are reviewed. In addition to the content of the successful first edition, a more detailed and practice oriented description of monitor metrics in Beltrami and diffusion equations is given for generating adaptive numerical grids. Also, new techniques developed by the author are presented, in particular a technique based on the inverted form of Beltrami’s partial differential equations with respect to control metrics. This technique allows the generation of adaptive grids for a wide variety of computational physics problems, including grid clustering to given function values and gradients, grid alignment with given vector fields, and combinations thereof. Applications of geometric methods to the analysis of numerical grid behavior as well as grid ge...

  4. Scientific component framework for W7-X using service oriented GRID middleware

    International Nuclear Information System (INIS)

    Werner, A.; Svensson, J.; Kuehner, G.; Bluhm, T.; Heimann, P.; Jakob, S.; Hennig, C.; Kroiss, H.; Laqua, H.; Lewerentz, M.; Riemann, H.; Schacht, J.; Spring, A.; Zilker, M.; Maier, J.

    2010-01-01

    Future fusion experiments, aiming to demonstrate steady state reactor operation, require physics driven plasma control based on increasingly complex plasma models. A precondition for establishing such control systems is widely automated data analysis, which can provide integration of multiple diagnostic on a large scale. Even high quality online data evaluation, which is essential for the scientific documentation of the experiment, has to be performed automatically due to the huge data sets being recorded in long discharge runs. An automated system that can handle these requirements will have to be built on reusable software components that can be maintained by the domain experts: diagnosticians, theorists, engineers and others. For Wendelstein 7-X a service oriented architecture seems to be appropriate, in which software components can be exposed as services with well defined interface contracts. Although grid computing has up to now been mainly used for remote job execution, a more promising service oriented middleware has emerged from the recent grid specification, the open grid service architecture (OGSA). It is based on stateful web services defined by the web service resource framework (WSRF) standard. In particular, the statefulness of services allows to setup complex models without unnecessary performance losses by frequent transmission of large and complex data sets. At present, the usability of this technology in the W7-X CoDaC context is under evaluation by first service implementations.

  5. The study and prioritization of job satisfaction dimensions in zanjan-based Refah Bank employees

    Directory of Open Access Journals (Sweden)

    Siavash Rashidi

    2013-01-01

    Full Text Available Theories associated with job satisfaction are based on the principle that all environmental elements could shape entire career satisfaction.  In literature, major studies illustrate that positive and negative emotions are largely associated with job satisfaction. Job satisfaction source is not only job position but also other factors such as the physical and social work environment, relationships with supervisors and colleagues, group culture and management style of the managers.  In this study, it is aimed to evaluate and prioritize the five dimensions of job satisfaction in Zanjan Refah Bank employees: (i the nature of the job (ii supervisor, (iii peer, (iv promotion and (v payment. In this study a field research was applied with a survey study. To testify the hypothesis, the Pearson parametric and Friedman test was conducted. The major findings of this study are (i there is a negative correlation exists between level of education and nature of the job  (ii job promotion and payment, (iii there is not any significant differences in job satisfaction between men and women.Keywords: Refah Bank, Zanjan, job satisfaction

  6. Experiences with the GLUE information schema in the LCG/EGEE production grid

    International Nuclear Information System (INIS)

    Burke, S; Andreozzi, S; Field, L

    2008-01-01

    A common information schema for the description of Grid resources and services is an essential requirement for interoperating Grid infrastructures, and its implementation interacts with every Grid component. In this context, the GLUE information schema was originally defined in 2002 as a joint project between the European DataGrid and DataTAG projects and the US iVDGL. The schema has major components to describe Computing and Storage Elements, and also generic Service and Site information. It has been used extensively in the LCG/EGEE Grid, for job submission, data management, service discovery and monitoring. In this paper we present the experience gained over the last five years, highlighting both successes and problems. In particular, we consider the importance of having a clear definition of schema attributes; the construction of standard information providers and difficulties encountered in mapping an abstract schema to diverse real systems; the configuration of publication in a way which suits system managers and the varying characteristics of Grid sites; the validation of published information; the ways in which information can be used (and misused) by Grid services and users; and issues related to managing schema upgrades in a large distributed system

  7. Integration of the Chinese HPC Grid in ATLAS Distributed Computing

    Science.gov (United States)

    Filipčič, A.; ATLAS Collaboration

    2017-10-01

    Fifteen Chinese High-Performance Computing sites, many of them on the TOP500 list of most powerful supercomputers, are integrated into a common infrastructure providing coherent access to a user through an interface based on a RESTful interface called SCEAPI. These resources have been integrated into the ATLAS Grid production system using a bridge between ATLAS and SCEAPI which translates the authorization and job submission protocols between the two environments. The ARC Computing Element (ARC-CE) forms the bridge using an extended batch system interface to allow job submission to SCEAPI. The ARC-CE was setup at the Institute for High Energy Physics, Beijing, in order to be as close as possible to the SCEAPI front-end interface at the Computing Network Information Center, also in Beijing. This paper describes the technical details of the integration between ARC-CE and SCEAPI and presents results so far with two supercomputer centers, Tianhe-IA and ERA. These two centers have been the pilots for ATLAS Monte Carlo Simulation in SCEAPI and have been providing CPU power since fall 2015.

  8. Dashboard Task Monitor for Managing ATLAS User Analysis on the Grid

    Science.gov (United States)

    Sargsyan, L.; Andreeva, J.; Jha, M.; Karavakis, E.; Kokoszkiewicz, L.; Saiz, P.; Schovancova, J.; Tuckett, D.; Atlas Collaboration

    2014-06-01

    The organization of the distributed user analysis on the Worldwide LHC Computing Grid (WLCG) infrastructure is one of the most challenging tasks among the computing activities at the Large Hadron Collider. The Experiment Dashboard offers a solution that not only monitors but also manages (kill, resubmit) user tasks and jobs via a web interface. The ATLAS Dashboard Task Monitor provides analysis users with a tool that is independent of the operating system and Grid environment. This contribution describes the functionality of the application and its implementation details, in particular authentication, authorization and audit of the management operations.

  9. Dashboard task monitor for managing ATLAS user analysis on the grid

    International Nuclear Information System (INIS)

    Sargsyan, L; Andreeva, J; Karavakis, E; Saiz, P; Tuckett, D; Jha, M; Kokoszkiewicz, L; Schovancova, J

    2014-01-01

    The organization of the distributed user analysis on the Worldwide LHC Computing Grid (WLCG) infrastructure is one of the most challenging tasks among the computing activities at the Large Hadron Collider. The Experiment Dashboard offers a solution that not only monitors but also manages (kill, resubmit) user tasks and jobs via a web interface. The ATLAS Dashboard Task Monitor provides analysis users with a tool that is independent of the operating system and Grid environment. This contribution describes the functionality of the application and its implementation details, in particular authentication, authorization and audit of the management operations.

  10. Energy efficiency analysis for flexible-grid OFDM-based optical networks

    DEFF Research Database (Denmark)

    Vizcaíno, Jorge López; Ye, Yabin; Tafur Monroy, Idelfonso

    2012-01-01

    As the Internet traffic grows, the energy efficiency gains more attention as a design factor for the planning and operation of telecommunication networks. This paper is devoted to the study of energy efficiency in optical transport networks, comparing the performance of an innovative flexible......-grid network based on Orthogonal Frequency Division Multiplexing (OFDM) with that of conventional fixed-grid Wavelength Division Multiplexing (WDM) networks with a Single Line Rate (SLR) and with a Mixed Line Rate (MLR) operation. The power consumption values of the network elements are introduced. Energy......-aware heuristic algorithms are proposed for the resource allocation both in static (offline) and dynamic (online) scenarios with time-varying demands for the Elastic-bandwidth OFDM-based network and the WDM networks (with SLR and MLR). The energy efficiency performance of the two network technologies under...

  11. Job Design, Training Effect and Job Satisfaction: Evidence from Work Placement at Audit Firms

    OpenAIRE

    Liu Guangyou

    2017-01-01

    This paper aims to test the relationships between job design aspects and job satisfaction, as well as the mediating role of training effect in these relationships. Regression and mediation analyses were performed based on the data collected from a questionnaire-based survey on the senior accounting students’ audit work placement at audit firms. I conclude that repeated tedious non-professional job aspect is negatively related to job satisfaction, whereas judgmental professional job aspect is ...

  12. A Data Transmission Algorithm Based on Dynamic Grid Division for Coal Goaf Temperature Monitoring

    Directory of Open Access Journals (Sweden)

    Qingsong Hu

    2014-01-01

    Full Text Available WSN (wireless sensor network is a perfect tool of temperature monitoring in coal goaf. Based on the three-zone theory of goaf, the GtmWSN model is proposed, and its dynamic features are analyzed. Accordingly, a data transmission scheme, named DTDGD, is worked out. Firstly, sink nodes conduct dynamic grid division on the GtmWSN according to virtual semicircle. Secondly, each node will confirm to which grid it belongs based on grid number. Finally, data will be delivered to sink nodes with greedy forward and hole avoidance. Simulation results and field data showed that the GtmWSN and DTDGD satisfied the lifetime need of goaf temperature monitoring.

  13. SoilGrids1km--global soil information based on automated mapping.

    Directory of Open Access Journals (Sweden)

    Tomislav Hengl

    Full Text Available BACKGROUND: Soils are widely recognized as a non-renewable natural resource and as biophysical carbon sinks. As such, there is a growing requirement for global soil information. Although several global soil information systems already exist, these tend to suffer from inconsistencies and limited spatial detail. METHODOLOGY/PRINCIPAL FINDINGS: We present SoilGrids1km--a global 3D soil information system at 1 km resolution--containing spatial predictions for a selection of soil properties (at six standard depths: soil organic carbon (g kg-1, soil pH, sand, silt and clay fractions (%, bulk density (kg m-3, cation-exchange capacity (cmol+/kg, coarse fragments (%, soil organic carbon stock (t ha-1, depth to bedrock (cm, World Reference Base soil groups, and USDA Soil Taxonomy suborders. Our predictions are based on global spatial prediction models which we fitted, per soil variable, using a compilation of major international soil profile databases (ca. 110,000 soil profiles, and a selection of ca. 75 global environmental covariates representing soil forming factors. Results of regression modeling indicate that the most useful covariates for modeling soils at the global scale are climatic and biomass indices (based on MODIS images, lithology, and taxonomic mapping units derived from conventional soil survey (Harmonized World Soil Database. Prediction accuracies assessed using 5-fold cross-validation were between 23-51%. CONCLUSIONS/SIGNIFICANCE: SoilGrids1km provide an initial set of examples of soil spatial data for input into global models at a resolution and consistency not previously available. Some of the main limitations of the current version of SoilGrids1km are: (1 weak relationships between soil properties/classes and explanatory variables due to scale mismatches, (2 difficulty to obtain covariates that capture soil forming factors, (3 low sampling density and spatial clustering of soil profile locations. However, as the SoilGrids system is

  14. Computer-Based Job and Occupational Data Collection Methods: Feasibility Study

    National Research Council Canada - National Science Library

    Mitchell, Judith I

    1998-01-01

    .... The feasibility study was conducted to assess the operational and logistical problems involved with the development, implementation, and evaluation of computer-based job and occupational data collection methods...

  15. Job insecurity , work-based support, job satisfaction, organisational commitment and general health of human resources professionals in a chemical industry / by Florence Nomhlangano Rani

    OpenAIRE

    Rani, Nomhlangano Florence

    2005-01-01

    The work environment in which South African employees have to function is highly demanding, offering them little in terms of job security, but simultaneously expecting them to give more in terms of inter alia flexibility, competency, and effort. Tracking and addressing chemical industry employees' functioning in areas that could affect their general health and consequent standard of service is essential. Job insecurity, work-based support, job satisfaction, organisational commitment and gener...

  16. Time-domain analysis of planar microstrip devices using a generalized Yee-algorithm based on unstructured grids

    Science.gov (United States)

    Gedney, Stephen D.; Lansing, Faiza

    1993-01-01

    The generalized Yee-algorithm is presented for the temporal full-wave analysis of planar microstrip devices. This algorithm has the significant advantage over the traditional Yee-algorithm in that it is based on unstructured and irregular grids. The robustness of the generalized Yee-algorithm is that structures that contain curved conductors or complex three-dimensional geometries can be more accurately, and much more conveniently modeled using standard automatic grid generation techniques. This generalized Yee-algorithm is based on the the time-marching solution of the discrete form of Maxwell's equations in their integral form. To this end, the electric and magnetic fields are discretized over a dual, irregular, and unstructured grid. The primary grid is assumed to be composed of general fitted polyhedra distributed throughout the volume. The secondary grid (or dual grid) is built up of the closed polyhedra whose edges connect the centroid's of adjacent primary cells, penetrating shared faces. Faraday's law and Ampere's law are used to update the fields normal to the primary and secondary grid faces, respectively. Subsequently, a correction scheme is introduced to project the normal fields onto the grid edges. It is shown that this scheme is stable, maintains second-order accuracy, and preserves the divergenceless nature of the flux densities. Finally, for computational efficiency the algorithm is structured as a series of sparse matrix-vector multiplications. Based on this scheme, the generalized Yee-algorithm has been implemented on vector and parallel high performance computers in a highly efficient manner.

  17. Relationship between job stress, occupational position and job satisfaction using a brief job stress questionnaire (BJSQ).

    Science.gov (United States)

    Kawada, Tomoyuki; Otsuka, Toshiaki

    2011-01-01

    Subjects with higher occupational position are speculated to have higher ability to handle with stress, and they were less affected by job stress. This study focused on the relationship between job satisfaction and three sub-scales of a brief job stress questionnaire (BJSQ) related to workload. This self-administered questionnaire was distributed to 371 employees of a company, and all the workers sent back their responses. Among the 57 items graded on a 4-point Likert-type scale to measure job stressors, psycho-physical complaints, and support for workers, the authors studied the influence of quantitative and qualitative job overload (six items), job control (three items), and support port (six items). The job satisfaction score estimated on a 4-point Likert-type scale was also used in relation to job stress determined using a 15-item scale from the BJSQ based on demand-control-support model. Occupational positions were classified into directors, managers, and general workers, and the content of job was classified into clerical workers, skilled technicians, and unskilled manual workers. All the scales on job stress presented acceptable alpha coefficients reflecting high internal consistency (job demand: 0.855, job control: 0.644, and support: 0.878, respectively). Principal axis factor analysis was conducted, and three factors were extracted; support, job demand and job control. There was a significant difference in the mean score among four groups divided by the job satisfaction level as evaluated by Dunnett's multiple comparison, and members who were dissatisfied with their job showed a high job demand, limited job control, and poor support. The mean score of support for managers were significantly higher (lower support) than that for general workers. The logistic regression analysis revealed that job control and support contributed significantly to job satisfaction. In addition, unskilled manual workers showed significantly higher job dissatisfaction compared

  18. Air Pollution Monitoring and Mining Based on Sensor Grid in London.

    Science.gov (United States)

    Ma, Yajie; Richards, Mark; Ghanem, Moustafa; Guo, Yike; Hassard, John

    2008-06-01

    In this paper, we present a distributed infrastructure based on wireless sensors network and Grid computing technology for air pollution monitoring and mining, which aims to develop low-cost and ubiquitous sensor networks to collect real-time, large scale and comprehensive environmental data from road traffic emissions for air pollution monitoring in urban environment. The main informatics challenges in respect to constructing the high-throughput sensor Grid are discussed in this paper. We present a twolayer network framework, a P2P e-Science Grid architecture, and the distributed data mining algorithm as the solutions to address the challenges. We simulated the system in TinyOS to examine the operation of each sensor as well as the networking performance. We also present the distributed data mining result to examine the effectiveness of the algorithm.

  19. Grid-based modeling for land use planning and environmental resource mapping.

    Energy Technology Data Exchange (ETDEWEB)

    Kuiper, J. A.

    1999-08-04

    Geographic Information System (GIS) technology is used by land managers and natural resource planners for examining resource distribution and conducting project planning, often by visually interpreting spatial data representing environmental or regulatory variables. Frequently, many variables influence the decision-making process, and modeling can improve results with even a small investment of time and effort. Presented are several grid-based GIS modeling projects, including: (1) land use optimization under environmental and regulatory constraints; (2) identification of suitable wetland mitigation sites; and (3) predictive mapping of prehistoric cultural resource sites. As different as the applications are, each follows a similar process of problem conceptualization, implementation of a practical grid-based GIS model, and evaluation of results.

  20. Improving ATLAS grid site reliability with functional tests using HammerCloud

    Science.gov (United States)

    Elmsheuser, Johannes; Legger, Federica; Medrano Llamas, Ramon; Sciacca, Gianfranco; van der Ster, Dan

    2012-12-01

    With the exponential growth of LHC (Large Hadron Collider) data in 2011, and more coming in 2012, distributed computing has become the established way to analyse collider data. The ATLAS grid infrastructure includes almost 100 sites worldwide, ranging from large national computing centers to smaller university clusters. These facilities are used for data reconstruction and simulation, which are centrally managed by the ATLAS production system, and for distributed user analysis. To ensure the smooth operation of such a complex system, regular tests of all sites are necessary to validate the site capability of successfully executing user and production jobs. We report on the development, optimization and results of an automated functional testing suite using the HammerCloud framework. Functional tests are short lightweight applications covering typical user analysis and production schemes, which are periodically submitted to all ATLAS grid sites. Results from those tests are collected and used to evaluate site performances. Sites that fail or are unable to run the tests are automatically excluded from the PanDA brokerage system, therefore avoiding user or production jobs to be sent to problematic sites.

  1. Proceedings of the second workshop of LHC Computing Grid, LCG-France

    International Nuclear Information System (INIS)

    Chollet, Frederique; Hernandez, Fabio; Malek, Fairouz; Gaelle, Shifrin

    2007-03-01

    The second LCG-France Workshop was held in Clermont-Ferrand on 14-15 March 2007. These sessions organized by IN2P3 and DAPNIA were attended by around 70 participants working with the Computing Grid of LHC in France. The workshop was a opportunity of exchanges of information between the French and foreign site representatives on one side and delegates of experiments on the other side. The event allowed enlightening the place of LHC Computing Task within the frame of W-LCG world project, the undergoing actions and the prospects in 2007 and beyond. The following communications were presented: 1. The current status of the LHC computation in France; 2.The LHC Grid infrastructure in France and associated resources; 3.Commissioning of Tier 1; 4.The sites of Tier-2s and Tier-3s; 5.Computing in ALICE experiment; 6.Computing in ATLAS experiment; 7.Computing in the CMS experiments; 8.Computing in the LHCb experiments; 9.Management and operation of computing grids; 10.'The VOs talk to sites'; 11.Peculiarities of ATLAS; 12.Peculiarities of CMS and ALICE; 13.Peculiarities of LHCb; 14.'The sites talk to VOs'; 15. Worldwide operation of Grid; 16.Following-up the Grid jobs; 17.Surveillance and managing the failures; 18. Job scheduling and tuning; 19.Managing the site infrastructure; 20.LCG-France communications; 21.Managing the Grid data; 22.Pointing the net infrastructure and site storage. 23.ALICE bulk transfers; 24.ATLAS bulk transfers; 25.CMS bulk transfers; 26. LHCb bulk transfers; 27.Access to LHCb data; 28.Access to CMS data; 29.Access to ATLAS data; 30.Access to ALICE data; 31.Data analysis centers; 32.D0 Analysis Farm; 33.Some CMS grid analyses; 34.PROOF; 35.Distributed analysis using GANGA; 36.T2 set-up for end-users. In their concluding remarks Fairouz Malek and Dominique Pallin stressed that the current workshop was more close to users while the tasks for tightening the links between the sites and the experiments were definitely achieved. The IN2P3 leadership expressed

  2. Beyond grid security

    International Nuclear Information System (INIS)

    Hoeft, B; Epting, U; Koenig, T

    2008-01-01

    While many fields relevant to Grid security are already covered by existing working groups, their remit rarely goes beyond the scope of the Grid infrastructure itself. However, security issues pertaining to the internal set-up of compute centres have at least as much impact on Grid security. Thus, this talk will present briefly the EU ISSeG project (Integrated Site Security for Grids). In contrast to groups such as OSCT (Operational Security Coordination Team) and JSPG (Joint Security Policy Group), the purpose of ISSeG is to provide a holistic approach to security for Grid computer centres, from strategic considerations to an implementation plan and its deployment. The generalised methodology of Integrated Site Security (ISS) is based on the knowledge gained during its implementation at several sites as well as through security audits, and this will be briefly discussed. Several examples of ISS implementation tasks at the Forschungszentrum Karlsruhe will be presented, including segregation of the network for administration and maintenance and the implementation of Application Gateways. Furthermore, the web-based ISSeG training material will be introduced. This aims to offer ISS implementation guidance to other Grid installations in order to help avoid common pitfalls

  3. Job Demand and Job Resources related to the turnover intention of public health nurses: An analysis using a Job Demands-Resources model.

    Science.gov (United States)

    Iguchi, Aya

    2016-01-01

    Objectives The purpose of this study was to investigate the job demands and job resources of public health nurses based on the Job Demands-Resources (JD-R) model, and to build a model that can estimate turnover intention based on job demands and job resources.Method By adding 12 items to the existing questionnaire, the author created a questionnaire consisting of 10 factors and 167 items, and used statistical analysis to examine job demands and job resources in relation to turnover intention.Results Out of 2,668 questionnaires sent, 1993 (72.5%) were returned. Considering sex-based differences in occupational stress, I analyzed women's answers in 1766 (66.2%) mails among the 1798 valid responses. The average age of respondents was 41.0±9.8 years, and the mean service duration was 17.0±10.0 years. For public health nurses, there was a turnover intention of 9.2%. The "job demands" section consisted of 29 items and 10 factors, while the "job resources" section consisted of 54 items and 22 factors. The result of examining the structure of job demands and job resources, leading to turnover intention was supported by the JD-R model. Turnover intention was strong and the Mental Component Summary (MCS) is low in those who had many job demands and few job resources (experiencing 'burn-out'). Enhancement of work engagement and turnover intention was weak in those who had many job resources. This explained approximately 60% of the dispersion to "burn-out", and approximately 40% to "work engagement", with four factors: work suitability, work significance, positive work self-balance, and growth opportunity of job resources.Conclusion This study revealed that turnover intention is strong in those who are burned out because of many job demands. Enhancement of work engagement and turnover intention is weak in those with many job resources. This suggests that suitable staffing and organized efforts to raise awareness of job significance are effective in reducing

  4. Flexible operation of parallel grid-connecting converters under unbalanced grid voltage

    DEFF Research Database (Denmark)

    Lu, Jinghang; Savaghebi, Mehdi; Guerrero, Josep M.

    2017-01-01

    -link voltage ripple, and overloading. Moreover, under grid voltage unbalance, the active power delivery ability is decreased due to the converter's current rating limitation. In this paper, a thorough study on the current limitation of the grid-connecting converter under grid voltage unbalance is conducted....... In addition, based on the principle that total output active power should be oscillation free, a coordinated control strategy is proposed for the parallel grid-connecting converters. The case study has been conducted to demonstrate the effectiveness of this proposed control strategy....

  5. Exact and heuristic solution approaches for the Integrated Job Scheduling and Constrained Network Routing Problem

    DEFF Research Database (Denmark)

    Gamst, M.

    2014-01-01

    problem. The methods are computationally evaluated on test instances arising from telecommunications with up to 500 jobs and 500 machines. Results show that solving the integrated job scheduling and constrained network routing problem to optimality is very difficult. The exact solution approach performs......This paper examines the problem of scheduling a number of jobs on a finite set of machines such that the overall profit of executed jobs is maximized. Each job has a certain demand, which must be sent to the executing machine via constrained paths. A job cannot start before all its demands have...... arrived at the machine. Furthermore, two resource demand transmissions cannot use the same edge in the same time period. The problem has application in grid computing, where a number of geographically distributed machines work together for solving large problems. The machines are connected through...

  6. Sustainable Power Supply Solutions for Off-Grid Base Stations

    Directory of Open Access Journals (Sweden)

    Asma Mohamad Aris

    2015-09-01

    Full Text Available The telecommunication sector plays a significant role in shaping the global economy and the way people share information and knowledge. At present, the telecommunication sector is liable for its energy consumption and the amount of emissions it emits in the environment. In the context of off-grid telecommunication applications, off-grid base stations (BSs are commonly used due to their ability to provide radio coverage over a wide geographic area. However, in the past, the off-grid BSs usually relied on emission-intensive power supply solutions such as diesel generators. In this review paper, various types of solutions (including, in particular, the sustainable solutions for powering BSs are discussed. The key aspects in designing an ideal power supply solution are reviewed, and these mainly include the pre-feasibility study and the thermal management of BSs, which comprise heating and cooling of the BS shelter/cabinets and BS electronic equipment and power supply components. The sizing and optimization approaches used to design the BSs’ power supply systems as well as the operational and control strategies adopted to manage the power supply systems are also reviewed in this paper.

  7. Bus.py: A GridLAB-D Communication Interface for Smart Distribution Grid Simulations

    Energy Technology Data Exchange (ETDEWEB)

    Hansen, Timothy M.; Palmintier, Bryan; Suryanarayanan, Siddharth; Maciejewski, Anthony A.; Siegel, Howard Jay

    2015-07-03

    As more Smart Grid technologies (e.g., distributed photovoltaic, spatially distributed electric vehicle charging) are integrated into distribution grids, static distribution simulations are no longer sufficient for performing modeling and analysis. GridLAB-D is an agent-based distribution system simulation environment that allows fine-grained end-user models, including geospatial and network topology detail. A problem exists in that, without outside intervention, once the GridLAB-D simulation begins execution, it will run to completion without allowing the real-time interaction of Smart Grid controls, such as home energy management systems and aggregator control. We address this lack of runtime interaction by designing a flexible communication interface, Bus.py (pronounced bus-dot-pie), that uses Python to pass messages between one or more GridLAB-D instances and a Smart Grid simulator. This work describes the design and implementation of Bus.py, discusses its usefulness in terms of some Smart Grid scenarios, and provides an example of an aggregator-based residential demand response system interacting with GridLAB-D through Bus.py. The small scale example demonstrates the validity of the interface and shows that an aggregator using said interface is able to control residential loads in GridLAB-D during runtime to cause a reduction in the peak load on the distribution system in (a) peak reduction and (b) time-of-use pricing cases.

  8. A composite passive damping method of the LLCL-filter based grid-tied inverter

    DEFF Research Database (Denmark)

    Wu, Weimin; Huang, Min; Sun, Yunjie

    2012-01-01

    This paper investigates the maximum and the minimum gain of the proportional resonant based grid current controller for a grid-tied inverter with a passive damped high-order power filter. It is found that the choice of the controller gain is limited to the local maximum amplitude determined by Q......-factor around the characteristic frequency of the filter and grid impedance. To obtain the Q-factor of a high-order system, an equivalent circuit analysis method is proposed and illustrated through several classical passive damped LCL- and LLCL-filters. It is shown that both the RC parallel damper...... that is in parallel with the capacitor of the LCL-filter or with the Lf-Cf resonant circuit of the LLCL-filter, and the RL series damper in series with the grid-side inductor have their own application limits. Thus, a composite passive damped LLCL-filter for the grid-tied inverter is proposed, which can effectively...

  9. A new service-oriented grid-based method for AIoT application and implementation

    Science.gov (United States)

    Zou, Yiqin; Quan, Li

    2017-07-01

    The traditional three-layer Internet of things (IoT) model, which includes physical perception layer, information transferring layer and service application layer, cannot express complexity and diversity in agricultural engineering area completely. It is hard to categorize, organize and manage the agricultural things with these three layers. Based on the above requirements, we propose a new service-oriented grid-based method to set up and build the agricultural IoT. Considering the heterogeneous, limitation, transparency and leveling attributes of agricultural things, we propose an abstract model for all agricultural resources. This model is service-oriented and expressed with Open Grid Services Architecture (OGSA). Information and data of agricultural things were described and encapsulated by using XML in this model. Every agricultural engineering application will provide service by enabling one application node in this service-oriented grid. Description of Web Service Resource Framework (WSRF)-based Agricultural Internet of Things (AIoT) and the encapsulation method were also discussed in this paper for resource management in this model.

  10. Improving energy efficiency and smart grid program analysis with agent-based end-use forecasting models

    International Nuclear Information System (INIS)

    Jackson, Jerry

    2010-01-01

    Electric utilities and regulators face difficult challenges evaluating new energy efficiency and smart grid programs prompted, in large part, by recent state and federal mandates and financial incentives. It is increasingly difficult to separate electricity use impacts of individual utility programs from the impacts of increasingly stringent appliance and building efficiency standards, increasing electricity prices, appliance manufacturer efficiency improvements, energy program interactions and other factors. This study reviews traditional approaches used to evaluate electric utility energy efficiency and smart-grid programs and presents an agent-based end-use modeling approach that resolves many of the shortcomings of traditional approaches. Data for a representative sample of utility customers in a Midwestern US utility are used to evaluate energy efficiency and smart grid program targets over a fifteen-year horizon. Model analysis indicates that a combination of the two least stringent efficiency and smart grid program scenarios provides peak hour reductions one-third greater than the most stringent smart grid program suggesting that reductions in peak demand requirements are more feasible when both efficiency and smart grid programs are considered together. Suggestions on transitioning from traditional end-use models to agent-based end-use models are provided.

  11. Power control based on particle swarm optimization of grid-connected inverter for hybrid renewable energy system

    International Nuclear Information System (INIS)

    García-Triviño, Pablo; Gil-Mena, Antonio José; Llorens-Iborra, Francisco; García-Vázquez, Carlos Andrés; Fernández-Ramírez, Luis M.; Jurado, Francisco

    2015-01-01

    Highlights: • Three PSO-based PI controllers for a grid-connected inverter were presented. • Two online PSO-based PI controllers were compared with an offline PSO-tuned PI. • The HRES and the inverter were evaluated under power changes and grid voltage sags. • Online ITAE-based PSO reduced ITAE (current THD) by 15.24% (5.32%) versus offline one. - Abstract: This paper is focused on the study of particle swarm optimization (PSO)-based PI controllers for the power control of a grid-connected inverter supplied from a hybrid renewable energy system. It is composed of two renewable energy sources (wind turbine and photovoltaic – PV – solar panels) and two energy storage systems (battery and hydrogen system, integrated by fuel cell and electrolyzer). Three PSO-based PI controllers are implemented: (1) conventional PI controller with offline tuning by PSO algorithm based on the integral time absolute error (ITAE) index; (2) PI controllers with online self-tuning by PSO algorithm based on the error; and (3) PI controllers with online self-tuning by PSO algorithm based on the ITAE index. To evaluate and compare the three controllers, the hybrid renewable energy system and the grid-connected inverter are simulated under changes in the active and reactive power values, as well as under a grid voltage sag. The results show that the online PSO-based PI controllers that optimize the ITAE index achieves the best response

  12. Application of synchronous grid-connected controller in the wind-solar-storage micro grid

    OpenAIRE

    Li, Hua; Ren, Yongfeng; Li, Le; Luo, Zhenpeng

    2016-01-01

    Recently, there has been an increasing interest in using distributed generators (DG) not only to inject power into the grid, but also to enhance the power quality. In this study, a space voltage pulse width modulation (SVPWM) control method is proposed for a synchronous grid-connected controller in a wind-solar-storage micro grid. This method is based on the appropriate topology of the synchronous controller. The wind-solar-storage micro grid is controlled to reconnect to the grid synchronous...

  13. Small-mammal density estimation: A field comparison of grid-based vs. web-based density estimators

    Science.gov (United States)

    Parmenter, R.R.; Yates, Terry L.; Anderson, D.R.; Burnham, K.P.; Dunnum, J.L.; Franklin, A.B.; Friggens, M.T.; Lubow, B.C.; Miller, M.; Olson, G.S.; Parmenter, Cheryl A.; Pollard, J.; Rexstad, E.; Shenk, T.M.; Stanley, T.R.; White, Gary C.

    2003-01-01

    blind” test allowed us to evaluate the influence of expertise and experience in calculating density estimates in comparison to simply using default values in programs CAPTURE and DISTANCE. While the rodent sample sizes were considerably smaller than the recommended minimum for good model results, we found that several models performed well empirically, including the web-based uniform and half-normal models in program DISTANCE, and the grid-based models Mb and Mbh in program CAPTURE (with AÌ‚ adjusted by species-specific full mean maximum distance moved (MMDM) values). These models produced accurate DÌ‚ values (with 95% confidence intervals that included the true D values) and exhibited acceptable bias but poor precision. However, in linear regression analyses comparing each model's DÌ‚ values to the true D values over the range of observed test densities, only the web-based uniform model exhibited a regression slope near 1.0; all other models showed substantial slope deviations, indicating biased estimates at higher or lower density values. In addition, the grid-based DÌ‚ analyses using full MMDM values for WÌ‚ area adjustments required a number of theoretical assumptions of uncertain validity, and we therefore viewed their empirical successes with caution. Finally, density estimates from the independent analysts were highly variable, but estimates from web-based approaches had smaller mean square errors and better achieved confidence-interval coverage of D than did grid-based approaches. Our results support the contention that web-based approaches for density estimation of small-mammal populations are both theoretically and empirically superior to grid-based approaches, even when sample size is far less than often recommended. In view of the increasing need for standardized environmental measures for comparisons among ecosystems and through time, analytical models based on distance sampling appear to offer accurate density estimation approaches for research

  14. SHOP: scaffold hopping by GRID-based similarity searches

    DEFF Research Database (Denmark)

    Bergmann, Rikke; Linusson, Anna; Zamora, Ismael

    2007-01-01

    A new GRID-based method for scaffold hopping (SHOP) is presented. In a fully automatic manner, scaffolds were identified in a database based on three types of 3D-descriptors. SHOP's ability to recover scaffolds was assessed and validated by searching a database spiked with fragments of known...... scaffolds were in the 31 top-ranked scaffolds. SHOP also identified new scaffolds with substantially different chemotypes from the queries. Docking analysis indicated that the new scaffolds would have similar binding modes to those of the respective query scaffolds observed in X-ray structures...

  15. Web-based computer adaptive assessment of individual perceptions of job satisfaction for hospital workplace employees.

    Science.gov (United States)

    Chien, Tsair-Wei; Lai, Wen-Pin; Lu, Chih-Wei; Wang, Weng-Chung; Chen, Shih-Chung; Wang, Hsien-Yi; Su, Shih-Bin

    2011-04-17

    To develop a web-based computer adaptive testing (CAT) application for efficiently collecting data regarding workers' perceptions of job satisfaction, we examined whether a 37-item Job Content Questionnaire (JCQ-37) could evaluate the job satisfaction of individual employees as a single construct. The JCQ-37 makes data collection via CAT on the internet easy, viable and fast. A Rasch rating scale model was applied to analyze data from 300 randomly selected hospital employees who participated in job-satisfaction surveys in 2008 and 2009 via non-adaptive and computer-adaptive testing, respectively. Of the 37 items on the questionnaire, 24 items fit the model fairly well. Person-separation reliability for the 2008 surveys was 0.88. Measures from both years and item-8 job satisfaction for groups were successfully evaluated through item-by-item analyses by using t-test. Workers aged 26 - 35 felt that job satisfaction was significantly worse in 2009 than in 2008. A Web-CAT developed in the present paper was shown to be more efficient than traditional computer-based or pen-and-paper assessments at collecting data regarding workers' perceptions of job content.

  16. Web-based computer adaptive assessment of individual perceptions of job satisfaction for hospital workplace employees

    Directory of Open Access Journals (Sweden)

    Chen Shih-Chung

    2011-04-01

    Full Text Available Abstract Background To develop a web-based computer adaptive testing (CAT application for efficiently collecting data regarding workers' perceptions of job satisfaction, we examined whether a 37-item Job Content Questionnaire (JCQ-37 could evaluate the job satisfaction of individual employees as a single construct. Methods The JCQ-37 makes data collection via CAT on the internet easy, viable and fast. A Rasch rating scale model was applied to analyze data from 300 randomly selected hospital employees who participated in job-satisfaction surveys in 2008 and 2009 via non-adaptive and computer-adaptive testing, respectively. Results Of the 37 items on the questionnaire, 24 items fit the model fairly well. Person-separation reliability for the 2008 surveys was 0.88. Measures from both years and item-8 job satisfaction for groups were successfully evaluated through item-by-item analyses by using t-test. Workers aged 26 - 35 felt that job satisfaction was significantly worse in 2009 than in 2008. Conclusions A Web-CAT developed in the present paper was shown to be more efficient than traditional computer-based or pen-and-paper assessments at collecting data regarding workers' perceptions of job content.

  17. BLAST in Gid (BiG): A Grid-Enabled Software Architecture and Implementation of Parallel and Sequential BLAST

    International Nuclear Information System (INIS)

    Aparicio, G.; Blanquer, I.; Hernandez, V.; Segrelles, D.

    2007-01-01

    The integration of High-performance computing tools is a key issue in biomedical research. Many computer-based applications have been migrated to High-Performance computers to deal with their computing and storage needs such as BLAST. However, the use of clusters and computing farm presents problems in scalability. The use of a higher layer of parallelism that splits the task into highly independent long jobs that can be executed in parallel can improve the performance maintaining the efficiency. Grid technologies combined with parallel computing resources are an important enabling technology. This work presents a software architecture for executing BLAST in a International Grid Infrastructure that guarantees security, scalability and fault tolerance. The software architecture is modular an adaptable to many other high-throughput applications, both inside the field of bio computing and outside. (Author)

  18. Active Power Quality Improvement Strategy for Grid-connected Microgrid Based on Hierarchical Control

    DEFF Research Database (Denmark)

    Wei, Feng; Sun, Kai; Guan, Yajuan

    2018-01-01

    proposes an active, unbalanced, and harmonic GCC suppression strategy based on hierarchical theory. The voltage error between the bus of the DCGC-MG and the grid’s PCC was transformed to the dq frame. On the basis of the grid, an additional compensator, which consists of multiple resonant voltage......When connected to a distorted grid utility, droop-controlled grid-connected microgrids (DCGC-MG) exhibit low equivalent impedance. The harmonic and unbalanced voltage at the point of common coupling (PCC) deteriorates the power quality of the grid-connected current (GCC) of DCGC-MG. This work...... regulators, was then added to the original secondary control to generate the negative fundamental and unbalanced harmonic voltage reference. Proportional integral and multiple resonant controllers were adopted as voltage controller at the original primary level to improve the voltage tracking performance...

  19. Operation of an InGrid based X-ray detector at the CAST experiment

    Directory of Open Access Journals (Sweden)

    Krieger Christoph

    2018-01-01

    During operation at the experiment, background rates in the order of 10−5 keV−1 cm−2 s−1 have been achieved by application of a likelihood based method discriminating the non-photon background originating mostly from cosmic rays. For continued operation in 2016, an upgraded InGrid based detector is to be installed among other improvements including decoupling and sampling of the signal induced on the grid as well as a veto scintillator to further lower the observed background rates and improving sensitivity.

  20. Air Pollution Monitoring and Mining Based on Sensor Grid in London

    Science.gov (United States)

    Ma, Yajie; Richards, Mark; Ghanem, Moustafa; Guo, Yike; Hassard, John

    2008-01-01

    In this paper, we present a distributed infrastructure based on wireless sensors network and Grid computing technology for air pollution monitoring and mining, which aims to develop low-cost and ubiquitous sensor networks to collect real-time, large scale and comprehensive environmental data from road traffic emissions for air pollution monitoring in urban environment. The main informatics challenges in respect to constructing the high-throughput sensor Grid are discussed in this paper. We present a two-layer network framework, a P2P e-Science Grid architecture, and the distributed data mining algorithm as the solutions to address the challenges. We simulated the system in TinyOS to examine the operation of each sensor as well as the networking performance. We also present the distributed data mining result to examine the effectiveness of the algorithm. PMID:27879895

  1. Air Pollution Monitoring and Mining Based on Sensor Grid in London

    Directory of Open Access Journals (Sweden)

    John Hassard

    2008-06-01

    Full Text Available In this paper, we present a distributed infrastructure based on wireless sensors network and Grid computing technology for air pollution monitoring and mining, which aims to develop low-cost and ubiquitous sensor networks to collect real-time, large scale and comprehensive environmental data from road traffic emissions for air pollution monitoring in urban environment. The main informatics challenges in respect to constructing the high-throughput sensor Grid are discussed in this paper. We present a twolayer network framework, a P2P e-Science Grid architecture, and the distributed data mining algorithm as the solutions to address the challenges. We simulated the system in TinyOS to examine the operation of each sensor as well as the networking performance. We also present the distributed data mining result to examine the effectiveness of the algorithm.

  2. Indonesia - Green Prosperity: Community-Based Off-Grid Renewable Energy Grant Portfolio

    Data.gov (United States)

    Millennium Challenge Corporation — Taken as a whole, this evaluation aims, to the extent possible, to validate the program logic underlying the portfolio of community-based off-grid renewable energy...

  3. New 2D adaptive mesh refinement algorithm based on conservative finite-differences with staggered grid

    Science.gov (United States)

    Gerya, T.; Duretz, T.; May, D. A.

    2012-04-01

    We present new 2D adaptive mesh refinement (AMR) algorithm based on stress-conservative finite-differences formulated for non-uniform rectangular staggered grid. The refinement approach is based on a repetitive cell splitting organized via a quad-tree construction (every parent cell is split into 4 daughter cells of equal size). Irrespective of the level of resolution every cell has 5 staggered nodes (2 horizontal velocities, 2 vertical velocities and 1 pressure) for which respective governing equations, boundary conditions and interpolation equations are formulated. The connectivity of the grid is achieved via cross-indexing of grid cells and basic nodal points located in their corners: four corner nodes are indexed for every cell and up to 4 surrounding cells are indexed for every node. The accuracy of the approach depends critically on the formulation of the stencil used at the "hanging" velocity nodes located at the boundaries between different levels of resolution. Most accurate results are obtained for the scheme based on the volume flux balance across the resolution boundary combined with stress-based interpolation of velocity orthogonal to the boundary. We tested this new approach with a number of 2D variable viscosity analytical solutions. Our tests demonstrate that the adaptive staggered grid formulation has convergence properties similar to those obtained in case of a standard, non-adaptive staggered grid formulation. This convergence is also achieved when resolution boundary crosses sharp viscosity contrast interfaces. The convergence rates measured are found to be insensitive to scenarios when the transition in grid resolution crosses sharp viscosity contrast interfaces. We compared various grid refinement strategies based on distribution of different field variables such as viscosity, density and velocity. According to these tests the refinement allows for significant (0.5-1 order of magnitude) increase in the computational accuracy at the same

  4. Deployment of the CMS software on the WLCG Grid

    International Nuclear Information System (INIS)

    Behrenhoff, W; Wissing, C; Kim, B; Blyweert, S; D'Hondt, J; Maes, J; Maes, M; Mulders, P Van; Villella, I; Vanelderen, L

    2011-01-01

    The CMS Experiment is taking high energy collision data at CERN. The computing infrastructure used to analyse the data is distributed round the world in a tiered structure. In order to use the 7 Tier-1 sites, the 50 Tier-2 sites and a still growing number of about 30 Tier-3 sites, the CMS software has to be available at those sites. Except for a very few sites the deployment and the removal of CMS software is managed centrally. Since the deployment team has no local accounts at the remote sites all installation jobs have to be sent via Grid jobs. Via a VOMS role the job has a high priority in the batch system and gains write privileges to the software area. Due to the lack of interactive access the installation jobs must be very robust against possible failures, in order not to leave a broken software installation. The CMS software is packaged in RPMs that are installed in the software area independent of the host OS. The apt-get tool is used to resolve package dependencies. This paper reports about the recent deployment experiences and the achieved performance.

  5. Job Design, Training Effect and Job Satisfaction: Evidence from Work Placement at Audit Firms

    Directory of Open Access Journals (Sweden)

    Liu Guangyou

    2017-01-01

    Full Text Available This paper aims to test the relationships between job design aspects and job satisfaction, as well as the mediating role of training effect in these relationships. Regression and mediation analyses were performed based on the data collected from a questionnaire-based survey on the senior accounting students’ audit work placement at audit firms. I conclude that repeated tedious non-professional job aspect is negatively related to job satisfaction, whereas judgmental professional job aspect is positively related to job satisfaction. I also conclude that training effect of work placement is playing a partial mediating role in the identified positive relationship while having no mediation in the negative one.

  6. Job satisfaction and intention to quit the job.

    Science.gov (United States)

    Suadicani, P; Bonde, J P; Olesen, K; Gyntelberg, F

    2013-03-01

    Negative psychosocial work conditions may influence the motivation of employees to adhere to their job. To elucidate the perception of psychosocial work conditions among Danish hospital employees who would quit their job if economically possible and those who would not. A cross-sectional questionnaire study of hospital employees. The questionnaire gave information on elements of the psychosocial work environment (job demands, job influence, job support, management quality, exposure to bullying), general health status, sick-leave during the preceding year, life style (leisure time physical activity, alcohol intake and smoking habits), age, sex and profession. There were 1809 participants with a response rate of 65%. About a quarter (26%) reported that they would quit their job if economically possible; this rose to 40% among the 17% who considered their health mediocre or bad. In a final logistic regression model, six factors were identified as independently associated with the wish to quit or not: self-assessed health status, meaningfulness of the job, quality of collaboration among colleagues, age, trustworthiness of closest superior(s) and exposure to bullying. Based on these factors it was possible to identify groups with fewer than 15% wishing to quit, and similarly, groups where 50% or more would quit if this was economically possible. Psychosocial work conditions, in particular meaningfulness of the job, were independently associated with intention to quit the job if economically possible and relevant within different job categories.

  7. Discrete Adjoint-Based Design Optimization of Unsteady Turbulent Flows on Dynamic Unstructured Grids

    Science.gov (United States)

    Nielsen, Eric J.; Diskin, Boris; Yamaleev, Nail K.

    2009-01-01

    An adjoint-based methodology for design optimization of unsteady turbulent flows on dynamic unstructured grids is described. The implementation relies on an existing unsteady three-dimensional unstructured grid solver capable of dynamic mesh simulations and discrete adjoint capabilities previously developed for steady flows. The discrete equations for the primal and adjoint systems are presented for the backward-difference family of time-integration schemes on both static and dynamic grids. The consistency of sensitivity derivatives is established via comparisons with complex-variable computations. The current work is believed to be the first verified implementation of an adjoint-based optimization methodology for the true time-dependent formulation of the Navier-Stokes equations in a practical computational code. Large-scale shape optimizations are demonstrated for turbulent flows over a tiltrotor geometry and a simulated aeroelastic motion of a fighter jet.

  8. Web Based Information System for Job Training Activities Using Personal Extreme Programming (PXP)

    Science.gov (United States)

    Asri, S. A.; Sunaya, I. G. A. M.; Rudiastari, E.; Setiawan, W.

    2018-01-01

    Job training is one of the subjects in university or polytechnic that involves many users and reporting activities. Time and distance became problems for users to reporting and to do obligations tasks during job training due to the location where the job training took place. This research tried to develop a web based information system of job training to overcome the problems. This system was developed using Personal Extreme Programming (PXP). PXP is one of the agile methods is combination of Extreme Programming (XP) and Personal Software Process (PSP). The information system that has developed and tested which are 24% of users are strongly agree, 74% are agree, 1% disagree and 0% strongly disagree about system functionality.

  9. Research and design of smart grid monitoring control via terminal based on iOS system

    Science.gov (United States)

    Fu, Wei; Gong, Li; Chen, Heli; Pan, Guangji

    2017-06-01

    Aiming at a series of problems existing in current smart grid monitoring Control Terminal, such as high costs, poor portability, simple monitoring system, poor software extensions, low system reliability when transmitting information, single man-machine interface, poor security, etc., smart grid remote monitoring system based on the iOS system has been designed. The system interacts with smart grid server so that it can acquire grid data through WiFi/3G/4G networks, and monitor each grid line running status, as well as power plant equipment operating conditions. When it occurs an exception in the power plant, incident information can be sent to the user iOS terminal equipment timely, which will provide troubleshooting information to help the grid staff to make the right decisions in a timely manner, to avoid further accidents. Field tests have shown the system realizes the integrated grid monitoring functions, low maintenance cost, friendly interface, high security and reliability, and it possesses certain applicable value.

  10. Grid for Earth Science Applications

    Science.gov (United States)

    Petitdidier, Monique; Schwichtenberg, Horst

    2013-04-01

    The civil society at large has addressed to the Earth Science community many strong requirements related in particular to natural and industrial risks, climate changes, new energies. The main critical point is that on one hand the civil society and all public ask for certainties i.e. precise values with small error range as it concerns prediction at short, medium and long term in all domains; on the other hand Science can mainly answer only in terms of probability of occurrence. To improve the answer or/and decrease the uncertainties, (1) new observational networks have been deployed in order to have a better geographical coverage and more accurate measurements have been carried out in key locations and aboard satellites. Following the OECD recommendations on the openness of research and public sector data, more and more data are available for Academic organisation and SMEs; (2) New algorithms and methodologies have been developed to face the huge data processing and assimilation into simulations using new technologies and compute resources. Finally, our total knowledge about the complex Earth system is contained in models and measurements, how we put them together has to be managed cleverly. The technical challenge is to put together databases and computing resources to answer the ES challenges. However all the applications are very intensive computing. Different compute solutions are available and depend on the characteristics of the applications. One of them is Grid especially efficient for independent or embarrassingly parallel jobs related to statistical and parametric studies. Numerous applications in atmospheric chemistry, meteorology, seismology, hydrology, pollution, climate and biodiversity have been deployed successfully on Grid. In order to fulfill requirements of risk management, several prototype applications have been deployed using OGC (Open geospatial Consortium) components with Grid middleware. The Grid has permitted via a huge number of runs to

  11. A comparative analysis of dynamic grids vs. virtual grids using the A3pviGrid framework.

    Science.gov (United States)

    Shankaranarayanan, Avinas; Amaldas, Christine

    2010-11-01

    With the proliferation of Quad/Multi-core micro-processors in mainstream platforms such as desktops and workstations; a large number of unused CPU cycles can be utilized for running virtual machines (VMs) as dynamic nodes in distributed environments. Grid services and its service oriented business broker now termed cloud computing could deploy image based virtualization platforms enabling agent based resource management and dynamic fault management. In this paper we present an efficient way of utilizing heterogeneous virtual machines on idle desktops as an environment for consumption of high performance grid services. Spurious and exponential increases in the size of the datasets are constant concerns in medical and pharmaceutical industries due to the constant discovery and publication of large sequence databases. Traditional algorithms are not modeled at handing large data sizes under sudden and dynamic changes in the execution environment as previously discussed. This research was undertaken to compare our previous results with running the same test dataset with that of a virtual Grid platform using virtual machines (Virtualization). The implemented architecture, A3pviGrid utilizes game theoretic optimization and agent based team formation (Coalition) algorithms to improve upon scalability with respect to team formation. Due to the dynamic nature of distributed systems (as discussed in our previous work) all interactions were made local within a team transparently. This paper is a proof of concept of an experimental mini-Grid test-bed compared to running the platform on local virtual machines on a local test cluster. This was done to give every agent its own execution platform enabling anonymity and better control of the dynamic environmental parameters. We also analyze performance and scalability of Blast in a multiple virtual node setup and present our findings. This paper is an extension of our previous research on improving the BLAST application framework

  12. Trends in life science grid: from computing grid to knowledge grid

    Directory of Open Access Journals (Sweden)

    Konagaya Akihiko

    2006-12-01

    Full Text Available Abstract Background Grid computing has great potential to become a standard cyberinfrastructure for life sciences which often require high-performance computing and large data handling which exceeds the computing capacity of a single institution. Results This survey reviews the latest grid technologies from the viewpoints of computing grid, data grid and knowledge grid. Computing grid technologies have been matured enough to solve high-throughput real-world life scientific problems. Data grid technologies are strong candidates for realizing "resourceome" for bioinformatics. Knowledge grids should be designed not only from sharing explicit knowledge on computers but also from community formulation for sharing tacit knowledge among a community. Conclusion Extending the concept of grid from computing grid to knowledge grid, it is possible to make use of a grid as not only sharable computing resources, but also as time and place in which people work together, create knowledge, and share knowledge and experiences in a community.

  13. Smart Solar Grid. Integration of high penetration of photovoltaic in municipal low voltage distribution grids; Smart Solar Grid. Integration hoher Anteile von Photovoltaik in kommunalen Niederspannungsverteilnetzen

    Energy Technology Data Exchange (ETDEWEB)

    Ruf, Holger; Heilscher, Gerd [Hochschule Ulm (Germany); Meier, Florian [SWU Netze GmbH, Ulm (Germany)

    2012-07-01

    The high rate of decentralized generation in low voltage grids especially photovoltaic (PV) put the distribution grid operators to new challenges. Grid operation and grid planning have to respect the volatility and dynamic of decentralized generation now and in the future and adapt their previous proceedings. In the frame of the project Smart Solar Grid was a test site defined in the grid area of the DSO Stadtwerke Ulm/Neu-Ulm GmbH (SWU) to analyze the impact of the PV rise and possible solutions for the grid planning in the future. The first analysis based upon secondly measurements of the first test site. From this were statistical evaluation of the load flows and power variations done. Furthermore were the roof potential analysis results of the test site validated. These data are the base for the development of a forecast system for grid condition parameter. (orig.)

  14. Micro grids toward the smart grid

    International Nuclear Information System (INIS)

    Guerrero, J.

    2011-01-01

    Worldwide electrical grids are expecting to become smarter in the near future, with interest in Microgrids likely to grow. A microgrid can be defined as a part of the grid with elements of prime energy movers, power electronics converters, distributed energy storage systems and local loads, that can operate autonomously but also interacting with main grid. Thus, the ability of intelligent Microgrids to operate in island mode or connected to the grid will be a keypoint to cope with new functionalities and the integration of renewable energy resources. The functionalities expected for these small grids are: black start operation, frequency and voltage stability, active and reactive power flow control, active power filter capabilities, and storage energy management. In this presentation, a review of the main concepts related to flexible Microgrids will be introduced, with examples of real Microgrids. AC and DC Microgrids to integrate renewable and distributed energy resources will also be presented, as well as distributed energy storage systems, and standardization issues of these Microgrids. Finally, Microgrid hierarchical control will be analyzed looking at three different levels: i) a primary control based on the droop method, including an output impedance virtual loop; ii) a secondary control, which enables restoring any deviations produced by the primary control; and iii) a tertiary control to manage the power flow between the microgrid and the external electrical distribution system.

  15. Modelling security properties in a grid-based operating system with anti-goals

    OpenAIRE

    Arenas, A.; Aziz, Benjamin; Bicarregui, J.; Matthews, B.; Yang, E.

    2008-01-01

    In this paper, we discuss the use of formal requirements-engineering techniques in capturing security requirements for a Grid-based operating system. We use KAOS goal model to represent two security goals for Grid systems, namely authorisation and single-sign on authentication. We apply goal-refinement to derive security requirements for these two security goals and we develop a model of antigoals and show how system vulnerabilities and threats to the security goals can arise from such anti-m...

  16. Synchrotron Imaging Computations on the Grid without the Computing Element

    International Nuclear Information System (INIS)

    Curri, A; Pugliese, R; Borghes, R; Kourousias, G

    2011-01-01

    Besides the heavy use of the Grid in the Synchrotron Radiation Facility (SRF) Elettra, additional special requirements from the beamlines had to be satisfied through a novel solution that we present in this work. In the traditional Grid Computing paradigm the computations are performed on the Worker Nodes of the grid element known as the Computing Element. A Grid middleware extension that our team has been working on, is that of the Instrument Element. In general it is used to Grid-enable instrumentation; and it can be seen as a neighbouring concept to that of the traditional Control Systems. As a further extension we demonstrate the Instrument Element as the steering mechanism for a series of computations. In our deployment it interfaces a Control System that manages a series of computational demanding Scientific Imaging tasks in an online manner. The instrument control in Elettra is done through a suitable Distributed Control System, a common approach in the SRF community. The applications that we present are for a beamline working in medical imaging. The solution resulted to a substantial improvement of a Computed Tomography workflow. The near-real-time requirements could not have been easily satisfied from our Grid's middleware (gLite) due to the various latencies often occurred during the job submission and queuing phases. Moreover the required deployment of a set of TANGO devices could not have been done in a standard gLite WN. Besides the avoidance of certain core Grid components, the Grid Security infrastructure has been utilised in the final solution.

  17. Grid-based virtual clinic for medical diagnosis tutoring | Yatchou ...

    African Journals Online (AJOL)

    La réalisation visée est un outil collaboratif d\\'enseignement pour les médecins du terrain et les étudiants en médecine au sein d\\'une organisation virtuelle. The emerging grid-based technologies are increasingly adopted to enhance education and provide better learning services. This is characterized all over the world, ...

  18. Thyristor based short circuit current injection in isolated grids

    OpenAIRE

    Hoff, Bjarte; Sharma, Pawan; Østrem, Trond

    2017-01-01

    This paper proposes a thyristor based short circuit current injector for providing short circuit current in isolated and weak grids, where sufficient fault current to trigger circuit breakers may not be available. This will allow the use of conventional miniature circuit breakers, which requires high fault current for instantaneous tripping. The method has been validated through experiments.

  19. Grid interoperability: joining grid information systems

    International Nuclear Information System (INIS)

    Flechl, M; Field, L

    2008-01-01

    A grid is defined as being 'coordinated resource sharing and problem solving in dynamic, multi-institutional virtual organizations'. Over recent years a number of grid projects, many of which have a strong regional presence, have emerged to help coordinate institutions and enable grids. Today, we face a situation where a number of grid projects exist, most of which are using slightly different middleware. Grid interoperation is trying to bridge these differences and enable Virtual Organizations to access resources at the institutions independent of their grid project affiliation. Grid interoperation is usually a bilateral activity between two grid infrastructures. Recently within the Open Grid Forum, the Grid Interoperability Now (GIN) Community Group is trying to build upon these bilateral activities. The GIN group is a focal point where all the infrastructures can come together to share ideas and experiences on grid interoperation. It is hoped that each bilateral activity will bring us one step closer to the overall goal of a uniform grid landscape. A fundamental aspect of a grid is the information system, which is used to find available grid services. As different grids use different information systems, interoperation between these systems is crucial for grid interoperability. This paper describes the work carried out to overcome these differences between a number of grid projects and the experiences gained. It focuses on the different techniques used and highlights the important areas for future standardization

  20. A Generic Danish Distribution Grid Model for Smart Grid Technology Testing

    DEFF Research Database (Denmark)

    Cha, Seung-Tae; Wu, Qiuwei; Østergaard, Jacob

    2012-01-01

    This paper describes the development of a generic Danish distribution grid model for smart grid technology testing based on the Bornholm power system. The frequency dependent network equivalent (FDNE) method has been used in order to accurately preserve the desired properties and characteristics...... as a generic Smart Grid benchmark model for testing purposes....... by comparing the transient response of the original Bornholm power system model and the developed generic model under significant fault conditions. The results clearly show that the equivalent generic distribution grid model retains the dynamic characteristics of the original system, and can be used...

  1. A Novel Quad Harmony Search Algorithm for Grid-Based Path Finding

    Directory of Open Access Journals (Sweden)

    Saso Koceski

    2014-09-01

    Full Text Available A novel approach to the problem of grid-based path finding has been introduced. The method is a block-based search algorithm, founded on the bases of two algorithms, namely the quad-tree algorithm, which offered a great opportunity for decreasing the time needed to compute the solution, and the harmony search (HS algorithm, a meta-heuristic algorithm used to obtain the optimal solution. This quad HS algorithm uses the quad-tree decomposition of free space in the grid to mark the free areas and treat them as a single node, which greatly improves the execution. The results of the quad HS algorithm have been compared to other meta-heuristic algorithms, i.e., ant colony, genetic algorithm, particle swarm optimization and simulated annealing, and it was proved to obtain the best results in terms of time and giving the optimal path.

  2. Phase-lock loop of Grid-connected Voltage Source Converter under non-ideal grid condition

    DEFF Research Database (Denmark)

    Wang, Haojie; Sun, Hai; Han, Minxiao

    2015-01-01

    It is a normal practice that the DC micro-grid is connected to AC main grid through Grid-connected Voltage Source Converter (G-VSC) for voltage support. Accurate control of DC micro-grid voltage is difficult for G-VSC under unbalanced grid condition as the fundamental positive-sequence component...... and distorted system voltage the proposed PLL can accurately detect the fundamental positive-sequence component of grid voltage thus accurate control of DC micro-grid voltage can be realized....... phase information cannot be accurately tracked. Based on analysis of the cause of double-frequency ripple when unbalance exists in main grid, a phase-locked loop (PLL) detection technique is proposed. Under the conditions of unsymmetrical system voltage, varying system frequency, single-phase system...

  3. A New Family of Multilevel Grid Connected Inverters Based on Packed U Cell Topology.

    Science.gov (United States)

    Pakdel, Majid; Jalilzadeh, Saeid

    2017-09-29

    In this paper a novel packed U cell (PUC) based multilevel grid connected inverter is proposed. Unlike the U cell arrangement which consists of two power switches and one capacitor, in the proposed converter topology a lower DC power supply from renewable energy resources such as photovoltaic arrays (PV) is used as a base power source. The proposed topology offers higher efficiency and lower cost using a small number of power switches and a lower DC power source which is supplied from renewable energy resources. Other capacitor voltages are extracted from the base lower DC power source using isolated DC-DC power converters. The operation principle of proposed transformerless multilevel grid connected inverter is analyzed theoretically. Operation of the proposed multilevel grid connected inverter is verified through simulation studies. An experimental prototype using STM32F407 discovery controller board is performed to verify the simulation results.

  4. Neighboring Structure Visualization on a Grid-based Layout.

    Science.gov (United States)

    Marcou, G; Horvath, D; Varnek, A

    2017-10-01

    Here, we describe an algorithm to visualize chemical structures on a grid-based layout in such a way that similar structures are neighboring. It is based on structure reordering with the help of the Hilbert Schmidt Independence Criterion, representing an empirical estimate of the Hilbert-Schmidt norm of the cross-covariance operator. The method can be applied to any layout of bi- or three-dimensional shape. The approach is demonstrated on a set of dopamine D5 ligands visualized on squared, disk and spherical layouts. © 2017 Wiley-VCH Verlag GmbH & Co. KGaA, Weinheim.

  5. Silicon-based metallic micro grid for electron field emission

    International Nuclear Information System (INIS)

    Kim, Jaehong; Jeon, Seok-Gy; Kim, Jung-Il; Kim, Geun-Ju; Heo, Duchang; Shin, Dong Hoon; Sun, Yuning; Lee, Cheol Jin

    2012-01-01

    A micro-scale metal grid based on a silicon frame for application to electron field emission devices is introduced and experimentally demonstrated. A silicon lattice containing aperture holes with an area of 80 × 80 µm 2 and a thickness of 10 µm is precisely manufactured by dry etching the silicon on one side of a double-polished silicon wafer and by wet etching the opposite side. Because a silicon lattice is more rigid than a pure metal lattice, a thin layer of Au/Ti deposited on the silicon lattice for voltage application can be more resistant to the geometric stress caused by the applied electric field. The micro-fabrication process, the images of the fabricated grid with 88% geometric transparency and the surface profile measurement after thermal feasibility testing up to 700 °C are presented. (paper)

  6. Cultural-Based Genetic Tabu Algorithm for Multiobjective Job Shop Scheduling

    Directory of Open Access Journals (Sweden)

    Yuzhen Yang

    2014-01-01

    Full Text Available The job shop scheduling problem, which has been dealt with by various traditional optimization methods over the decades, has proved to be an NP-hard problem and difficult in solving, especially in the multiobjective field. In this paper, we have proposed a novel quadspace cultural genetic tabu algorithm (QSCGTA to solve such problem. This algorithm provides a different structure from the original cultural algorithm in containing double brief spaces and population spaces. These spaces deal with different levels of populations globally and locally by applying genetic and tabu searches separately and exchange information regularly to make the process more effective towards promising areas, along with modified multiobjective domination and transform functions. Moreover, we have presented a bidirectional shifting for the decoding process of job shop scheduling. The computational results we presented significantly prove the effectiveness and efficiency of the cultural-based genetic tabu algorithm for the multiobjective job shop scheduling problem.

  7. Grid-enabled SEE++, A Grid-Based Medical Decision Support System for Eye Muscle Surgery Conference

    CERN Document Server

    Schreiner, W.; Buchberger, M.; Kaltofen, T.

    2006-01-01

    JKU/RISC currently develops in cooperation with Upper Austrian Research (UAR) the SEE-GRID software system. SEE-GRID is based on the SEE++ software for the biomechanical 3D simulation of the human eye and its muscles. SEE++ simulates the common eye muscle surgery techniques in a graphic interactive way that is familiar to an experienced surgeon. SEE++ is world-wide the most advanced software for this purpose; it is used by various hospitals and medical doctors for surgery training and planning, SEE++ deals with the support of diagnosis and treatment of strabismus, which is the common name given to usually persistent or regularly occuring misalignment of the eyes. Strabismus is a visual defect in which eyes point in different directions. A person suffering from it may see double images due to misaligned eyes. SEE++ is able to simulate the result of the Hess-Lancaster test, from which the pathological reason of strabismus can be estimated. The outcome of such an examination is two gaze patterns of blue points a...

  8. The Czech National Grid Infrastructure

    Science.gov (United States)

    Chudoba, J.; Křenková, I.; Mulač, M.; Ruda, M.; Sitera, J.

    2017-10-01

    The Czech National Grid Infrastructure is operated by MetaCentrum, a CESNET department responsible for coordinating and managing activities related to distributed computing. CESNET as the Czech National Research and Education Network (NREN) provides many e-infrastructure services, which are used by 94% of the scientific and research community in the Czech Republic. Computing and storage resources owned by different organizations are connected by fast enough network to provide transparent access to all resources. We describe in more detail the computing infrastructure, which is based on several different technologies and covers grid, cloud and map-reduce environment. While the largest part of CPUs is still accessible via distributed torque servers, providing environment for long batch jobs, part of infrastructure is available via standard EGI tools in EGI, subset of NGI resources is provided into EGI FedCloud environment with cloud interface and there is also Hadoop cluster provided by the same e-infrastructure.A broad spectrum of computing servers is offered; users can choose from standard 2 CPU servers to large SMP machines with up to 6 TB of RAM or servers with GPU cards. Different groups have different priorities on various resources, resource owners can even have an exclusive access. The software is distributed via AFS. Storage servers offering up to tens of terabytes of disk space to individual users are connected via NFS4 on top of GPFS and access to long term HSM storage with peta-byte capacity is also provided. Overview of available resources and recent statistics of usage will be given.

  9. Investigating the effect of job stress and emotional intelligence on job performance

    Directory of Open Access Journals (Sweden)

    Mojtaba Rafiee

    2013-09-01

    Full Text Available Researchers and scholars of management and behavioral sciences have tried to determine effective factors, which influence on efficiency and effectiveness in order to increase organization performance and they have tried to identify factors, which create job stress. In this research, we investigate the effect of job stress on job performance through emotional, organizational and moral intelligence. The study is a descriptive-analytic one, which is based on correlation, uses survey method to gather data and they are analyzed using structural equation modeling. The population of this research includes all the personnel of Registry Organization in Arak city. The results suggest that job stress influences on job performance through organizational intelligence and moral intelligence, but job stress does not influence on job performance through emotional intelligence. Regarding research hypotheses, results and findings after analyzing obtained data suggest that job stress influences on emotional, organizational and moral intelligence, but job stress does not influence on job performance. In addition, the results show that organizational and moral intelligence influence on job performance but emotional intelligence does not influence on job performance.

  10. The play grid

    DEFF Research Database (Denmark)

    Fogh, Rune; Johansen, Asger

    2013-01-01

    In this paper we propose The Play Grid, a model for systemizing different play types. The approach is psychological by nature and the actual Play Grid is based, therefore, on two pairs of fundamental and widely acknowledged distinguishing characteristics of the ego, namely: extraversion vs. intro...

  11. Integration of distributed energy resources into low voltage grid: A market-based multiperiod optimization model

    Energy Technology Data Exchange (ETDEWEB)

    Mashhour, Elahe; Moghaddas-Tafreshi, S.M. [Faculty of Electrical Engineering, K.N. Toosi University of Technology, Seyd Khandan, P.O. Box 16315-1355, Shariati, Tehran (Iran)

    2010-04-15

    This paper develops a multiperiod optimization model for an interconnected micro grid with hierarchical control that participates in wholesale energy market to maximize its benefit (i.e. revenues-costs). In addition to the operational constraints of distributed energy resources (DER) including both inter-temporal and non-inter-temporal types, the adequacy and steady-state security constraints of micro grid and its power losses are incorporated in the optimization model. In the presented model, DER are integrated into low voltage grid considering both technical and economical aspects. This integration as a micro grid can participate in wholesale energy market as an entity with dual role including producer and consumer based on the direction of exchanged power. The developed model is evaluated by testing on a micro grid considering different cases and the results are analyzed. (author)

  12. PERFORMANCE BASED PAY AS A DETERMINANT OF JOB SATISFACTION: A STUDY IN MALAYSIA GIATMARA CENTERS

    Directory of Open Access Journals (Sweden)

    Azman ISMAIL

    2011-01-01

    Full Text Available Compensation management literature highlights that performance based payhas two major characteristics: participation in pay systems and adequacy ofpay. The ability of management to properly implement such pay systems maylead to increased job satisfaction in organizations. Though, the nature of thisrelationship is interesting, little is known about the influence of performancebased pay on job satisfaction in compensation management literature.Therefore, this study was conducted to examine the relationship between payfor performance and job satisfaction in Malaysian GIATMARA centers. Theresults of exploratory factor analysis confirmed that measurement scalesused in this study satisfactorily met the standards of validity and reliabilityanalyses. An outcome of stepwise regression analysis shows thatdeterminant of job satisfaction is performance based pay. Further, this resultconfirms that pay for performance is an important antecedent for jobsatisfaction in the studied organizations.

  13. A dynamic system for ATLAS software installation on OSG grid sites

    International Nuclear Information System (INIS)

    Zhao, X; Maeno, T; Wenaus, T; Leuhring, F; Youssef, S; Brunelle, J; De Salvo, A; Thompson, A S

    2010-01-01

    A dynamic and reliable system for installing the ATLAS software releases on Grid sites is crucial to guarantee the timely and smooth start of ATLAS production and reduce its failure rate. In this paper, we discuss the issues encountered in the previous software installation system, and introduce the new approach, which is built upon the new development in the areas of the ATLAS workload management system (PanDA), and software package management system (pacman). It is also designed to integrate with the EGEE ATLAS software installation framework. In the new system, ATLAS software releases are packaged as pacball, a uniquely identifiable and reproducible self-installing data file. The distribution of pacballs to remote sites is managed by ATLAS data management system (DQ2) and PanDA server. The installation on remote sites is automatically triggered by the PanDA pilot jobs. The installation job payload connects to a central ATLAS software installation portal, making the information of installation status easily accessible across OSG and EGEE Grids. The issues encountered in running the new system in production, and our future plan for improvement, will also be discussed.

  14. A Damping Grid Strapdown Inertial Navigation System Based on a Kalman Filter for Ships in Polar Regions.

    Science.gov (United States)

    Huang, Weiquan; Fang, Tao; Luo, Li; Zhao, Lin; Che, Fengzhu

    2017-07-03

    The grid strapdown inertial navigation system (SINS) used in polar navigation also includes three kinds of periodic oscillation errors as common SINS are based on a geographic coordinate system. Aiming ships which have the external information to conduct a system reset regularly, suppressing the Schuler periodic oscillation is an effective way to enhance navigation accuracy. The Kalman filter based on the grid SINS error model which applies to the ship is established in this paper. The errors of grid-level attitude angles can be accurately estimated when the external velocity contains constant error, and then correcting the errors of the grid-level attitude angles through feedback correction can effectively dampen the Schuler periodic oscillation. The simulation results show that with the aid of external reference velocity, the proposed external level damping algorithm based on the Kalman filter can suppress the Schuler periodic oscillation effectively. Compared with the traditional external level damping algorithm based on the damping network, the algorithm proposed in this paper can reduce the overshoot errors when the state of grid SINS is switched from the non-damping state to the damping state, and this effectively improves the navigation accuracy of the system.

  15. Development of an international matrix-solver prediction system on a French-Japanese international grid computing environment

    International Nuclear Information System (INIS)

    Suzuki, Yoshio; Kushida, Noriyuki; Tatekawa, Takayuki; Teshima, Naoya; Caniou, Yves; Guivarch, Ronan; Dayde, Michel; Ramet, Pierre

    2010-01-01

    The 'Research and Development of International Matrix-Solver Prediction System (REDIMPS)' project aimed at improving the TLSE sparse linear algebra expert website by establishing an international grid computing environment between Japan and France. To help users in identifying the best solver or sparse linear algebra tool for their problems, we have developed an interoperable environment between French and Japanese grid infrastructures (respectively managed by DIET and AEGIS). Two main issues were considered. The first issue is how to submit a job from DIET to AEGIS. The second issue is how to bridge the difference of security between DIET and AEGIS. To overcome these issues, we developed APIs to communicate between different grid infrastructures by improving the client API of AEGIS. By developing a server deamon program (SeD) of DIET which behaves like an AEGIS user, DIET can call functions in AEGIS: authentication, file transfer, job submission, and so on. To intensify the security, we also developed functionalities to authenticate DIET sites and DIET users in order to access AEGIS computing resources. By this study, the set of software and computers available within TLSE to find an appropriate solver is enlarged over France (DIET) and Japan (AEGIS). (author)

  16. Age-Differential Effects of Job Characteristics on Job Attraction: A Policy-Capturing Study.

    Science.gov (United States)

    Zacher, Hannes; Dirkers, Bodil T; Korek, Sabine; Hughes, Brenda

    2017-01-01

    Based on an integration of job design and lifespan developmental theories, Truxillo et al. (2012) proposed that job characteristics interact with employee age in predicting important work outcomes. Using an experimental policy-capturing design, we investigated age-differential effects of four core job characteristics (i.e., job autonomy, task variety, task significance, and feedback from the job) on job attraction (i.e., individuals' rating of job attractiveness). Eighty-two employees between 19 and 65 years ( M age = 41, SD = 14) indicated their job attraction for each of 40 hypothetical job descriptions in which the four job characteristics were systematically manipulated (in total, participants provided 3,280 ratings). Results of multilevel analyses showed that the positive effects of task variety, task significance, and feedback from the job were stronger for younger compared to older employees, whereas we did not find significant age-differential effects of job autonomy on job attraction. These findings are only partially consistent with propositions of Truxillo et al.'s (2012) lifespan perspective on job design.

  17. Exploring Horticultural Employees' Attitudes Toward Their Jobs: A Qualitative Analysis Based on Herzberg's Theory of Job Satisfaction

    OpenAIRE

    Bitsch, Vera; Hogberg, Michael

    2005-01-01

    Job satisfaction is likely the most studied work-related attitude and is assumed to influence a variety of behaviors. This study analyzes the job satisfaction of agricultural employees using Herzberg’s theory, which is broadly employed in management. Fourteen horticultural businesses participated in case studies of labor-management practices. Fifteen nonsupervisory employee interviews were analyzed regarding job satisfaction. Components of job satisfaction relevant to horticultural employee...

  18. Smart homes as a base for smart grids; Smart Home als Basis fuer Smart Grid

    Energy Technology Data Exchange (ETDEWEB)

    Segbusch, Klaus von [ABB AG, Mannheim (Germany). Team Business Development Smart Grids; Struwe, Christian [Busch-Jaeger Elektro GmbH, Luedenscheid (Germany)

    2010-09-15

    Integration of renewable energy sources requires more intelligent distribution systems, i.e. so-called smart grids. For this, it is necessary to integrate the end customers in grid operation, giving them financial incentives, information in near real time from the utility, and means for automatic control of their consumption. (orig.)

  19. Novel grid-based optical Braille conversion: from scanning to wording

    Science.gov (United States)

    Yoosefi Babadi, Majid; Jafari, Shahram

    2011-12-01

    Grid-based optical Braille conversion (GOBCO) is explained in this article. The grid-fitting technique involves processing scanned images taken from old hard-copy Braille manuscripts, recognising and converting them into English ASCII text documents inside a computer. The resulted words are verified using the relevant dictionary to provide the final output. The algorithms employed in this article can be easily modified to be implemented on other visual pattern recognition systems and text extraction applications. This technique has several advantages including: simplicity of the algorithm, high speed of execution, ability to help visually impaired persons and blind people to work with fax machines and the like, and the ability to help sighted people with no prior knowledge of Braille to understand hard-copy Braille manuscripts.

  20. Analysis of turbine-grid interaction of grid-connected wind turbine using HHT

    Science.gov (United States)

    Chen, A.; Wu, W.; Miao, J.; Xie, D.

    2018-05-01

    This paper processes the output power of the grid-connected wind turbine with the denoising and extracting method based on Hilbert Huang transform (HHT) to discuss the turbine-grid interaction. At first, the detailed Empirical Mode Decomposition (EMD) and the Hilbert Transform (HT) are introduced. Then, on the premise of decomposing the output power of the grid-connected wind turbine into a series of Intrinsic Mode Functions (IMFs), energy ratio and power volatility are calculated to detect the unessential components. Meanwhile, combined with vibration function of turbine-grid interaction, data fitting of instantaneous amplitude and phase of each IMF is implemented to extract characteristic parameters of different interactions. Finally, utilizing measured data of actual parallel-operated wind turbines in China, this work accurately obtains the characteristic parameters of turbine-grid interaction of grid-connected wind turbine.

  1. Utility-based Reinforcement Learning for Reactive Grids

    OpenAIRE

    Perez , Julien; Germain-Renaud , Cécile; Kégl , Balázs; Loomis , C.

    2008-01-01

    International audience; Large scale production grids are an important case for autonomic computing. They follow a mutualization paradigm: decision-making (human or automatic) is distributed and largely independent, and, at the same time, it must implement the highlevel goals of the grid management. This paper deals with the scheduling problem with two partially conflicting goals: fairshare and Quality of Service (QoS). Fair sharing is a wellknown issue motivated by return on investment for pa...

  2. Grid connection of active stall wind farms using a VSC based DC transmission system

    DEFF Research Database (Denmark)

    Iov, F.; Sørensen, Poul Ejnar; Hansen, A.D.

    2005-01-01

    Currently, there is an increasing trend to connect large MW wind farms to the transmission system. Requirements that focus on the influence of the farms on the grid stability and power quality, and on the control capabilities of wind farms have already been established. The main trends of modern...... wind turbines/farms are clearly the variable speed operation and a grid connection through a power electronic interface, especially using doubly fed induction generators. Using power electronics the control capabilities of these wind turbines/farms are extended and thus the grid requirements...... are fulfilled. However, the traditional squirrel-cage generators based wind turbines/wind farms directly connected to the grid have less control capabilities. These wind turbines/farms cannot regulate their production and contribute to power system stability. A DC transmission system for connection...

  3. H∞ Robust Current Control for DFIG Based Wind Turbine subject to Grid Voltage Distortions

    DEFF Research Database (Denmark)

    Wang, Yun; Wu, Qiuwei; Gong, Wenming

    2016-01-01

    This paper proposes an H∞ robust current controller for doubly fed induction generator (DFIG) based wind turbines (WTs) subject to grid voltage distortions. The controller is to mitigate the impact of the grid voltage distortions on rotor currents with DFIG parameter perturbation. The grid voltage...... distortions considered include asymmetric voltage dips and grid background harmonics. An uncertain DFIG model is developed with uncertain factors originating from distorted stator voltage, and changed generator parameters due to the flux saturation effect, the skin effect, etc. Weighting functions...... are designed to efficiently track the unbalanced current components and the 5th and 7th background harmonics. The robust stability (RS) and robust performance (RP) of the proposed controller are verified by the structured singular value µ. The performance of the H∞ robust current controller was demonstrated...

  4. A comprehensive WSN-based approach to efficiently manage a Smart Grid.

    Science.gov (United States)

    Martinez-Sandoval, Ruben; Garcia-Sanchez, Antonio-Javier; Garcia-Sanchez, Felipe; Garcia-Haro, Joan; Flynn, David

    2014-10-10

    The Smart Grid (SG) is conceived as the evolution of the current electrical grid representing a big leap in terms of efficiency, reliability and flexibility compared to today's electrical network. To achieve this goal, the Wireless Sensor Networks (WSNs) are considered by the scientific/engineering community to be one of the most suitable technologies to apply SG technology to due to their low-cost, collaborative and long-standing nature. However, the SG has posed significant challenges to utility operators-mainly very harsh radio propagation conditions and the lack of appropriate systems to empower WSN devices-making most of the commercial widespread solutions inadequate. In this context, and as a main contribution, we have designed a comprehensive ad-hoc WSN-based solution for the Smart Grid (SENSED-SG) that focuses on specific implementations of the MAC, the network and the application layers to attain maximum performance and to successfully deal with any arising hurdles. Our approach has been exhaustively evaluated by computer simulations and mathematical analysis, as well as validation within real test-beds deployed in controlled environments. In particular, these test-beds cover two of the main scenarios found in a SG; on one hand, an indoor electrical substation environment, implemented in a High Voltage AC/DC laboratory, and, on the other hand, an outdoor case, deployed in the Transmission and Distribution segment of a power grid. The results obtained show that SENSED-SG performs better and is more suitable for the Smart Grid than the popular ZigBee WSN approach.

  5. A derived heuristics based multi-objective optimization procedure for micro-grid scheduling

    Science.gov (United States)

    Li, Xin; Deb, Kalyanmoy; Fang, Yanjun

    2017-06-01

    With the availability of different types of power generators to be used in an electric micro-grid system, their operation scheduling as the load demand changes with time becomes an important task. Besides satisfying load balance constraints and the generator's rated power, several other practicalities, such as limited availability of grid power and restricted ramping of power output from generators, must all be considered during the operation scheduling process, which makes it difficult to decide whether the optimization results are accurate and satisfactory. In solving such complex practical problems, heuristics-based customized optimization algorithms are suggested. However, due to nonlinear and complex interactions of variables, it is difficult to come up with heuristics in such problems off-hand. In this article, a two-step strategy is proposed in which the first task deciphers important heuristics about the problem and the second task utilizes the derived heuristics to solve the original problem in a computationally fast manner. Specifically, the specific operation scheduling is considered from a two-objective (cost and emission) point of view. The first task develops basic and advanced level knowledge bases offline from a series of prior demand-wise optimization runs and then the second task utilizes them to modify optimized solutions in an application scenario. Results on island and grid connected modes and several pragmatic formulations of the micro-grid operation scheduling problem clearly indicate the merit of the proposed two-step procedure.

  6. Direct care worker's perceptions of job satisfaction following implementation of work-based learning.

    Science.gov (United States)

    Lopez, Cynthia; White, Diana L; Carder, Paula C

    2014-02-01

    The purpose of this study was to understand the impact of a work-based learning program on the work lives of Direct Care Workers (DCWs) at assisted living (AL) residences. The research questions were addressed using focus group data collected as part of a larger evaluation of a work-based learning (WBL) program called Jobs to Careers. The theoretical perspective of symbolic interactionism was used to frame the qualitative data analysis. Results indicated that the WBL program impacted DCWs' job satisfaction through the program curriculum and design and through three primary categories: relational aspects of work, worker identity, and finding time. This article presents a conceptual model for understanding how these categories are interrelated and the implications for WBL programs. Job satisfaction is an important topic that has been linked to quality of care and reduced turnover in long-term care settings.

  7. Project Scheduling Heuristics-Based Standard PSO for Task-Resource Assignment in Heterogeneous Grid

    Directory of Open Access Journals (Sweden)

    Ruey-Maw Chen

    2011-01-01

    Full Text Available The task scheduling problem has been widely studied for assigning resources to tasks in heterogeneous grid environment. Effective task scheduling is an important issue for the performance of grid computing. Meanwhile, the task scheduling problem is an NP-complete problem. Hence, this investigation introduces a named “standard“ particle swarm optimization (PSO metaheuristic approach to efficiently solve the task scheduling problems in grid. Meanwhile, two promising heuristics based on multimode project scheduling are proposed to help in solving interesting scheduling problems. They are the best performance resource heuristic and the latest finish time heuristic. These two heuristics applied to the PSO scheme are for speeding up the search of the particle and improving the capability of finding a sound schedule. Moreover, both global communication topology and local ring communication topology are also investigated for efficient study of proposed scheme. Simulation results demonstrate that the proposed approach in this investigation can successfully solve the task-resource assignment problems in grid computing and similar scheduling problems.

  8. Commutative discrete filtering on unstructured grids based on least-squares techniques

    International Nuclear Information System (INIS)

    Haselbacher, Andreas; Vasilyev, Oleg V.

    2003-01-01

    The present work is concerned with the development of commutative discrete filters for unstructured grids and contains two main contributions. First, building on the work of Marsden et al. [J. Comp. Phys. 175 (2002) 584], a new commutative discrete filter based on least-squares techniques is constructed. Second, a new analysis of the discrete commutation error is carried out. The analysis indicates that the discrete commutation error is not only dependent on the number of vanishing moments of the filter weights, but also on the order of accuracy of the discrete gradient operator. The results of the analysis are confirmed by grid-refinement studies

  9. Triple-layer smart grid business model

    DEFF Research Database (Denmark)

    Ma, Zheng; Lundgaard, Morten; Jørgensen, Bo Nørregaard

    2016-01-01

    Viewing the smart grid with the theory of business models may open opportunities in understanding and capturing values in new markets. This study tries to discover and map the smart grid ecosystem-based business model framework with two different environments (sub-Saharan Africa and Denmark......), and identifies the parameters for the smart grid solutions to the emerging markets. This study develops a triple-layer business model including the organizational (Niche), environmental (Intermediate), and global (Dominators) factors. The result uncovers an interface of market factors and stakeholders...... in a generic smart grid constellation. The findings contribute the transferability potential of the smart grid solutions between countries, and indicate the potential to export and import smart grid solutions based on the business modeling....

  10. Comprehensive analysis of the dynamic behavior of grid-connected DFIG-based wind turbines under LVRT conditions

    DEFF Research Database (Denmark)

    Alsmadi, Yazan M.; Xu, Longya; Blaabjerg, Frede

    2015-01-01

    ) capability of wind turbines during grid faults is one of the core requirements to ensure stability in the power grid during transients. The doubly-fed induction generators (DFIGs) offer several advantages when utilized in wind turbines, but discussions about their LVRT capabilities are limited. This paper...... presents a comprehensive study of the LVRT of grid-connected DFIG-based wind turbines. It provides a detailed investigation of the transient characteristics and the dynamic behavior of DFIGs during symmetrical and asymmetrical grid voltage sags. A detailed theoretical study supported by computer......Power generation and grid stability have become key issues in the last decade. The high penetration of large capacity wind generation into the electric power grid has led to serious concerns about their influence on the dynamic behavior of power systems. The Low-Voltage Ride-Through (LVRT...

  11. A Security Monitoring Framework For Virtualization Based HEP Infrastructures

    Science.gov (United States)

    Gomez Ramirez, A.; Martinez Pedreira, M.; Grigoras, C.; Betev, L.; Lara, C.; Kebschull, U.; ALICE Collaboration

    2017-10-01

    High Energy Physics (HEP) distributed computing infrastructures require automatic tools to monitor, analyze and react to potential security incidents. These tools should collect and inspect data such as resource consumption, logs and sequence of system calls for detecting anomalies that indicate the presence of a malicious agent. They should also be able to perform automated reactions to attacks without administrator intervention. We describe a novel framework that accomplishes these requirements, with a proof of concept implementation for the ALICE experiment at CERN. We show how we achieve a fully virtualized environment that improves the security by isolating services and Jobs without a significant performance impact. We also describe a collected dataset for Machine Learning based Intrusion Prevention and Detection Systems on Grid computing. This dataset is composed of resource consumption measurements (such as CPU, RAM and network traffic), logfiles from operating system services, and system call data collected from production Jobs running in an ALICE Grid test site and a big set of malware samples. This malware set was collected from security research sites. Based on this dataset, we will proceed to develop Machine Learning algorithms able to detect malicious Jobs.

  12. An Extended Design of the "Grid-Enabled SEE++ System" Based on Globus Toolkit 4 and gLite Conference

    CERN Document Server

    Schreiner, W.; Buchberger, M.; Kaltofen, T.

    2006-01-01

    "Grid-Enabled SEE++" based on the SEE++ software system for the biomechanical 3D simulation of the human eye and its muscles. SEE++ simulates the common eye muscle surgery techniques in a graphic interactive way that is familiar to an experienced surgeon. The goal of "Grid-Enabled SEE++" is to adapt and to extend SEE++ in several steps and to develop an efficient grid-based tool for "Evidence Based Medicine", which supports the surgeons in choosing optimal surgery techniques for the treatments of different syndromes of strabismus. In our previous work, we combined the SEE++ software with the Globus (pre-Web Service) middleware and developed a parallel version of the simulation of the "Hess-Lancaster test" (typical medical examination). By this, we demonstrated how a noticeable speedup can be achieved in SEE++ by the exploitation of the computational power of the Grid. Furthermore, we reported the prototype implementation of a medical database component for "Grid-Enabled SEE++". Finally, we designed a so calle...

  13. Active Distribution Grid Management based on Robust AC Optimal Power Flow

    DEFF Research Database (Denmark)

    Soares, Tiago; Bessa, Richard J.; Pinson, Pierre

    2017-01-01

    Further integration of distributed renewable energy sources in distribution systems requires a paradigm change in grid management by the distribution system operators (DSO). DSOs are currently moving to an operational planning approach based on activating flexibility from distributed energy resou...

  14. How should grid operators govern smart grid innovation projects? An embedded case study approach

    International Nuclear Information System (INIS)

    Reuver, Mark de; Lei, Telli van der; Lukszo, Zofia

    2016-01-01

    Grid operators increasingly have to collaborate with other actors in order to realize smart grid innovations. For routine maintenance, grid operators typically acquire technologies in one-off transactions, but the innovative nature of smart grid projects may require more collaborate relationships. This paper studies how a transactional versus relational approach to governing smart grid innovation projects affects incentives for other actors to collaborate. We analyse 34 cases of smart grid innovation projects based on extensive archival data as well as interviews. We find that projects relying on relational governance are more likely to provide incentives for collaboration. Especially non-financial incentives such as reputational benefits and shared intellectual property rights are more likely to be found in projects relying on relational governance. Policy makers that wish to stimulate smart grid innovation projects should consider stimulating long-term relationships between grid operators and third parties, because such relationships are more likely to produce incentives for collaboration. - Highlights: • Smart grids require collaboration between grid operators and other actors. • We contrast transactional and relational governance of smart grid projects. • Long-term relations produce more incentives for smart grid collaboration. • Non-financial incentives are more important in long-term relations. • Policy makers should stimulate long-term relations to stimulate smart grids.

  15. Grid simulator for power quality assessment of micro-grids

    DEFF Research Database (Denmark)

    Carrasco, Joaquin Eloy Garcia; Vasquez, Juan Carlos; Guerrero, Josep M.

    2013-01-01

    voltages, low-order harmonics and flicker. The aim of this equipment is to test the performance of a given system under such distorted voltages. A prototype of the simulator, consisting of two inverters connected back-to-back to a 380 V three-phase grid and feeding a micro-grid composed of two......-inverter interfaced distributed generators and a critical load was built and tested. A set of experimental results for linear purely resistive loads, non-linear loads and current-controlled inverters is presented to prove the capabilities of the simulator. Finally, a case study is presented by testing a micro-grid.......In this study, a grid simulator based on a back-to-back inverter topology with resonant controllers is presented. The simulator is able to generate three-phase voltages for a range of amplitudes and frequencies with different types of perturbations, such as voltage sags, steady-state unbalanced...

  16. Operation and Control of a Direct-Driven PMSG-Based Wind Turbine System with an Auxiliary Parallel Grid-Side Converter

    Directory of Open Access Journals (Sweden)

    Jiawei Chu

    2013-07-01

    Full Text Available In this paper, based on the similarity, in structure and principle, between a grid-connected converter for a direct-driven permanent magnet synchronous generator (D-PMSG and an active power filter (APF, a new D-PMSG-based wind turbine (WT system configuration that includes not only an auxiliary converter in parallel with the grid-side converter, but also a coordinated control strategy, is proposed to enhance the low voltage ride through (LVRT capability and improve power quality. During normal operation, the main grid-side converter maintains the DC-link voltage constant, whereas the auxiliary grid-side converter functions as an APF with harmonic suppression and reactive power compensation to improve the power quality. During grid faults, a hierarchical coordinated control scheme for the generator-side converter, main grid-side converter and auxiliary grid-side converter, depending on the grid voltage sags, is presented to enhance the LVRT capability of the direct-driven PMSG WT. The feasibility and the effectiveness of the proposed system’s topology and hierarchical coordinated control strategy were verified using MATLAB/Simulink simulations.

  17. Personality and Education Mining based Job Advisory System

    OpenAIRE

    Rajendra S. Choudhary; Rajul Kukreja; Nitika Jain; Shikha Jain

    2014-01-01

    Every job demands an employee with some specific qualities in addition to the basic educational qualification. For example, an introvert person cannot be a good leader despite of a very good academic qualification. Thinking and logical ability is required for a person to be a successful software engineer. So, the aim of this paper is to present a novel approach for advising an ideal job to the job seeker while considering his personality trait and educational qualification both. Very well-kno...

  18. Advanced fault ride-through control of DFIG based wind turbines including grid connection via VSC-HVDC

    Energy Technology Data Exchange (ETDEWEB)

    Feltes, Christian

    2012-07-01

    With the growing renewable energy share in the power generation mix it becomes inevitable that also these new generation technologies participate on the provision of grid services to guarantee stable operation of the grid, especially when one considers the decreasing number of conventional power plants in operation as a result of the expansion of wind based generation plants. These so-called ancillary services include frequency / active power control, voltage / reactive power control and fault ride-through (FRT) with fast voltage control and are stipulated in modern grid codes. In the context of this thesis advanced control algorithms have been developed for wind turbines based on doubly-fed induction generator (DFIG) to allow safe FRT during symmetrical and unsymmetrical faults. This covers the control for conventional AC grid connection as well as for the connection through voltage source converter (VSC) based high voltage direct current transmission (HVDC). Currently, the DFIG is the most used generator technology in modem wind turbines, since it combines a relatively simple slip-ring induction machine with a frequency converter rated to only approx. 30% of the total power. This makes the DFIG a cost-effective concept, which offers a variable speed range and a high degree of flexibility in control. However, due to the direct coupling of the generator stator circuit to the grid, grid faults are a special challenge for the frequency converter, its protection circuits and control algorithms. As base for the detailed evaluation of the impact of grid faults to the DFIG, this thesis contains the analytical derivation of the DFIG short circuit currents under consideration of frequency converter control. The DFIG concept presented in this thesis makes use of a DC chopper in the frequency converter, which allows safe FRT with grid voltage support through both converter sides. The developed control contains a new algorithm for a clear separation and control of positive

  19. New Generation GridPix: Development and characterisation of pixelated gaseous detectors based on the Timepix3 chip

    CERN Document Server

    AUTHOR|(CDS)2082958; Hessey, Nigel

    Due to the increasing demands of high energy physics experiments there is a need for particle detectors which enable high precision measurements. In this regard, the GridPix detector is a novel detector concept which combines the benefits of a pixel chip with an integrated gas amplification structure. The resulting unit is a detector sensitive to single electrons with a great potential for particle tracking and energy loss measurements. This thesis is focusing on the development of a new generation of GridPix detectors based on the Timepix3 chip, which implements a high resolution Time to Digital Converter (TDC) in each pixel. After an introductory chapter describing the motivation behind GridPix, the manuscript presents the physics of gaseous detectors in chapter 2 along with the gaseous detectors used for particle tracking in chapter 3. Chapters 4 and 5 are focusing on the tracking performance of GridPix detectors. Chapter 4 presents results obtained with a GridPix detector based on a small scale prototy...

  20. Job Creation and Job Types

    DEFF Research Database (Denmark)

    Kuhn, Johan M.; Malchow-Møller, Nikolaj; Sørensen, Anders

    We extend earlier analyses of the job creation of start-ups vs. established firms by taking into consideration the educational content of the jobs created and destroyed. We define educationspecific measures of job creation and job destruction at the firm level, and we use these to construct...... a measure of “surplus job creation” defined as jobs created on top of any simultaneous destruction of similar jobs in incumbent firms in the same region and industry. Using Danish employer-employee data from 2002-7, which identify the start-ups and which cover almost the entire private sector......, these measures allow us to provide a more nuanced assessment of the role of entrepreneurial firms in the job-creation process than previous studies. Our findings show that while start-ups are responsible for the entire overall net job creation, incumbents account for more than a third of net job creation within...

  1. QoS Differential Scheduling in Cognitive-Radio-Based Smart Grid Networks: An Adaptive Dynamic Programming Approach.

    Science.gov (United States)

    Yu, Rong; Zhong, Weifeng; Xie, Shengli; Zhang, Yan; Zhang, Yun

    2016-02-01

    As the next-generation power grid, smart grid will be integrated with a variety of novel communication technologies to support the explosive data traffic and the diverse requirements of quality of service (QoS). Cognitive radio (CR), which has the favorable ability to improve the spectrum utilization, provides an efficient and reliable solution for smart grid communications networks. In this paper, we study the QoS differential scheduling problem in the CR-based smart grid communications networks. The scheduler is responsible for managing the spectrum resources and arranging the data transmissions of smart grid users (SGUs). To guarantee the differential QoS, the SGUs are assigned to have different priorities according to their roles and their current situations in the smart grid. Based on the QoS-aware priority policy, the scheduler adjusts the channels allocation to minimize the transmission delay of SGUs. The entire transmission scheduling problem is formulated as a semi-Markov decision process and solved by the methodology of adaptive dynamic programming. A heuristic dynamic programming (HDP) architecture is established for the scheduling problem. By the online network training, the HDP can learn from the activities of primary users and SGUs, and adjust the scheduling decision to achieve the purpose of transmission delay minimization. Simulation results illustrate that the proposed priority policy ensures the low transmission delay of high priority SGUs. In addition, the emergency data transmission delay is also reduced to a significantly low level, guaranteeing the differential QoS in smart grid.

  2. ATLAS job transforms: a data driven workflow engine

    International Nuclear Information System (INIS)

    Stewart, G A; Breaden-Madden, W B; Maddocks, H J; Harenberg, T; Sandhoff, M; Sarrazin, B

    2014-01-01

    The need to run complex workflows for a high energy physics experiment such as ATLAS has always been present. However, as computing resources have become even more constrained, compared to the wealth of data generated by the LHC, the need to use resources efficiently and manage complex workflows within a single grid job have increased. In ATLAS, a new Job Transform framework has been developed that we describe in this paper. This framework manages the multiple execution steps needed to 'transform' one data type into another (e.g., RAW data to ESD to AOD to final ntuple) and also provides a consistent interface for the ATLAS production system. The new framework uses a data driven workflow definition which is both easy to manage and powerful. After a transform is defined, jobs are expressed simply by specifying the input data and the desired output data. The transform infrastructure then executes only the necessary substeps to produce the final data products. The global execution cost of running the job is minimised and the transform can adapt to scenarios where data can be produced along different execution paths. Transforms for specific physics tasks which support up to 60 individual substeps have been successfully run. As the new transforms infrastructure has been deployed in production many features have been added to the framework which improve reliability, quality of error reporting and also provide support for multi-process jobs.

  3. An Optimal Integrated Control Scheme for Permanent Magnet Synchronous Generator-Based Wind Turbines under Asymmetrical Grid Fault Conditions

    Directory of Open Access Journals (Sweden)

    Dan Wang

    2016-04-01

    Full Text Available In recent years, the increasing penetration level of wind energy into power systems has brought new issues and challenges. One of the main concerns is the issue of dynamic response capability during outer disturbance conditions, especially the fault-tolerance capability during asymmetrical faults. In order to improve the fault-tolerance and dynamic response capability under asymmetrical grid fault conditions, an optimal integrated control scheme for the grid-side voltage-source converter (VSC of direct-driven permanent magnet synchronous generator (PMSG-based wind turbine systems is proposed in this paper. The optimal control strategy includes a main controller and an additional controller. In the main controller, a double-loop controller based on differential flatness-based theory is designed for grid-side VSC. Two parts are involved in the design process of the flatness-based controller: the reference trajectories generation of flatness output and the implementation of the controller. In the additional control aspect, an auxiliary second harmonic compensation control loop based on an improved calculation method for grid-side instantaneous transmission power is designed by the quasi proportional resonant (Quasi-PR control principle, which is able to simultaneously restrain the second harmonic components in active power and reactive power injected into the grid without the respective calculation for current control references. Moreover, to reduce the DC-link overvoltage during grid faults, the mathematical model of DC-link voltage is analyzed and a feedforward modified control factor is added to the traditional DC voltage control loop in grid-side VSC. The effectiveness of the optimal control scheme is verified in PSCAD/EMTDC simulation software.

  4. Securing the smart grid information exchange

    Energy Technology Data Exchange (ETDEWEB)

    Fries, Steffen; Falk, Rainer [Siemens AG, Corporate Technology, Muenchen (Germany)

    2012-07-01

    The smart grid is based on information exchange between various stakeholders using open communication technologies, to control the physical electric grid through the information grid. Protection against cyber attacks is essential to ensure a reliable operation of the smart grid. This challenge is addressed by various regulatory, standardization, and research activities. After giving an overview of the security demand of a smart grid, existing and appearing standardization activities are described. (orig.)

  5. Developing a Grid-based search and categorization tool

    CERN Document Server

    Haya, Glenn; Vigen, Jens

    2003-01-01

    Grid technology has the potential to improve the accessibility of digital libraries. The participants in Project GRACE (Grid Search And Categorization Engine) are in the process of developing a search engine that will allow users to search through heterogeneous resources stored in geographically distributed digital collections. What differentiates this project from current search tools is that GRACE will be run on the European Data Grid, a large distributed network, and will not have a single centralized index as current web search engines do. In some cases, the distributed approach offers advantages over the centralized approach since it is more scalable, can be used on otherwise inaccessible material, and can provide advanced search options customized for each data source.

  6. Advances in Grid Computing for the Fabric for Frontier Experiments Project at Fermilab

    Science.gov (United States)

    Herner, K.; Alba Hernandez, A. F.; Bhat, S.; Box, D.; Boyd, J.; Di Benedetto, V.; Ding, P.; Dykstra, D.; Fattoruso, M.; Garzoglio, G.; Kirby, M.; Kreymer, A.; Levshina, T.; Mazzacane, A.; Mengel, M.; Mhashilkar, P.; Podstavkov, V.; Retzke, K.; Sharma, N.; Teheran, J.

    2017-10-01

    The Fabric for Frontier Experiments (FIFE) project is a major initiative within the Fermilab Scientific Computing Division charged with leading the computing model for Fermilab experiments. Work within the FIFE project creates close collaboration between experimenters and computing professionals to serve high-energy physics experiments of differing size, scope, and physics area. The FIFE project has worked to develop common tools for job submission, certificate management, software and reference data distribution through CVMFS repositories, robust data transfer, job monitoring, and databases for project tracking. Since the projects inception the experiments under the FIFE umbrella have significantly matured, and present an increasingly complex list of requirements to service providers. To meet these requirements, the FIFE project has been involved in transitioning the Fermilab General Purpose Grid cluster to support a partitionable slot model, expanding the resources available to experiments via the Open Science Grid, assisting with commissioning dedicated high-throughput computing resources for individual experiments, supporting the efforts of the HEP Cloud projects to provision a variety of back end resources, including public clouds and high performance computers, and developing rapid onboarding procedures for new experiments and collaborations. The larger demands also require enhanced job monitoring tools, which the project has developed using such tools as ElasticSearch and Grafana. in helping experiments manage their large-scale production workflows. This group in turn requires a structured service to facilitate smooth management of experiment requests, which FIFE provides in the form of the Production Operations Management Service (POMS). POMS is designed to track and manage requests from the FIFE experiments to run particular workflows, and support troubleshooting and triage in case of problems. Recently a new certificate management infrastructure called

  7. SoilGrids1km — Global Soil Information Based on Automated Mapping

    Science.gov (United States)

    Hengl, Tomislav; de Jesus, Jorge Mendes; MacMillan, Robert A.; Batjes, Niels H.; Heuvelink, Gerard B. M.; Ribeiro, Eloi; Samuel-Rosa, Alessandro; Kempen, Bas; Leenaars, Johan G. B.; Walsh, Markus G.; Gonzalez, Maria Ruiperez

    2014-01-01

    Background Soils are widely recognized as a non-renewable natural resource and as biophysical carbon sinks. As such, there is a growing requirement for global soil information. Although several global soil information systems already exist, these tend to suffer from inconsistencies and limited spatial detail. Methodology/Principal Findings We present SoilGrids1km — a global 3D soil information system at 1 km resolution — containing spatial predictions for a selection of soil properties (at six standard depths): soil organic carbon (g kg−1), soil pH, sand, silt and clay fractions (%), bulk density (kg m−3), cation-exchange capacity (cmol+/kg), coarse fragments (%), soil organic carbon stock (t ha−1), depth to bedrock (cm), World Reference Base soil groups, and USDA Soil Taxonomy suborders. Our predictions are based on global spatial prediction models which we fitted, per soil variable, using a compilation of major international soil profile databases (ca. 110,000 soil profiles), and a selection of ca. 75 global environmental covariates representing soil forming factors. Results of regression modeling indicate that the most useful covariates for modeling soils at the global scale are climatic and biomass indices (based on MODIS images), lithology, and taxonomic mapping units derived from conventional soil survey (Harmonized World Soil Database). Prediction accuracies assessed using 5–fold cross-validation were between 23–51%. Conclusions/Significance SoilGrids1km provide an initial set of examples of soil spatial data for input into global models at a resolution and consistency not previously available. Some of the main limitations of the current version of SoilGrids1km are: (1) weak relationships between soil properties/classes and explanatory variables due to scale mismatches, (2) difficulty to obtain covariates that capture soil forming factors, (3) low sampling density and spatial clustering of soil profile locations. However, as the SoilGrids

  8. Cobacabana (control of balance by card-based navigation) : A card-based system for job shop control

    NARCIS (Netherlands)

    Land, M.J.

    Existing card-based production control systems such as Kanban are mostly dedicated to repetitive production environments. Cards-based systems for job shop control are lacking, while particularly this industry segment shows a need for simple control systems. This paper aims at filling the gap by

  9. Passivity-Based Stability Assessment of Grid-Connected VSCs - An Overview

    DEFF Research Database (Denmark)

    Harnefors, Lennart; Wang, Xiongfei; Yepes, Alejandro G.

    2016-01-01

    The interconnection stability of a grid-connected voltage-source converter (VSC) can be assessed by the passivity properties of the VSC input admittance. If critical grid resonances fall within regions where the input admittance acts passively, i.e., has nonnegative real part, then their destabil......The interconnection stability of a grid-connected voltage-source converter (VSC) can be assessed by the passivity properties of the VSC input admittance. If critical grid resonances fall within regions where the input admittance acts passively, i.e., has nonnegative real part...

  10. Connection between Job Motivation, Job Satisfaction and Work Performance in Romanian Trade Enterprises

    Directory of Open Access Journals (Sweden)

    Dina Maria LUT

    2012-12-01

    Full Text Available On defining the connection between job motivation and job satisfaction and between job satisfaction and work performance is still not complete agreement among experts, given the complexity of the phenomena studied. Over time, the relationship between job motivation and job satisfaction has been addressed in many ways in the organizational psychology. The issue of the connection between job satisfaction and work performance is also at least as controversial as the issues mentioned above with respect to these phenomena. Related to this connection, in the literature existed, over time, at least three points of view which held that: job satisfaction leads to work performance; work performance leads to job satisfaction; rewards are involved between job satisfaction and work performance. The paper aim is to present some aspects related to the connection between job motivation, job satisfaction and work performance. The issues presented in this paper lead to the conclusion that in the work process, the connection between job satisfaction and work performance or the connection between job motivation and job satisfaction are not constant or linear connection. These connections should be nuanced interpreted, depending on the conditions of occurrence. These conclusions are based on the results of a survey conducted among enterprises dealing mostly in trade sector, in Romanian Western Region.

  11. Evaluation of a Positive Youth Development Program Based on the Repertory Grid Test

    Directory of Open Access Journals (Sweden)

    Daniel T. L. Shek

    2012-01-01

    Full Text Available The repertory grid test, based on personal construct psychology, was used to evaluate the effectiveness of Project P.A.T.H.S. (Positive Adolescent Training through Holistic Social Programmes in Hong Kong. One hundred and four program participants (n=104 were randomly invited to complete a repertory grid based on personal construct theory in order to provide both quantitative and qualitative data for measuring self-identity changes after joining the program. Findings generally showed that the participants perceived that they understood themselves better and had stronger resilience after joining the program. Participants also saw themselves as closer to their ideal selves and other positive role figures (but farther away from a loser after joining the program. This study provides additional support for the effectiveness of the Tier 1 Program of Project P.A.T.H.S. in the Chinese context. This study also shows that the repertory grid test is a useful evaluation method to measure self-identity changes in participants in positive youth development programs.

  12. PV-Powered CoMP-Based Green Cellular Networks with a Standby Grid Supply

    Directory of Open Access Journals (Sweden)

    Abu Jahid

    2017-01-01

    Full Text Available This paper proposes a novel framework for PV-powered cellular networks with a standby grid supply and an essential energy management technique for achieving envisaged green networks. The proposal considers an emerging cellular network architecture employing two types of coordinated multipoint (CoMP transmission techniques for serving the subscribers. Under the proposed framework, each base station (BS is powered by an individual PV solar energy module having an independent storage device. BSs are also connected to the conventional grid supply for meeting additional energy demand. We also propose a dynamic inter-BS solar energy sharing policy through a transmission line for further greening the proposed network by minimizing the consumption from the grid supply. An extensive simulation-based study in the downlink of a Long-Term Evolution (LTE cellular system is carried out for evaluating the energy efficiency performance of the proposed framework. System performance is also investigated for identifying the impact of various system parameters including storage factor, storage capacity, solar generation capacity, transmission line loss, and different CoMP techniques.

  13. Research on the Method of Urban Waterlogging Flood Routing Based on Hexagonal Grid

    Directory of Open Access Journals (Sweden)

    LAI Guangling

    2016-12-01

    Full Text Available An evolution of the urban waterlogging flood routing was studied in this paper based on the method of hexagonal grid modeling. Using the method of discrete grid, established an urban geometry model on account of the regular multi-scale discrete grid. With the fusion of 3D topographic survey data and 2D building vector data, formed a regular network model of surface. This model took multi special block into account, such as urban terrain and buildings. On this basis, a method of reverse flow deduction was proposed, which was an inverse computation from the state of flood to the evolution process. That is, based on the water depth of flood, made use of the connectivity with the outfall to calculate the range of water logging, and then implemented the urban waterlogging flood simulation deduction. The test indicated that, this method can implement the evolution of urban waterlogging scenario deduction effectively. And the correlational research could provide scientific basis for urban disaster prevention and emergency decision-making.

  14. Model Penilaian Risiko Kebakaran Perkotaan dengan Sistem Pakar berbasis GIS Grid-Based

    Directory of Open Access Journals (Sweden)

    Sabrillah Taridala

    2017-12-01

    Full Text Available Abstrak Kota Kendari merupakan suatu kawasan perkotaan dengan luas wilayah terkecil dan jumlah penduduk terpadat di Provinsi Sulawesi Tenggara. Bencana kebakaran di Kota Kendari sering terjadi dan telah menimbulkan kerugian yang cukup banyak, hingga menelan korban jiwa. Penelitian ini bertujuan untuk melakukan penilaian terhadap tingkat risiko bencana kebakaran di Kota Kendari dengan menggunakan pendekatan Sistem Pakar (Expert System berbasis Sistem Informasi Geografis (SIG. Hasil penelitian menujukkan bahwa tingkat risiko kebakaran di Kota Kendari terklasifikasi dalam empat kelas, yaitu tingkat risiko kebakaran sangat tinggi sebanyak 206 grid, tingkat risiko kebakaran tinggi sebanyak 6.815 grid, tingkat risiko kebakaran rendah sebanyak 46.175 grid, dan tingkat risiko kebakaran sangat rendah sebanyak 54.640 grid. Tingkat risiko kebakaran sangat tinggi di Kota Kendari merupakan kawasan terbangun yang berpenduduk padat dengan dominasi jenis material bangunan kayu dan campuran, terletak pada daerah dengan morfologi berbukit, dan aksesibilitas hanya dilalui oleh jalan umum yang memiliki lebar jalur lalu lintas <4 meter. Wilayah dengan tingkat risiko sangat rendah merupakan kawasan non-terbangun yang didominasi oleh badan air (sungai dan rawa, hutan dan sebagian kawasan pertanian (kebun. Kawasan tersebut bermorfologi datar, berbukit dan bergunung. Abstract Kendari city is an urban area with the smallest area and the densest population in Southeast Sulawesi Province. Fire disaster in the city of Kendari often occurs and has caused considerable losses, to claim casualties. This study aims to assess the risk degree of fire disaster in Kendari City using Expert System Approach based on Geographic Information System (GIS. The results showed that the degrees of fire risk in Kendari City were classified into four classes, ie very high fire risk degree, 206 grid, high fire risk degree, 6,815 grid, low fire risk degree, 46.175 grid, and very low fire risk, as

  15. Analysis of the Multi Strategy Goal Programming for Micro-Grid Based on Dynamic ant Genetic Algorithm

    Science.gov (United States)

    Qiu, J. P.; Niu, D. X.

    Micro-grid is one of the key technologies of the future energy supplies. Take economic planning. reliability, and environmental protection of micro grid as a basis for the analysis of multi-strategy objective programming problems for micro grid which contains wind power, solar power, and battery and micro gas turbine. Establish the mathematical model of each power generation characteristics and energy dissipation. and change micro grid planning multi-objective function under different operating strategies to a single objective model based on AHP method. Example analysis shows that in combination with dynamic ant mixed genetic algorithm can get the optimal power output of this model.

  16. WebGIS based on semantic grid model and web services

    Science.gov (United States)

    Zhang, WangFei; Yue, CaiRong; Gao, JianGuo

    2009-10-01

    ontology based on Grid technology and Web Services.

  17. PERFORMANCE BASED PAY AS A DETERMINANT OF JOB SATISFACTION: A STUDY IN MALAYSIA GIATMARA CENTERS

    OpenAIRE

    Azman ISMAIL; Nurhana M RAFIUDDIN; Mohd Hamran MOHAMAD; Norashikin Sahol HAMID; Aniza WAMIN; Nurzawani ZAKARIA

    2011-01-01

    Compensation management literature highlights that performance based pay has two major characteristics: participation in pay systems and adequacy of pay. The ability of management to properly implement such pay systems may lead to increased job satisfaction in organizations. Though, the nature of this relationship is interesting, little is known about the influence of performance based pay on job satisfaction in compensation management literature. Therefore, this study was conducted to examin...

  18. Grid-based lattice summation of electrostatic potentials by assembled rank-structured tensor approximation

    Science.gov (United States)

    Khoromskaia, Venera; Khoromskij, Boris N.

    2014-12-01

    Our recent method for low-rank tensor representation of sums of the arbitrarily positioned electrostatic potentials discretized on a 3D Cartesian grid reduces the 3D tensor summation to operations involving only 1D vectors however retaining the linear complexity scaling in the number of potentials. Here, we introduce and study a novel tensor approach for fast and accurate assembled summation of a large number of lattice-allocated potentials represented on 3D N × N × N grid with the computational requirements only weakly dependent on the number of summed potentials. It is based on the assembled low-rank canonical tensor representations of the collected potentials using pointwise sums of shifted canonical vectors representing the single generating function, say the Newton kernel. For a sum of electrostatic potentials over L × L × L lattice embedded in a box the required storage scales linearly in the 1D grid-size, O(N) , while the numerical cost is estimated by O(NL) . For periodic boundary conditions, the storage demand remains proportional to the 1D grid-size of a unit cell, n = N / L, while the numerical cost reduces to O(N) , that outperforms the FFT-based Ewald-type summation algorithms of complexity O(N3 log N) . The complexity in the grid parameter N can be reduced even to the logarithmic scale O(log N) by using data-sparse representation of canonical N-vectors via the quantics tensor approximation. For justification, we prove an upper bound on the quantics ranks for the canonical vectors in the overall lattice sum. The presented approach is beneficial in applications which require further functional calculus with the lattice potential, say, scalar product with a function, integration or differentiation, which can be performed easily in tensor arithmetics on large 3D grids with 1D cost. Numerical tests illustrate the performance of the tensor summation method and confirm the estimated bounds on the tensor ranks.

  19. Wireless Communications in Smart Grid

    Science.gov (United States)

    Bojkovic, Zoran; Bakmaz, Bojan

    Communication networks play a crucial role in smart grid, as the intelligence of this complex system is built based on information exchange across the power grid. Wireless communications and networking are among the most economical ways to build the essential part of the scalable communication infrastructure for smart grid. In particular, wireless networks will be deployed widely in the smart grid for automatic meter reading, remote system and customer site monitoring, as well as equipment fault diagnosing. With an increasing interest from both the academic and industrial communities, this chapter systematically investigates recent advances in wireless communication technology for the smart grid.

  20. Extending the Fermi-LAT Data Processing Pipeline to the Grid

    Science.gov (United States)

    Zimmer, S.; Arrabito, L.; Glanzman, T.; Johnson, T.; Lavalley, C.; Tsaregorodtsev, A.

    2012-12-01

    The Data Handling Pipeline (“Pipeline”) has been developed for the Fermi Gamma-Ray Space Telescope (Fermi) Large Area Telescope (LAT) which launched in June 2008. Since then it has been in use to completely automate the production of data quality monitoring quantities, reconstruction and routine analysis of all data received from the satellite and to deliver science products to the collaboration and the Fermi Science Support Center. Aside from the reconstruction of raw data from the satellite (Level 1), data reprocessing and various event-level analyses are also reasonably heavy loads on the pipeline and computing resources. These other loads, unlike Level 1, can run continuously for weeks or months at a time. In addition it receives heavy use in performing production Monte Carlo tasks. In daily use it receives a new data download every 3 hours and launches about 2000 jobs to process each download, typically completing the processing of the data before the next download arrives. The need for manual intervention has been reduced to less than 0.01% of submitted jobs. The Pipeline software is written almost entirely in Java and comprises several modules. The software comprises web-services that allow online monitoring and provides charts summarizing work flow aspects and performance information. The server supports communication with several batch systems such as LSF and BQS and recently also Sun Grid Engine and Condor. This is accomplished through dedicated job control services that for Fermi are running at SLAC and the other computing site involved in this large scale framework, the Lyon computing center of IN2P3. While being different in the logic of a task, we evaluate a separate interface to the Dirac system in order to communicate with EGI sites to utilize Grid resources, using dedicated Grid optimized systems rather than developing our own. More recently the Pipeline and its associated data catalog have been generalized for use by other experiments, and are

  1. Demands–abilities fit, work beliefs, meaningful work and engagement in nature-based jobs

    Directory of Open Access Journals (Sweden)

    Nellie de Crom

    2018-03-01

    Full Text Available Orientation: Meaningful work and personal engagement are important dimensions of flourishing of employees, especially when individuals work in challenging jobs. Research purpose: This study aimed to investigate the relationship between demands–abilities fit, work beliefs, meaningful work and engagement in individuals in nature-based jobs. Motivation for the study: Individuals working in nature often work under challenging circumstances without the necessary resources. A research gap exists regarding the effects of demands–abilities fit and work beliefs on meaningful work. It is also not clear how these antecedents and meaningful work will impact the engagement of individuals working in nature. Research approach, design and method: A cross-sectional survey was used with a convenience sample of 161 nature-based employees. Data were collected using a structured online questionnaire consisting of items from the demands–abilities fit scale, work–life questionnaire, work and meaning Inventory, work engagement scale and a biographical questionnaire. Main findings: Work beliefs (calling, career and job and demands–abilities fit predicted a large percentage of the variance in meaning making. Work beliefs (calling and job and demands–abilities fit also predicted a large percentage of the variance in greater good motivations. Demands–abilities fit and a calling work orientation indirectly affected work engagement via meaningful work. The scales which measured calling and job orientations showed insufficient discriminant validity in relation to the scales which measured positive meaning and work engagement. Practical and managerial implications: Managers should consider implementing interventions to affect the demands–abilities fit (through human resource management interventions and work beliefs of individuals working in nature (through job crafting. Promoting perceptions of meaningful work might contribute to higher personal engagement

  2. A Theorem on Grid Access Control

    Institute of Scientific and Technical Information of China (English)

    XU ZhiWei(徐志伟); BU GuanYing(卜冠英)

    2003-01-01

    The current grid security research is mainly focused on the authentication of grid systems. A problem to be solved by grid systems is to ensure consistent access control. This problem is complicated because the hosts in a grid computing environment usually span multiple autonomous administrative domains. This paper presents a grid access control model, based on asynchronous automata theory and the classic Bell-LaPadula model. This model is useful to formally study the confidentiality and integrity problems in a grid computing environment. A theorem is proved, which gives the necessary and sufficient conditions to a grid to maintain confidentiality.These conditions are the formalized descriptions of local (node) relations or relationship between grid subjects and node subjects.

  3. The MammoGrid Project Grids Architecture

    CERN Document Server

    McClatchey, Richard; Hauer, Tamas; Estrella, Florida; Saiz, Pablo; Rogulin, Dmitri; Buncic, Predrag; Clatchey, Richard Mc; Buncic, Predrag; Manset, David; Hauer, Tamas; Estrella, Florida; Saiz, Pablo; Rogulin, Dmitri

    2003-01-01

    The aim of the recently EU-funded MammoGrid project is, in the light of emerging Grid technology, to develop a European-wide database of mammograms that will be used to develop a set of important healthcare applications and investigate the potential of this Grid to support effective co-working between healthcare professionals throughout the EU. The MammoGrid consortium intends to use a Grid model to enable distributed computing that spans national borders. This Grid infrastructure will be used for deploying novel algorithms as software directly developed or enhanced within the project. Using the MammoGrid clinicians will be able to harness the use of massive amounts of medical image data to perform epidemiological studies, advanced image processing, radiographic education and ultimately, tele-diagnosis over communities of medical "virtual organisations". This is achieved through the use of Grid-compliant services [1] for managing (versions of) massively distributed files of mammograms, for handling the distri...

  4. Urban micro-grids

    International Nuclear Information System (INIS)

    Faure, Maeva; Salmon, Martin; El Fadili, Safae; Payen, Luc; Kerlero, Guillaume; Banner, Arnaud; Ehinger, Andreas; Illouz, Sebastien; Picot, Roland; Jolivet, Veronique; Michon Savarit, Jeanne; Strang, Karl Axel

    2017-02-01

    ENEA Consulting published the results of a study on urban micro-grids conducted in partnership with the Group ADP, the Group Caisse des Depots, ENEDIS, Omexom, Total and the Tuck Foundation. This study offers a vision of the definition of an urban micro-grid, the value brought by a micro-grid in different contexts based on real case studies, and the upcoming challenges that micro-grid stakeholders will face (regulation, business models, technology). The electric production and distribution system, as the backbone of an increasingly urbanized and energy dependent society, is urged to shift towards a more resilient, efficient and environment-friendly infrastructure. Decentralisation of electricity production into densely populated areas is a promising opportunity to achieve this transition. A micro-grid enhances local production through clustering electricity producers and consumers within a delimited electricity network; it has the ability to disconnect from the main grid for a limited period of time, offering an energy security service to its customers during grid outages for example. However: The islanding capability is an inherent feature of the micro-grid concept that leads to a significant premium on electricity cost, especially in a system highly reliant on intermittent electricity production. In this case, a smart grid, with local energy production and no islanding capability, can be customized to meet relevant sustainability and cost savings goals at lower costs For industrials, urban micro-grids can be economically profitable in presence of high share of reliable energy production and thermal energy demand micro-grids face strong regulatory challenges that should be overcome for further development Whether islanding is or is not implemented into the system, end-user demand for a greener, more local, cheaper and more reliable energy, as well as additional services to the grid, are strong drivers for local production and consumption. In some specific cases

  5. Benchmarking of Grid Fault Modes in Single-Phase Grid-Connected Photovoltaic Systems

    DEFF Research Database (Denmark)

    Yang, Yongheng; Blaabjerg, Frede; Zou, Zhixiang

    2013-01-01

    Pushed by the booming installations of singlephase photovoltaic (PV) systems, the grid demands regarding the integration of PV systems are expected to be modified. Hence, the future PV systems should become more active with functionalities of Low Voltage Ride-Through (LVRT) and grid support...... phase systems under grid faults. The intent of this paper is to present a benchmarking of grid fault modes that might come in future single-phase PV systems. In order to map future challenges, the relevant synchronization and control strategies are discussed. Some faulty modes are studied experimentally...... and provided at the end of this paper. It is concluded that there are extensive control possibilities in single-phase PV systems under grid faults. The Second Order General Integral based PLL technique might be the most promising candidate for future single-phase PV systems because of its fast adaptive...

  6. A Novel Framework Based on the Improved Job Demands-Resources (JD-R) Model to Understand the Impact of Job Characteristics on Job Burnout from the View of Emotion Regulation Theory.

    Science.gov (United States)

    Yang, Naiding; Lu, Jintao; Ye, Jinfu

    2018-03-01

    It has been suggested that individual job characteristics have a significant impact on job burnout, and the process is subject to the regulation of demographic variables. However, the influence path of job characteristics on job burnout is still a "black box". On the basis of a systematic literature review by employing Pub Med, Science Direct, Web of Science, Google Scholar, CNKI and Scopus for required information with the several keywords "Job burnout", "Emotion regulation", "Personality traits", and "Psychological stress", in this study, an improved mine rescue workers-oriented job demands-resources (JD-R) model was put forward. Then, a novel analysis framework, to explore the impact of job characteristics on job burnout from the view of emotion regulation theory, was proposed combining the personality trait theory. This study argues that job burnout is influenced by job demands through expressive suppression and by job resources through cognitive reappraisal respectively. Further more, job demands and job resources have the opposite effects on job burnout through the "loss-path" caused by job pressure and the "gain-path" arised from job motivation, respectively. Extrovert personality traits can affect the way the individual processes the information of work environment and then how individual further adopts emotion regulation strategies, finally resulting in indirectly affecting the influence path of mine rescue workers' job characteristics on job burnout. This present study can help managers to realize the importance of employees' psychological stress and job burnout problems. The obtained conclusions provide significant decision-making references for managers in intervening job burnout, managing emotional stress and mental health of employees.

  7. An Enhanced LVRT Scheme for DFIG-based WECSs under Both Balanced and Unbalanced Grid Voltage Sags

    DEFF Research Database (Denmark)

    Mohammadi, Jafar; Afsharnia, Saeed; Ebrahimzadeh, Esmaeil

    2017-01-01

    reactive power into the grid. The passive compensator is based on a three-phase stator damping resistor (SDR) located in series with the stator windings. The proposed scheme decreases the negative effects of grid voltage sags in the DFIG system including the rotor over-currents, electromagnetic torque...

  8. Job-Structure and Job-Related Information

    OpenAIRE

    川上, 善郎

    1981-01-01

    The requirements of job-related information in many domain, such as personnel selection, placement, training, personnel appraisal, job evaluation, job design etc, have developed many techniques of job analysis.In this paper, several approaches to analyze the job characteristics are reviewed; (a) conventional approach, (b) worker-oriented approach, and (c) perceived job characteristics approach.In addition, new direction of job-related information is discussed.

  9. The job self-efficacy and job involvement of clinical nursing teachers.

    Science.gov (United States)

    Yang, Hui-Ling; Kao, Yu-Hsiu; Huang, Yi-Ching

    2006-09-01

    This paper explored the present status of self-efficacy and job involvement of clinical nursing teachers and investigated the predictive power of teachers' personal background variables on such, as well as the relationship between self-efficacy and job involvement. A total of 419 participants in the survey sample were chosen among clinical nursing teachers at 19 public and private institutes of technology and junior colleges in Taiwan in 2004. The self-developed structural questionnaire was categorized into three sections, including personal background data, job self-efficacy related to the clinical teaching inventory and job involvement related to clinical teaching inventory. Of the total 419 questionnaires distributed for this cross-sectional survey, 266 valid copies were registered, at a recovery rate of 63%. Findings indicated that both the job self-efficacy and job involvement of clinical nursing teachers are at a medium to high level and that significant differences exist in job self-efficacy and job involvement based on differences in age, marital status, teaching seniority, teacher qualifications, and job satisfaction. Second, samples have significantly different performance in self-efficacy due to differences in education level attained and the medical institution to which nursing teachers had been assigned. Self-efficacy and job involvement are significantly positively correlated. These results can serve as a reference for the cultivation of nursing teachers and reform of clinical nursing education in the future.

  10. Coding Model and Mapping Method of Spherical Diamond Discrete Grids Based on Icosahedron

    Directory of Open Access Journals (Sweden)

    LIN Bingxian

    2016-12-01

    Full Text Available Discrete Global Grid(DGG provides a fundamental environment for global-scale spatial data's organization and management. DGG's encoding scheme, which blocks coordinate transformation between different coordination reference frames and reduces the complexity of spatial analysis, contributes a lot to the multi-scale expression and unified modeling of spatial data. Compared with other kinds of DGGs, Diamond Discrete Global Grid(DDGG based on icosahedron is beneficial to the spherical spatial data's integration and expression for much better geometric properties. However, its structure seems more complicated than DDGG on octahedron due to its initial diamond's edges cannot fit meridian and parallel. New challenges are posed when it comes to the construction of hierarchical encoding system and mapping relationship with geographic coordinates. On this issue, this paper presents a DDGG's coding system based on the Hilbert curve and designs conversion methods between codes and geographical coordinates. The study results indicate that this encoding system based on the Hilbert curve can express space scale and location information implicitly with the similarity between DDG and planar grid put into practice, and balances efficiency and accuracy of conversion between codes and geographical coordinates in order to support global massive spatial data's modeling, integrated management and all kinds of spatial analysis.

  11. Incentive-compatible demand-side management for smart grids based on review strategies

    Science.gov (United States)

    Xu, Jie; van der Schaar, Mihaela

    2015-12-01

    Demand-side load management is able to significantly improve the energy efficiency of smart grids. Since the electricity production cost depends on the aggregate energy usage of multiple consumers, an important incentive problem emerges: self-interested consumers want to increase their own utilities by consuming more than the socially optimal amount of energy during peak hours since the increased cost is shared among the entire set of consumers. To incentivize self-interested consumers to take the socially optimal scheduling actions, we design a new class of protocols based on review strategies. These strategies work as follows: first, a review stage takes place in which a statistical test is performed based on the daily prices of the previous billing cycle to determine whether or not the other consumers schedule their electricity loads in a socially optimal way. If the test fails, the consumers trigger a punishment phase in which, for a certain time, they adjust their energy scheduling in such a way that everybody in the consumer set is punished due to an increased price. Using a carefully designed protocol based on such review strategies, consumers then have incentives to take the socially optimal load scheduling to avoid entering this punishment phase. We rigorously characterize the impact of deploying protocols based on review strategies on the system's as well as the users' performance and determine the optimal design (optimal billing cycle, punishment length, etc.) for various smart grid deployment scenarios. Even though this paper considers a simplified smart grid model, our analysis provides important and useful insights for designing incentive-compatible demand-side management schemes based on aggregate energy usage information in a variety of practical scenarios.

  12. Calculation approaches for grid usage fees to influence the load curve in the distribution grid level

    International Nuclear Information System (INIS)

    Illing, Bjoern

    2014-01-01

    Dominated by the energy policy the decentralized German energy market is changing. One mature target of the government is to increase the contribution of renewable generation to the gross electricity consumption. In order to achieve this target disadvantages like an increased need for capacity management occurs. Load reduction and variable grid fees offer the grid operator solutions to realize capacity management by influencing the load profile. The evolution of the current grid fees towards more causality is required to adapt these approaches. Two calculation approaches are developed in this assignment. On the one hand multivariable grid fees keeping the current components demand and energy charge. Additional to the grid costs grid load dependent parameters like the amount of decentralized feed-ins, time and local circumstances as well as grid capacities are considered. On the other hand the grid fee flat-rate which represents a demand based model on a monthly level. Both approaches are designed to meet the criteria for future grid fees. By means of a case study the effects of the grid fees on the load profile at the low voltage grid is simulated. Thereby the consumption is represented by different behaviour models and the results are scaled at the benchmark grid area. The resulting load curve is analyzed concerning the effects of peak load reduction as well as the integration of renewable energy sources. Additionally the combined effect of grid fees and electricity tariffs is evaluated. Finally the work discusses the launching of grid fees in the tense atmosphere of politics, legislation and grid operation. Results of this work are two calculation approaches designed for grid operators to define the grid fees. Multivariable grid fees are based on the current calculation scheme. Hereby demand and energy charges are weighted by time, locational and load related dependencies. The grid fee flat-rate defines a limitation in demand extraction. Different demand levels

  13. Damping Methods for Resonances Caused by LCL-Filter-Based Current-Controlled Grid-Tied Power Inverters

    DEFF Research Database (Denmark)

    Wu, Weimin; Liu, Yuan; He, Yuanbin

    2017-01-01

    Grid-tied voltage source inverters using LCL filter have been widely adopted in distributed power generation systems (DPGSs). As high-order LCL filters contain multiple resonant frequencies, switching harmonics generated by the inverter and current harmonics generated by the active/passive loads...... innovative damping methods have been proposed. A comprehensive overview on those contributions and their classification on the inverter- and grid-side damping measures are presented. Based on the concept of the impedance-based stability analysis, all damping methods can ensure the system stability...

  14. Design and implementation of a web-based data grid management system for enterprise PACS backup and disaster recovery

    Science.gov (United States)

    Zhou, Zheng; Ma, Kevin; Talini, Elisa; Documet, Jorge; Lee, Jasper; Liu, Brent

    2007-03-01

    A cross-continental Data Grid infrastructure has been developed at the Image Processing and Informatics (IPI) research laboratory as a fault-tolerant image data backup and disaster recovery solution for Enterprise PACS. The Data Grid stores multiple copies of the imaging studies as well as the metadata, such as patient and study information, in geographically distributed computers and storage devices involving three different continents: America, Asia and Europe. This effectively prevents loss of image data and accelerates data recovery in the case of disaster. However, the lack of centralized management system makes the administration of the current Data Grid difficult. Three major challenges exist in current Data Grid management: 1. No single user interface to access and administrate each geographically separate component; 2. No graphical user interface available, resulting in command-line-based administration; 3. No single sign-on access to the Data Grid; administrators have to log into every Grid component with different corresponding user names/passwords. In this paper we are presenting a prototype of a unique web-based access interface for both Data Grid administrators and users. The interface has been designed to be user-friendly; it provides necessary instruments to constantly monitor the current status of the Data Grid components and their contents from any locations, contributing to longer system up-time.

  15. Design of a nonlinear backstepping control strategy of grid interconnected wind power system based PMSG

    Science.gov (United States)

    Errami, Y.; Obbadi, A.; Sahnoun, S.; Benhmida, M.; Ouassaid, M.; Maaroufi, M.

    2016-07-01

    This paper presents nonlinear backstepping control for Wind Power Generation System (WPGS) based Permanent Magnet Synchronous Generator (PMSG) and connected to utility grid. The block diagram of the WPGS with PMSG and the grid side back-to-back converter is established with the dq frame of axes. This control scheme emphasises the regulation of the dc-link voltage and the control of the power factor at changing wind speed. Besides, in the proposed control strategy of WPGS, Maximum Power Point Tracking (MPPT) technique and pitch control are provided. The stability of the regulators is assured by employing Lyapunov analysis. The proposed control strategy for the system has been validated by MATLAB simulations under varying wind velocity and the grid fault condition. In addition, a comparison of simulation results based on the proposed Backstepping strategy and conventional Vector Control is provided.

  16. Emotional intelligence, teamwork effectiveness, and job performance: the moderating role of job context.

    Science.gov (United States)

    Farh, Crystal I C Chien; Seo, Myeong-Gu; Tesluk, Paul E

    2012-07-01

    We advance understanding of the role of ability-based emotional intelligence (EI) and its subdimensions in the workplace by examining the mechanisms and context-based boundary conditions of the EI-performance relationship. Using a trait activation framework, we theorize that employees with higher overall EI and emotional perception ability exhibit higher teamwork effectiveness (and subsequent job performance) when working in job contexts characterized by high managerial work demands because such contexts contain salient emotion-based cues that activate employees' emotional capabilities. A sample of 212 professionals from various organizations and industries indicated support for the salutary effect of EI, above and beyond the influence of personality, cognitive ability, emotional labor job demands, job complexity, and demographic control variables. Theoretical and practical implications of the potential value of EI for workplace outcomes under contexts involving managerial complexity are discussed. (PsycINFO Database Record (c) 2012 APA, all rights reserved).

  17. Impedance Based Analysis and Design of Harmonic Resonant Controller for a Wide Range of Grid Impedance

    DEFF Research Database (Denmark)

    Kwon, Jun Bum; Wang, Xiongfei; Blaabjerg, Frede

    2014-01-01

    This paper investigates the effect of grid impedance variation on harmonic resonant current controllers for gridconnected voltage source converters by means of impedance-based analysis. It reveals that the negative harmonic resistances tend to be derived from harmonic resonant controllers...... in the closed-loop output admittance of converter. Such negative resistances may interact with the grid impedance resulting in steady state error or unstable harmonic compensation. To deal with this problem, a design guideline for harmonic resonant controllers under a wide range of grid impedance is proposed...

  18. Multi-Agent Model-Based Optimization for Future Electrical Grids

    NARCIS (Netherlands)

    Bajracharya, G.

    2014-01-01

    The electricity grid is one of the most complex systems created by human beings. It consists of an intricate network of components such as generators, transmission and distribution lines, transformers, breakers, various controllers, and various measurement and monitoring systems. The grid has been

  19. The Application of Stationary VOC-PR with PLL for Grid side Converter-based Wind Power Generation System

    DEFF Research Database (Denmark)

    Guo, Yougui; Zeng, Ping; Li, Lijuan

    2010-01-01

    Voltage oriented control PR is combined with space vector modulation and phase locked loop to control the grid side converter in wind power generation system in this paper. First the mathematical models of grid side converter and LCL filter as well as grid are given. Then the control strategy...... of grid side converter-based wind power generation system is given in detail. Finally the simulation model consisting of the grid side converter wind power generation system is set up. The simulation results have verified that the control strategy is feasible to be used for control of gird currents......, active power, reactive power and DC-link voltage in wind power generation system. It has laid a good basis for the real system development....

  20. A Visual Analysis Approach for Inferring Personal Job and Housing Locations Based on Public Bicycle Data

    Directory of Open Access Journals (Sweden)

    Xiaoying Shi

    2017-07-01

    Full Text Available Information concerning the home and workplace of residents is the basis of analyzing the urban job-housing spatial relationship. Traditional methods conduct time-consuming user surveys to obtain personal job and housing location information. Some new methods define rules to detect personal places based on human mobility data. However, because the travel patterns of residents are variable, simple rule-based methods are unable to generalize highly changing and complex travel modes. In this paper, we propose a visual analysis approach to assist the analyzer in inferring personal job and housing locations interactively based on public bicycle data. All users are first clustered to find potential commuting users. Then, several visual views are designed to find the key candidate stations for a specific user, and the visited temporal pattern of stations and the user’s hire behavior are analyzed, which helps with the inference of station semantic meanings. Finally, a number of users’ job and housing locations are detected by the analyzer and visualized. Our approach can manage the complex and diverse cycling habits of users. The effectiveness of the approach is shown through case studies based on a real-world public bicycle dataset.

  1. Job Attitudes of Agricultural Middle Managers

    OpenAIRE

    Bitsch, Vera

    2006-01-01

    The paper analyzes middle managers' job attitudes, in particular job satisfaction, based on case studies. Employees' job satisfaction is expected to reduce human resource management risks, leading to higher loyalty, organizational commitment and motivation and resulting in less turnover. Components of job satisfaction include achievement, recognition, work itself, job security, supervision, interpersonal relationships, compensation, organization, personal life and working conditions. They cau...

  2. Advances in Grid Computing for the FabrIc for Frontier Experiments Project at Fermialb

    Energy Technology Data Exchange (ETDEWEB)

    Herner, K. [Fermilab; Alba Hernandex, A. F. [Fermilab; Bhat, S. [Fermilab; Box, D. [Fermilab; Boyd, J. [Fermilab; Di Benedetto, V. [Fermilab; Ding, P. [Fermilab; Dykstra, D. [Fermilab; Fattoruso, M. [Fermilab; Garzoglio, G. [Fermilab; Kirby, M. [Fermilab; Kreymer, A. [Fermilab; Levshina, T. [Fermilab; Mazzacane, A. [Fermilab; Mengel, M. [Fermilab; Mhashilkar, P. [Fermilab; Podstavkov, V. [Fermilab; Retzke, K. [Fermilab; Sharma, N. [Fermilab; Teheran, J. [Fermilab

    2016-01-01

    The FabrIc for Frontier Experiments (FIFE) project is a major initiative within the Fermilab Scientic Computing Division charged with leading the computing model for Fermilab experiments. Work within the FIFE project creates close collaboration between experimenters and computing professionals to serve high-energy physics experiments of diering size, scope, and physics area. The FIFE project has worked to develop common tools for job submission, certicate management, software and reference data distribution through CVMFS repositories, robust data transfer, job monitoring, and databases for project tracking. Since the projects inception the experiments under the FIFE umbrella have signicantly matured, and present an increasingly complex list of requirements to service providers. To meet these requirements, the FIFE project has been involved in transitioning the Fermilab General Purpose Grid cluster to support a partitionable slot model, expanding the resources available to experiments via the Open Science Grid, assisting with commissioning dedicated high-throughput computing resources for individual experiments, supporting the eorts of the HEP Cloud projects to provision a variety of back end resources, including public clouds and high performance computers, and developing rapid onboarding procedures for new experiments and collaborations. The larger demands also require enhanced job monitoring tools, which the project has developed using such tools as ElasticSearch and Grafana. in helping experiments manage their large-scale production work ows. This group in turn requires a structured service to facilitate smooth management of experiment requests, which FIFE provides in the form of the Production Operations Management Service (POMS). POMS is designed to track and manage requests from the FIFE experiments to run particular work ows, and support troubleshooting and triage in case of problems. Recently a new certicate management infrastructure called Distributed

  3. Implementation of fuzzy-sliding mode based control of a grid connected photovoltaic system.

    Science.gov (United States)

    Menadi, Abdelkrim; Abdeddaim, Sabrina; Ghamri, Ahmed; Betka, Achour

    2015-09-01

    The present work describes an optimal operation of a small scale photovoltaic system connected to a micro-grid, based on both sliding mode and fuzzy logic control. Real time implementation is done through a dSPACE 1104 single board, controlling a boost chopper on the PV array side and a voltage source inverter (VSI) on the grid side. The sliding mode controller tracks permanently the maximum power of the PV array regardless of atmospheric condition variations, while The fuzzy logic controller (FLC) regulates the DC-link voltage, and ensures via current control of the VSI a quasi-total transit of the extracted PV power to the grid under a unity power factor operation. Simulation results, carried out via Matlab-Simulink package were approved through experiment, showing the effectiveness of the proposed control techniques. Copyright © 2015. Published by Elsevier Ltd.

  4. Coupling mechanism of electric vehicle and grid under the background of smart grid

    Science.gov (United States)

    Dong, Mingyu; Li, Dezhi; Chen, Rongjun; Shu, Han; He, Yongxiu

    2018-02-01

    With the development of smart distribution technology in the future, electric vehicle users can not only charge reasonably based on peak-valley price, they can also discharge electricity into the power grid to realize their economic benefit when it’s necessary and thus promote peak load shifting. According to the characteristic that future electric vehicles can discharge, this paper studies the interaction effect between electric vehicles and the grid based on TOU (time of use) Price Strategy. In this paper, four scenarios are used to compare the change of grid load after implementing TOU Price Strategy. The results show that the wide access of electric vehicles can effectively reduce peak and valley difference.

  5. Risk-Based Two-Stage Stochastic Optimization Problem of Micro-Grid Operation with Renewables and Incentive-Based Demand Response Programs

    Directory of Open Access Journals (Sweden)

    Pouria Sheikhahmadi

    2018-03-01

    Full Text Available The operation problem of a micro-grid (MG in grid-connected mode is an optimization one in which the main objective of the MG operator (MGO is to minimize the operation cost with optimal scheduling of resources and optimal trading energy with the main grid. The MGO can use incentive-based demand response programs (DRPs to pay an incentive to the consumers to change their demands in the peak hours. Moreover, the MGO forecasts the output power of renewable energy resources (RERs and models their uncertainties in its problem. In this paper, the operation problem of an MGO is modeled as a risk-based two-stage stochastic optimization problem. To model the uncertainties of RERs, two-stage stochastic programming is considered and conditional value at risk (CVaR index is used to manage the MGO’s risk-level. Moreover, the non-linear economic models of incentive-based DRPs are used by the MGO to change the peak load. The numerical studies are done to investigate the effect of incentive-based DRPs on the operation problem of the MGO. Moreover, to show the effect of the risk-averse parameter on MGO decisions, a sensitivity analysis is carried out.

  6. A novel approach for UI charge reduction using AMI based load prioritization in smart grid

    Directory of Open Access Journals (Sweden)

    Avani Pujara

    2017-09-01

    Full Text Available System frequency is vital part for power system balance. As per India Electricity Grid code frequency should be in the range of 49.5 Hz–50.5 Hz. Deviation from above mentioned range is charged as Unscheduled Interchange (UI charge. This paper proposes a new method for load and frequency control based on control of third parameter of three-part Availability Based Tariff (ABT i.e. Unscheduled Interchange charges. New circuit is designed considering prioritization of load and using Advanced Metering Infrastructure (AMI under Smart Grid environment.

  7. The effect of job organizational factors on job satisfaction in two automotive industries in Malaysia.

    Science.gov (United States)

    Dawal, Siti Zawiah Md; Taha, Zahari

    2007-12-01

    A methodology is developed in diagnosing the effect of job organizational factors on job satisfaction in two automotive industries in Malaysia. One hundred and seventy male subjects of age 18-40 years with the mean age of 26.8 and standard deviation (SD) of 5.3 years and the mean work experience of 6.5 years and SD of 4.9 years took part in the study. Five job organizational factors were tested in the study including job rotation, work method, training, problem solving and goal setting. A job organization questionnaire was designed and was based on respondents' perception in relation to job satisfaction. The results showed that job organization factors were significantly related to job satisfaction. Job rotation, work method, training and goal setting showed strong correlation with job satisfaction while problem solving had intermediate correlation in the first automotive industry. On the other hand, most job organization factors showed intermediate correlation with job satisfaction in the second automotive industry except the training factor which had low correlation with job satisfaction. These results highlight that job rotation, work methods, problem solving and goal setting are outstanding factors in the study of job satisfaction for automotive industries.

  8. Long range Debye-Hückel correction for computation of grid-based electrostatic forces between biomacromolecules

    International Nuclear Information System (INIS)

    Mereghetti, Paolo; Martinez, Michael; Wade, Rebecca C

    2014-01-01

    Brownian dynamics (BD) simulations can be used to study very large molecular systems, such as models of the intracellular environment, using atomic-detail structures. Such simulations require strategies to contain the computational costs, especially for the computation of interaction forces and energies. A common approach is to compute interaction forces between macromolecules by precomputing their interaction potentials on three-dimensional discretized grids. For long-range interactions, such as electrostatics, grid-based methods are subject to finite size errors. We describe here the implementation of a Debye-Hückel correction to the grid-based electrostatic potential used in the SDA BD simulation software that was applied to simulate solutions of bovine serum albumin and of hen egg white lysozyme. We found that the inclusion of the long-range electrostatic correction increased the accuracy of both the protein-protein interaction profiles and the protein diffusion coefficients at low ionic strength. An advantage of this method is the low additional computational cost required to treat long-range electrostatic interactions in large biomacromolecular systems. Moreover, the implementation described here for BD simulations of protein solutions can also be applied in implicit solvent molecular dynamics simulations that make use of gridded interaction potentials

  9. Do job demands and job control affect problem-solving?

    Science.gov (United States)

    Bergman, Peter N; Ahlberg, Gunnel; Johansson, Gun; Stoetzer, Ulrich; Aborg, Carl; Hallsten, Lennart; Lundberg, Ingvar

    2012-01-01

    The Job Demand Control model presents combinations of working conditions that may facilitate learning, the active learning hypothesis, or have detrimental effects on health, the strain hypothesis. To test the active learning hypothesis, this study analysed the effects of job demands and job control on general problem-solving strategies. A population-based sample of 4,636 individuals (55% women, 45% men) with the same job characteristics measured at two times with a three year time lag was used. Main effects of demands, skill discretion, task authority and control, and the combined effects of demands and control were analysed in logistic regressions, on four outcomes representing general problem-solving strategies. Those reporting high on skill discretion, task authority and control, as well as those reporting high demand/high control and low demand/high control job characteristics were more likely to state using problem solving strategies. Results suggest that working conditions including high levels of control may affect how individuals cope with problems and that workplace characteristics may affect behaviour in the non-work domain.

  10. [Subjective job strain and job satisfaction among neurologists in German hospitals].

    Science.gov (United States)

    Bauer, J; Bendels, M H K; Groneberg, D A

    2016-06-01

    The number of sick leaves due to job strain is increasing. This study's scope is to examine working conditions of neurologists in hospitals in regard to job strain and job satisfaction. This study is part of the iCEPT-Study. The iCEPT-Study was conducted as a web based survey among physicians (n = 7090) in German hospitals. The focus was on working conditions regarding job strain. Job strain was measured by a questionnaire consisting of items and scales from the short version of the Effort-Reward-Imbalance (ERI) questionnaire and the short questionnaire for working analysis (KFZA). By calculation ratios of distinct scales according to validated stress models a conclusion could be drawn as to whether or not job strain was present. The total number of n = 354 neurologists were analyzed. The response rate was at 18.2 %. Job strain was encountered by 52.0 % (95 %-KI: 46.7|57.2) of all neurologists and no significant gender difference was present. However, resident neurologists were significantly more often exposed to job strain than attending neurologists (OR = 2.9; 95 %-KI: 1.6-4.7; p job satisfaction, 59.6 % (95 %-KI: 54.5-64.7) of all respondents stated to be satisfied with their job. Significantly more men were satisfied than women (OR = 1.5; 95 %-KI: 1.0-2.4; p job than residents (OR = 2.9; 95 %-KI: 1.7-4.8; p job strain among neurologists in German hospitals. Keeping the negative implications of mental and physical health in mind, the working conditions of neurologists must be improved. As shown in this study, a possible way to do so is to increase job control in order to decrease a major stressor at work.

  11. Global Population Density Grid Time Series Estimates

    Data.gov (United States)

    National Aeronautics and Space Administration — Global Population Density Grid Time Series Estimates provide a back-cast time series of population density grids based on the year 2000 population grid from SEDAC's...

  12. Modeling and Control of VSC based DC Connection for Active Stall Wind Farms to Grid

    DEFF Research Database (Denmark)

    Iov, Florin; Sorensen, Poul; Hansen, Anca-Daniela

    2005-01-01

    Currently, there is an increasing trend to connect large MW wind farms to the transmission system. Therefore the power system becomes more vulnerable and dependent on the wind energy production. At the same time requirements that focus on the influence of the farms on the grid stability and power...... quality, and on the control capabilities of wind farms have already been established. The main trends of modem wind turbines/farms are clearly the variable speed operation and a grid connection through a power electronic interface, especially using doubly-fed induction generators. Using power electronics...... the control capabilities of these wind turbines/farms are extended and thus the grid requirements are fulfilled. However, the traditional squirrel-cage generators based wind turbines/wind farms directly connected to the grid does not have such control capabilities. They produce maximum possible power...

  13. Modelling and Control of VSC based DC Connection for Active Stall Wind Farms to Grid

    DEFF Research Database (Denmark)

    Iov, Florin; Sørensen, Poul; Hansen, Anca Daniela

    2006-01-01

    Currently, there is an increasing trend to connect large MW wind farms to the transmission system. Therefore the power system becomes more vulnerable and dependent on the wind energy production. At the same time requirements that focus on the influence of the farms on the grid stability and power...... quality, and on the control capabilities of wind farms have already been established. The main trends of modern wind turbines/farms are clearly the variable speed operation and a grid connection through a power electronic interface, especially using doubly-fed induction generators. Using power electronics...... the control capabilities of these wind turbines/farms are extended and thus the grid requirements are fulfilled.  However, the traditional squirrel-cage generators based wind turbines/wind farms directly connected to the grid does not have such control capabilities. They produce maximum possible power...

  14. Low-voltage ride-through of a droop-based three-phase four-wire grid-connected microgrid

    DEFF Research Database (Denmark)

    Sadeghkhani, Iman; Golshan, Mohamad Esmail Hamedani; Mehrizi-Sani, Ali

    2018-01-01

    system operations during abnormal grid conditions. The objective of this paper is to propose an LVRT scheme that improves the power quality of the entire microgrid. The developed method is implemented as the controller of the interface voltage-sourced converter (VSC) of a distributed energy resource...... control of each phase and does not require calculation of symmetrical components. Moreover, it can be employed in the VSC control systems with various reference frames and is effective for droop-based grid-connected microgrids with both single-phase and three-phase four-wire configurations. The proposed......The ability of riding through the grid disturbances can increase the integration of microgrids into the distribution system. Consequently, a grid-connected microgrid should provide ancillary services such as low voltage ride-through (LVRT) capability and reactive power support to sustain the power...

  15. Implementation of grid-connected to/from off-grid transference for micro-grid inverters

    OpenAIRE

    Heredero Peris, Daniel; Chillón Antón, Cristian; Pages Gimenez, Marc; Gross, Gabriel Igor; Montesinos Miracle, Daniel

    2013-01-01

    This paper presents the transfer of a microgrid converter from/to on-grid to/from off-grid when the converter is working in two different modes. In the first transfer presented method, the converter operates as a Current Source Inverter (CSI) when on-grid and as a Voltage Source Inverter (VSI) when off-grid. In the second transfer method, the converter is operated as a VSI both, when operated on-grid and off-grid. The two methods are implemented successfully in a real pla...

  16. A sparse grid based method for generative dimensionality reduction of high-dimensional data

    Science.gov (United States)

    Bohn, Bastian; Garcke, Jochen; Griebel, Michael

    2016-03-01

    Generative dimensionality reduction methods play an important role in machine learning applications because they construct an explicit mapping from a low-dimensional space to the high-dimensional data space. We discuss a general framework to describe generative dimensionality reduction methods, where the main focus lies on a regularized principal manifold learning variant. Since most generative dimensionality reduction algorithms exploit the representer theorem for reproducing kernel Hilbert spaces, their computational costs grow at least quadratically in the number n of data. Instead, we introduce a grid-based discretization approach which automatically scales just linearly in n. To circumvent the curse of dimensionality of full tensor product grids, we use the concept of sparse grids. Furthermore, in real-world applications, some embedding directions are usually more important than others and it is reasonable to refine the underlying discretization space only in these directions. To this end, we employ a dimension-adaptive algorithm which is based on the ANOVA (analysis of variance) decomposition of a function. In particular, the reconstruction error is used to measure the quality of an embedding. As an application, the study of large simulation data from an engineering application in the automotive industry (car crash simulation) is performed.

  17. Flexible DWDM Grid Manipulation Using Four Wave Mixing-based Time Lenses

    DEFF Research Database (Denmark)

    Røge, Kasper Meldgaard; Guan, Pengyu; Mulvad, Hans Christian Hansen

    2014-01-01

    An experimental demonstration of dense wavelength-division multiplexing (DWDM) grid manipulation is carried out using two time lenses. A DWDM spectrum is compressed from a 100-GHz to a 28-GHz grid with error-free performance.......An experimental demonstration of dense wavelength-division multiplexing (DWDM) grid manipulation is carried out using two time lenses. A DWDM spectrum is compressed from a 100-GHz to a 28-GHz grid with error-free performance....

  18. Job Creation and Job Types

    DEFF Research Database (Denmark)

    Kuhn, Johan Moritz; Malchow-Møller, Nikolaj; Sørensen, Anders

    2016-01-01

    We extend earlier analyses of the job creation of start-ups versus established firms by considering the educational content of the jobs created and destroyed. We define education-specific measures of job creation and job destruction at the firm level, and we use these measures to construct a meas...

  19. Job Creation and Job Types

    DEFF Research Database (Denmark)

    Kuhn, Johan M.; Malchow-Møller, Nikolaj; Sørensen, Anders

    We extend earlier analyses of the job creation of start-ups vs. established firms by taking into consideration the educational content of the jobs created and destroyed. We define educationspecific measures of job creation and job destruction at the firm level, and we use these to construct a mea...

  20. Integrating renewables in distribution grids. Storage, regulation and the interaction of different stakeholders in future grids

    Energy Technology Data Exchange (ETDEWEB)

    Nykamp, S.

    2013-10-18

    In recent years, the transition of the power supply chain towards a sustainable system based on 'green' electricity generation out of renewable energy sources (RES-E) has become a main challenge for grid operators and further stakeholders in the power system. To enable the evaluation of new concepts for the integration of RES-E, first the feed-in characteristics of photovoltaic, wind and biomass generators located in a distribution grid area and based on numerous measured feed-in data are studied in this thesis. The achieved insights from the feed-in profiles can be used for the dimensioning of grid assets. Furthermore, the results are useful for the evaluation of congestion management or for the dimensioning of storage assets in distribution grids. The latter aspect is analyzed in detail such that suitable storage characteristics for an introduction in the grid are determined. An economic approach is presented to derive break-even points for storage assets as a substitute to conventional reinforcements. For a case study from a real world low voltage grid with reinforcement needs, these break-even points are determined and the main influencing parameters are evaluated. A further important question in this context concerns the role DSOs (distribution system operators) may play with the operation of decentralized storage assets since several stakeholders may be interested in using the flexibility provided by these assets. This unclear responsibility also applies to the steering of adjustable consumption devices such as electric heat pumps or electric cars. For decentralized storage assets as well as heat pump appliances, optimal operation modes based on the optimization objectives for a DSO and a trader are derived. It is shown based on real world data that choosing a 'copperplate' scenario is not only technically insufficient for a global balance of the consumption and generation. It may even be harmful for the society from a welfare economic