WorldWideScience

Sample records for atlas production system

  1. ATLAS production system

    CERN Document Server

    Borodin, Mikhail; The ATLAS collaboration; De, Kaushik; Klimentov, Alexei; Golubkov, Dmitry; Maeno, Tadashi; Mashinistov, Ruslan; Wenaus, Torre; Padolski, Siarhei

    2016-01-01

    The second generation of the ATLAS production system called ProdSys2 is a distributed workload manager which used by thousands of physicists to analyze the data remotely, with the volume of processed data is beyond the exabyte scale, across a more than hundred heterogeneous sites. It achieves high utilization by combining dynamic job definition based on many criterias, such as input and output size, memory requirements and CPU consumption with manageable scheduling policies and by supporting different kind of computational resources, such as GRID, clouds, supercomputers and volunteering computers. Besides jobs definition Production System also includes flexible web user interface, which implements user-friendly environment for main ATLAS workflows, e.g. simple way of combining different data flows, and real-time monitoring, optimised for using with huge amount of information to present. We present an overview of the ATLAS Production System major components: job and task definition, workflow manager web user i...

  2. The Next Generation ATLAS Production System

    CERN Document Server

    Borodin, Mikhail; The ATLAS collaboration; Golubkov, Dmitry; Klimentov, Alexei; Maeno, Tadashi; Mashinistov, Ruslan; Vaniachine, Alexandre

    2015-01-01

    The ATLAS experiment at LHC data processing and simulation grows continuously, as more data and more use cases emerge. For data processing the ATLAS experiment adopted the data transformation approach, where software applications transform the input data into outputs. In the ATLAS production system, each data transformation is represented by a task, a collection of many jobs, dynamically submitted by the ATLAS workload management system (PanDA/JEDI) and executed on the Grid, clouds and supercomputers. Patterns in ATLAS data transformation workflows composed of many tasks provided a scalable production system framework for template definitions of the many-tasks workflows. User interface and system logic of these workflows are being implemented in the Database Engine for Tasks (DEFT). Such development required using modern computing technologies and approaches. We report technical details of this development: database implementation, server logic and Web user interface technologies.

  3. The ATLAS Production System Evolution

    CERN Document Server

    Borodin, Mikhail; The ATLAS collaboration

    2017-01-01

    The second generation of the ATLAS Production System called ProdSys2 is a distributed workload manager that runs daily hundreds of thousands of jobs, from dozens of different ATLAS-specific workflows, across more than a hundred heterogeneous sites. It achieves high utilization by combining dynamic job definition based upon many criteria, such as input and output size, memory requirements and CPU consumption, with manageable scheduling policies and by supporting different kinds of computational resources, such as GRID, clouds, supercomputers and volunteer computers. The system dynamically assigns a group of jobs (task) to a group of geographically distributed computing resources. Dynamic assignment and resource utilization is one of the major features of the system. The Production System has a sophisticated job fault recovery mechanism, which efficiently allows running multi-terabyte tasks without human intervention. We have implemented new features which allow automatic task submission and chaining of differe...

  4. Task Management in the New ATLAS Production System

    CERN Document Server

    De, K; The ATLAS collaboration; Klimentov, A; Potekhin, M; Vaniachine, A

    2013-01-01

    The ATLAS Production System is the top level workflow manager which translates physicists' needs for production level processing into actual workflows executed across about a hundred processing sites used globally by ATLAS. As the production workload increased in volume and complexity in recent years (the ATLAS production tasks count is above one million, with each task containing hundreds or thousands of jobs) there is a need to upgrade the Production System to meet the challenging requirements of the next LHC run while minimizing the operating costs. Providing a front-end and a management layer for petascale data processing and analysis, the new Production System contains generic subsystems that can be used in a wider range of applications. The main subsystems are the Database Engine for Tasks (DEFT) and the Job Execution and Definition Interface (JEDI). Based on users' requests, the DEFT subsystem manages inter-dependent groups of tasks (Meta-Tasks) and generates corresponding data processing workflows. Th...

  5. Task Management in the New ATLAS Production System

    CERN Document Server

    De, K; The ATLAS collaboration; Klimentov, A; Potekhin, M; Vaniachine, A

    2014-01-01

    The ATLAS Production System is the top level workflow manager which translates physicists' needs for production level processing into actual workflows executed across about a hundred processing sites used globally by ATLAS. As the production workload increased in volume and complexity in recent years (the ATLAS production tasks count is above one million, with each task containing hundreds or thousands of jobs) there is a need to upgrade the Production System to meet the challenging requirements of the next LHC run while minimizing the operating costs. Providing a front-end and a management layer for petascale data processing and analysis, the new Production System contains generic subsystems that can be used in a wider range of applications. The main subsystems are the Database Engine for Tasks (DEFT) and the Job Execution and Definition Interface (JEDI). Based on users' requests, the DEFT subsystem manages inter-dependent groups of tasks (Meta-Tasks) and generates corresponding data processing workflows. Th...

  6. Task management in the new ATLAS production system

    International Nuclear Information System (INIS)

    De, K; Golubkov, D; Klimentov, A; Potekhin, M; Vaniachine, A

    2014-01-01

    This document describes the design of the new Production System of the ATLAS experiment at the LHC [1]. The Production System is the top level workflow manager which translates physicists' needs for production level processing and analysis into actual workflows executed across over a hundred Grid sites used globally by ATLAS. As the production workload increased in volume and complexity in recent years (the ATLAS production tasks count is above one million, with each task containing hundreds or thousands of jobs) there is a need to upgrade the Production System to meet the challenging requirements of the next LHC run while minimizing the operating costs. In the new design, the main subsystems are the Database Engine for Tasks (DEFT) and the Job Execution and Definition Interface (JEDI). Based on users' requests, DEFT manages inter-dependent groups of tasks (Meta-Tasks) and generates corresponding data processing workflows. The JEDI component then dynamically translates the task definitions from DEFT into actual workload jobs executed in the PanDA Workload Management System [2]. We present the requirements, design parameters, basics of the object model and concrete solutions utilized in building the new Production System and its components.

  7. PanDA: distributed production and distributed analysis system for ATLAS

    International Nuclear Information System (INIS)

    Maeno, T

    2008-01-01

    A new distributed software system was developed in the fall of 2005 for the ATLAS experiment at the LHC. This system, called PANDA, provides an integrated service architecture with late binding of jobs, maximal automation through layered services, tight binding with ATLAS Distributed Data Management system [1], advanced error discovery and recovery procedures, and other features. In this talk, we will describe the PANDA software system. Special emphasis will be placed on the evolution of PANDA based on one and half year of real experience in carrying out Computer System Commissioning data production [2] for ATLAS. The architecture of PANDA is well suited for the computing needs of the ATLAS experiment, which is expected to be one of the first HEP experiments to operate at the petabyte scale

  8. Integration of Titan supercomputer at OLCF with ATLAS Production System

    CERN Document Server

    AUTHOR|(SzGeCERN)643806; The ATLAS collaboration; De, Kaushik; Klimentov, Alexei; Nilsson, Paul; Oleynik, Danila; Padolski, Siarhei; Panitkin, Sergey; Wenaus, Torre

    2017-01-01

    The PanDA (Production and Distributed Analysis) workload management system was developed to meet the scale and complexity of distributed computing for the ATLAS experiment. PanDA managed resources are distributed worldwide, on hundreds of computing sites, with thousands of physicists accessing hundreds of Petabytes of data and the rate of data processing already exceeds Exabyte per year. While PanDA currently uses more than 200,000 cores at well over 100 Grid sites, future LHC data taking runs will require more resources than Grid computing can possibly provide. Additional computing and storage resources are required. Therefore ATLAS is engaged in an ambitious program to expand the current computing model to include additional resources such as the opportunistic use of supercomputers. In this paper we will describe a project aimed at integration of ATLAS Production System with Titan supercomputer at Oak Ridge Leadership Computing Facility (OLCF). Current approach utilizes modified PanDA Pilot framework for jo...

  9. Integration of Titan supercomputer at OLCF with ATLAS production system

    CERN Document Server

    Panitkin, Sergey; The ATLAS collaboration

    2016-01-01

    The PanDA (Production and Distributed Analysis) workload management system was developed to meet the scale and complexity of distributed computing for the ATLAS experiment. PanDA managed resources are distributed worldwide, on hundreds of computing sites, with thousands of physicists accessing hundreds of Petabytes of data and the rate of data processing already exceeds Exabyte per year. While PanDA currently uses more than 200,000 cores at well over 100 Grid sites, future LHC data taking runs will require more resources than Grid computing can possibly provide. Additional computing and storage resources are required. Therefore ATLAS is engaged in an ambitious program to expand the current computing model to include additional resources such as the opportunistic use of supercomputers. In this talk we will describe a project aimed at integration of ATLAS Production System with Titan supercomputer at Oak Ridge Leadership Computing Facility (OLCF). Current approach utilizes modified PanDA Pilot framework for job...

  10. Scaling up ATLAS production system for the LHC Run 2 and beyond: project ProdSys2

    CERN Document Server

    Borodin, Mikhail; The ATLAS collaboration; García Navarro, José Enrique; Golubkov, Dmitry; Klimentov, Alexei; Maeno, Tadashi; Vaniachine, Alexandre

    2015-01-01

    The Big Data processing needs of the ATLAS experiment grow continuously, as more data and more use cases emerge. For Big Data processing the ATLAS experiment adopted the data transformation approach, where software applications transform the input data into outputs. In the ATLAS production system, each data transformation is represented by a task, a collection of many jobs, submitted by the ATLAS workload management system (PanDA) and executed on the Grid. Our experience shows that the rate of tasks submission grows exponentially over the years. To scale up the ATLAS production system for new challenges, we started the ProdSys2 project. PanDA has been upgraded with the Job Execution and Definition Interface (JEDI). Patterns in ATLAS data transformation workflows composed of many tasks, provided a scalable production system framework for template definitions of the many-tasks workflows. These workflows are being implemented in the Database Engine for Tasks (DEfT) that generates individual tasks for processing ...

  11. Integration of Titan supercomputer at OLCF with ATLAS Production System

    Science.gov (United States)

    Barreiro Megino, F.; De, K.; Jha, S.; Klimentov, A.; Maeno, T.; Nilsson, P.; Oleynik, D.; Padolski, S.; Panitkin, S.; Wells, J.; Wenaus, T.; ATLAS Collaboration

    2017-10-01

    The PanDA (Production and Distributed Analysis) workload management system was developed to meet the scale and complexity of distributed computing for the ATLAS experiment. PanDA managed resources are distributed worldwide, on hundreds of computing sites, with thousands of physicists accessing hundreds of Petabytes of data and the rate of data processing already exceeds Exabyte per year. While PanDA currently uses more than 200,000 cores at well over 100 Grid sites, future LHC data taking runs will require more resources than Grid computing can possibly provide. Additional computing and storage resources are required. Therefore ATLAS is engaged in an ambitious program to expand the current computing model to include additional resources such as the opportunistic use of supercomputers. In this paper we will describe a project aimed at integration of ATLAS Production System with Titan supercomputer at Oak Ridge Leadership Computing Facility (OLCF). Current approach utilizes modified PanDA Pilot framework for job submission to Titan’s batch queues and local data management, with lightweight MPI wrappers to run single node workloads in parallel on Titan’s multi-core worker nodes. It provides for running of standard ATLAS production jobs on unused resources (backfill) on Titan. The system already allowed ATLAS to collect on Titan millions of core-hours per month, execute hundreds of thousands jobs, while simultaneously improving Titans utilization efficiency. We will discuss the details of the implementation, current experience with running the system, as well as future plans aimed at improvements in scalability and efficiency. Notice: This manuscript has been authored, by employees of Brookhaven Science Associates, LLC under Contract No. DE-AC02-98CH10886 with the U.S. Department of Energy. The publisher by accepting the manuscript for publication acknowledges that the United States Government retains a non-exclusive, paid-up, irrevocable, world-wide license to

  12. Predictive analytics tools to adjust and monitor performance metrics for the ATLAS Production System

    CERN Document Server

    Titov, Mikhail; The ATLAS collaboration

    2017-01-01

    Every scientific workflow involves an organizational part which purpose is to plan an analysis process thoroughly according to defined schedule, thus to keep work progress efficient. Having such information as an estimation of the processing time or possibility of system outage (abnormal behaviour) will improve the planning process, provide an assistance to monitor system performance and predict its next state. The ATLAS Production System is an automated scheduling system that is responsible for central production of Monte-Carlo data, highly specialized production for physics groups, as well as data pre-processing and analysis using such facilities as grid infrastructures, clouds and supercomputers. With its next generation (ProdSys2) the processing rate is around 2M tasks per year that is more than 365M jobs per year. ProdSys2 evolves to accommodate a growing number of users and new requirements from the ATLAS Collaboration, physics groups and individual users. ATLAS Distributed Computing in its current stat...

  13. FEDERAL USERS CONFERENCE PRODUCT LINE TOOL SET (PLTS) MAP PRODUCTION SYSTEM (MPS) ATLAS CUSTOM GRIDS [Rev 0 was draft

    Energy Technology Data Exchange (ETDEWEB)

    HAYENGA, J.L.

    2006-12-19

    Maps, and more importantly Atlases, are assisting the user community in managing a large land area with complex issues, the most complex of which is the management of nuclear waste. The techniques and experiences discussed herein were gained while developing several atlases for use at the US Department of Energy's Hanford Site. The user community requires the ability to locate not only waste sites, but other features as well. Finding a specific waste site on a map and in the field is a difficult task at a site the size of Hanford. To find a specific waste site, the user begins by locating the item or object in an index, then locating the feature on the corresponding map within an atlas. Locating features requires a method for indexing them. The location index and how to place it on a map or atlas is the central theme presented in this article. The user requirements for atlases forced the design team to develop new and innovative solutions for requirements that Product Line Tool Set (PLTS) Map Production System (MPS)-Atlas was not designed to handle. The layout of the most complex atlases includes custom reference grids, multiple data frames, multiple map series, and up to 250 maps. All of these functional requirements are at the extreme edge of the capabilities of PLTS MPS-Atlas. This document outlines the setup of an atlas using PLTS MPS-Atlas to meet these requirements.

  14. FEDERAL USERS CONFERENCE PRODUCT LINE TOOL SET (PLTS) MAP PRODUCTION SYSTEM (MPS) ATLAS CUSTOM GRIDS [Rev 0 was draft

    International Nuclear Information System (INIS)

    HAYENGA, J.L.

    2006-01-01

    Maps, and more importantly Atlases, are assisting the user community in managing a large land area with complex issues, the most complex of which is the management of nuclear waste. The techniques and experiences discussed herein were gained while developing several atlases for use at the US Department of Energy's Hanford Site. The user community requires the ability to locate not only waste sites, but other features as well. Finding a specific waste site on a map and in the field is a difficult task at a site the size of Hanford. To find a specific waste site, the user begins by locating the item or object in an index, then locating the feature on the corresponding map within an atlas. Locating features requires a method for indexing them. The location index and how to place it on a map or atlas is the central theme presented in this article. The user requirements for atlases forced the design team to develop new and innovative solutions for requirements that Product Line Tool Set (PLTS) Map Production System (MPS)-Atlas was not designed to handle. The layout of the most complex atlases includes custom reference grids, multiple data frames, multiple map series, and up to 250 maps. All of these functional requirements are at the extreme edge of the capabilities of PLTS MPS-Atlas. This document outlines the setup of an atlas using PLTS MPS-Atlas to meet these requirements

  15. Grid production with the ATLAS Event Service

    CERN Document Server

    Benjamin, Douglas; The ATLAS collaboration

    2018-01-01

    ATLAS has developed and previously presented a new computing architecture, the Event Service, that allows real time delivery of fine grained workloads which process dispatched events (or event ranges) and immediately streams outputs. The principal aim was to profit from opportunistic resources such as commercial cloud, supercomputing, and volunteer computing, and otherwise unused cycles on clusters and grids. During the development and deployment phase, its utility also on the grid and conventional clusters for the exploitation of otherwise unused cycles became apparent. Here we describe our experience commissioning the Event Service on the grid in the ATLAS production system. We study the performance compared with standard simulation production. We describe the integration with the ATLAS data management system to ensure scalability and compatibility with object stores. Finally, we outline the remaining steps towards a fully commissioned system.

  16. Production Performance of the ATLAS Semiconductor Tracker Readout System

    CERN Document Server

    Mitsou, V A

    2006-01-01

    The ATLAS Semiconductor Tracker (SCT) together with the pixel and the transition radiation detectors will form the tracking system of the ATLAS experiment at LHC. It will consist of 20000 single-sided silicon microstrip sensors assembled back-to-back into modules mounted on four concentric barrels and two end-cap detectors formed by nine disks each. The SCT module production and testing has finished while the macro-assembly is well under way. After an overview of the layout and the operating environment of the SCT, a description of the readout electronics design and operation requirements will be given. The quality control procedure and the DAQ software for assuring the electrical functionality of hybrids and modules will be discussed. The focus will be on the electrical performance results obtained during the assembly and testing of the end-cap SCT modules.

  17. AGIS: The ATLAS Grid Information System

    CERN Document Server

    Anisenkov, Alexey; Di Girolamo, Alessandro; Gayazov, Stavro; Klimentov, Alexei; Oleynik, Danila; Senchenko, Alexander

    2012-01-01

    ATLAS is a particle physics experiment at the Large Hadron Collider at CERN. The experiment produces petabytes of data annually through simulation production and tens petabytes of data per year from the detector itself. The ATLAS Computing model embraces the Grid paradigm and a high degree of decentralization and computing resources able to meet ATLAS requirements of petabytes scale data operations. In this paper we present ATLAS Grid Information System (AGIS) designed to integrate configuration and status information about resources, services and topology of whole ATLAS Grid needed by ATLAS Distributed Computing applications and services.

  18. AGIS: The ATLAS Grid Information System

    OpenAIRE

    Anisenkov, Alexey; Belov, Sergey; Di Girolamo, Alessandro; Gayazov, Stavro; Klimentov, Alexei; Oleynik, Danila; Senchenko, Alexander

    2012-01-01

    ATLAS is a particle physics experiment at the Large Hadron Collider at CERN. The experiment produces petabytes of data annually through simulation production and tens petabytes of data per year from the detector itself. The ATLAS Computing model embraces the Grid paradigm and a high degree of decentralization and computing resources able to meet ATLAS requirements of petabytes scale data operations. In this paper we present ATLAS Grid Information System (AGIS) designed to integrate configurat...

  19. AGIS: The ATLAS Grid Information System

    Science.gov (United States)

    Anisenkov, A.; Di Girolamo, A.; Klimentov, A.; Oleynik, D.; Petrosyan, A.; Atlas Collaboration

    2014-06-01

    ATLAS, a particle physics experiment at the Large Hadron Collider at CERN, produced petabytes of data annually through simulation production and tens of petabytes of data per year from the detector itself. The ATLAS computing model embraces the Grid paradigm and a high degree of decentralization and computing resources able to meet ATLAS requirements of petabytes scale data operations. In this paper we describe the ATLAS Grid Information System (AGIS), designed to integrate configuration and status information about resources, services and topology of the computing infrastructure used by the ATLAS Distributed Computing applications and services.

  20. ATLAS Grid Data Processing: system evolution and scalability

    CERN Document Server

    Golubkov, D; The ATLAS collaboration; Klimentov, A; Minaenko, A; Nevski, P; Vaniachine, A; Walker, R

    2012-01-01

    The production system for Grid Data Processing handles petascale ATLAS data reprocessing and Monte Carlo activities. The production system empowered further data processing steps on the Grid performed by dozens of ATLAS physics groups with coordinated access to computing resources worldwide, including additional resources sponsored by regional facilities. The system provides knowledge management of configuration parameters for massive data processing tasks, reproducibility of results, scalable database access, orchestrated workflow and performance monitoring, dynamic workload sharing, automated fault tolerance and petascale data integrity control. The system evolves to accommodate a growing number of users and new requirements from our contacts in ATLAS main areas: Trigger, Physics, Data Preparation and Software & Computing. To assure scalability, the next generation production system architecture development is in progress. We report on scaling up the production system for a growing number of users provi...

  1. Pre-Production Validation of the ATLAS Level-1 Calorimeter Trigger System

    CERN Document Server

    Achenbach, R; Barnett, B M; Bauss, B; Belkin, A; Bohm, C; Brawn, I P; Davis, A O; Edwards, J; Eisenhandler, E F; Föhlisch, F; Gee, C N P; Geweniger, C; Gillman, A R; Hanke, P; Hellman, S; Hidvégi, A; Hillier, S J; Kluge, E E; Landon, M; Mahboubi, K; Mahout, G; Meier, K; Mirea, A; Moye, T H; Perera, V J O; Qian, W; Rieke, S; Rühr, F; Sankey, D P C; Schäfer, U; Schmitt, K; Schultz-Coulon, H C; Silverstein, S; Staley, R J; Tapprogge, S; Thomas, J P; Trefzger, T; Typaldos, D; Watkins, P M; Watson, A; Weber, G A; Weber, P; 14th IEEE - NPSS Real Time Conference 2005 Nuclear Plasma Sciences Society

    2005-01-01

    The Level-1 Calorimeter Trigger is a major part of the first stage of event selection for the ATLAS experiment at the LHC. It is a digital, pipelined system with several stages of processing, largely based on FPGAs, which perform programmable algorithms in parallel with a fixed latency to process about 300 Gbyte/s of input data. The real-time output consists of counts of different types of trigger objects and energy sums. Prototypes of all module types have been undergoing intensive testing before final production during 2005. Verification of their correct operation has been performed standalone and in the ATLAS test-beam at CERN. Results from these investigations will be presented, along with a description of the methodology used to perform the tests.

  2. Atlas event production on the EGEE infrastructure

    CERN Document Server

    Espinal, X; Perini, L; Rod, W

    2007-01-01

    ATLAS is one of the four LHC (Large Hadron Collider) experiments at CERN, is devoted to study proton-proton and ion-ion collisions at 14TeV. ATLAS collaboration is composed of about 2000 scientists spread around the world. The activity of the experiment requirements for next year is of about 300TB of storage and a CPU power of about 13 Mski2sk, and is relying on GRID philosophy and EGEE infrastructure. Simulated events are distributed over EGEE by the Atlas production system. Data has to be processed and must be accessible by a huge number of scientists for analysis. The throughput of data for Atlas experiment is expected to be of 320 MB/s with an integrated amount of data per year of ~10Pb. The processing and storage need a distributed share of resources, spread worldwide and interconnected with GRID technologies as the requirements are so demanding for the LHC. In that sense event production is the way to produce, process and store data for analysis before the experiment startup, and is performed in a distr...

  3. Rotating target wheel system for super-heavy element production at ATLAS

    CERN Document Server

    Greene, J P; Falout, J; Janssens, R V F

    2004-01-01

    A new scattering chamber housing a large diameter rotating target wheel has been designed and constructed in front of the Fragment Mass Analyzer (FMA) for the production of very heavy nuclei (Z greater than 100) using beams from the Argonne Tandem Linear Accelerator System (ATLAS). In addition to the target and drive system, the chamber is extensively instrumented in order to monitor target performance and deterioration. Capabilities also exist to install rotating entrance and exit windows for gas cooling of the target within the scattering chamber. The design and initial tests are described.

  4. The ATLAS detector control system

    International Nuclear Information System (INIS)

    Schlenker, S.; Arfaoui, S.; Franz, S.

    2012-01-01

    The ATLAS experiment is one of the multi-purpose experiments at the Large Hadron Collider (LHC), constructed to study elementary particle interactions in collisions of high-energy proton beams. Twelve different sub-detectors as well as the common experimental infrastructure are supervised by the Detector Control System (DCS). The DCS enables equipment supervision of all ATLAS sub-detectors by using a system of more that 130 server machines running the industrial SCADA product PVSS. This highly distributed system reads, processes and archives of the order of 10 6 operational parameters. Higher level control system layers allow for automatic control procedures, efficient error recognition and handling, and manage the communication with external systems such as the LHC. First, this contribution describes the status of the ATLAS DCS and the experience gained during the LHC commissioning and the first physics data taking operation period. Secondly, the future evolution and maintenance constraints for the coming years and the LHC high luminosity upgrades are outlined. (authors)

  5. The ATLAS Detector Control System

    CERN Document Server

    Schlenker, S; Kersten, S; Hirschbuehl, D; Braun, H; Poblaguev, A; Oliveira Damazio, D; Talyshev, A; Zimmermann, S; Franz, S; Gutzwiller, O; Hartert, J; Mindur, B; Tsarouchas, CA; Caforio, D; Sbarra, C; Olszowska, J; Hajduk, Z; Banas, E; Wynne, B; Robichaud-Veronneau, A; Nemecek, S; Thompson, PD; Mandic, I; Deliyergiyev, M; Polini, A; Kovalenko, S; Khomutnikov, V; Filimonov, V; Bindi, M; Stanecka, E; Martin, T; Lantzsch, K; Hoffmann, D; Huber, J; Mountricha, E; Santos, HF; Ribeiro, G; Barillari, T; Habring, J; Arabidze, G; Boterenbrood, H; Hart, R; Marques Vinagre, F; Lafarguette, P; Tartarelli, GF; Nagai, K; D'Auria, S; Chekulaev, S; Phillips, P; Ertel, E; Brenner, R; Leontsinis, S; Mitrevski, J; Grassi, V; Karakostas, K; Iakovidis, G.; Marchese, F; Aielli, G

    2011-01-01

    The ATLAS experiment is one of the multi-purpose experiments at the Large Hadron Collider (LHC), constructed to study elementary particle interactions in collisions of high-energy proton beams. Twelve different sub-detectors as well as the common experimental infrastructure are supervised by the Detector Control System (DCS). The DCS enables equipment supervision of all ATLAS sub-detectors by using a system of >130 server machines running the industrial SCADA product PVSS. This highly distributed system reads, processes and archives of the order of 106 operational parameters. Higher level control system layers allow for automatic control procedures, efficient error recognition and handling, and manage the communication with external systems such as the LHC. This contribution firstly describes the status of the ATLAS DCS and the experience gained during the LHC commissioning and the first physics data taking operation period. Secondly, the future evolution and maintenance constraints for the coming years an...

  6. Evolution of the ATLAS PanDA Production and Distributed Analysis System

    International Nuclear Information System (INIS)

    Maeno, T; Wenaus, T; Fine, V; Potekhin, M; Panitkin, S; De, K; Nilsson, P; Stradling, A; Walker, R; Compostella, G

    2012-01-01

    The PanDA (Production and Distributed Analysis) system has been developed to meet ATLAS production and analysis requirements for a data-driven workload management system capable of operating at LHC data processing scale. PanDA has performed well with high reliability and robustness during the two years of LHC data-taking, while being actively evolved to meet the rapidly changing requirements for analysis use cases. We will present an overview of system evolution including automatic rebrokerage and reattempt for analysis jobs, adaptation for the CernVM File System, support for the multi-cloud model through which Tier-2 sites act as members of multiple clouds, pledged resource management and preferential brokerage, and monitoring improvements. We will also describe results from the analysis of two years of PanDA usage statistics, current issues, and plans for the future.

  7. A dynamic system for ATLAS software installation on OSG grid sites

    International Nuclear Information System (INIS)

    Zhao, X; Maeno, T; Wenaus, T; Leuhring, F; Youssef, S; Brunelle, J; De Salvo, A; Thompson, A S

    2010-01-01

    A dynamic and reliable system for installing the ATLAS software releases on Grid sites is crucial to guarantee the timely and smooth start of ATLAS production and reduce its failure rate. In this paper, we discuss the issues encountered in the previous software installation system, and introduce the new approach, which is built upon the new development in the areas of the ATLAS workload management system (PanDA), and software package management system (pacman). It is also designed to integrate with the EGEE ATLAS software installation framework. In the new system, ATLAS software releases are packaged as pacball, a uniquely identifiable and reproducible self-installing data file. The distribution of pacballs to remote sites is managed by ATLAS data management system (DQ2) and PanDA server. The installation on remote sites is automatically triggered by the PanDA pilot jobs. The installation job payload connects to a central ATLAS software installation portal, making the information of installation status easily accessible across OSG and EGEE Grids. The issues encountered in running the new system in production, and our future plan for improvement, will also be discussed.

  8. The ATLAS Production System Evolution: New Data Processing and Analysis Paradigm for the LHC Run2 and High-Luminosity

    Science.gov (United States)

    Barreiro, F. H.; Borodin, M.; De, K.; Golubkov, D.; Klimentov, A.; Maeno, T.; Mashinistov, R.; Padolski, S.; Wenaus, T.; ATLAS Collaboration

    2017-10-01

    The second generation of the ATLAS Production System called ProdSys2 is a distributed workload manager that runs daily hundreds of thousands of jobs, from dozens of different ATLAS specific workflows, across more than hundred heterogeneous sites. It achieves high utilization by combining dynamic job definition based on many criteria, such as input and output size, memory requirements and CPU consumption, with manageable scheduling policies and by supporting different kind of computational resources, such as GRID, clouds, supercomputers and volunteer-computers. The system dynamically assigns a group of jobs (task) to a group of geographically distributed computing resources. Dynamic assignment and resources utilization is one of the major features of the system, it didn’t exist in the earliest versions of the production system where Grid resources topology was predefined using national or/and geographical pattern. Production System has a sophisticated job fault-recovery mechanism, which efficiently allows to run multi-Terabyte tasks without human intervention. We have implemented “train” model and open-ended production which allow to submit tasks automatically as soon as new set of data is available and to chain physics groups data processing and analysis with central production by the experiment. We present an overview of the ATLAS Production System and its major components features and architecture: task definition, web user interface and monitoring. We describe the important design decisions and lessons learned from an operational experience during the first year of LHC Run2. We also report the performance of the designed system and how various workflows, such as data (re)processing, Monte-Carlo and physics group production, users analysis, are scheduled and executed within one production system on heterogeneous computing resources.

  9. Quarkonia production in small and large systems measured by ATLAS

    CERN Document Server

    Lopez, Jorge; The ATLAS collaboration

    2018-01-01

    The experimentally observed dissociation and regeneration of bound quarkonium states in heavy-ion collisions provide a powerful tool to probe the dynamics of the hot, dense plasma. These measurements are sensitive to the effects of color screening, color recombination, or other, new suppression mechanisms. In the large-statistics Run 2 lead-lead and proton-lead collision data, these phenomena can be probed with unprecedented precision. Measurements of the ground and excited quarkonia states, as well as their separation into prompt and non-prompt components, provide further opportunities to study the dynamics of heavy parton energy loss in these large systems. In addition, quarkonium production rates, and their excited to ground states ratios, in small, asymmetric systems are an interesting probe of cold nuclear matter effects. In this talk, the latest ATLAS results on quarkonia production will be presented, including new, differential measurements of charmonium suppression and azimuthal modulation in lead-lea...

  10. Multilevel Workflow System in the ATLAS Experiment

    International Nuclear Information System (INIS)

    Borodin, M; De, K; Navarro, J Garcia; Golubkov, D; Klimentov, A; Maeno, T; Vaniachine, A

    2015-01-01

    The ATLAS experiment is scaling up Big Data processing for the next LHC run using a multilevel workflow system comprised of many layers. In Big Data processing ATLAS deals with datasets, not individual files. Similarly a task (comprised of many jobs) has become a unit of the ATLAS workflow in distributed computing, with about 0.8M tasks processed per year. In order to manage the diversity of LHC physics (exceeding 35K physics samples per year), the individual data processing tasks are organized into workflows. For example, the Monte Carlo workflow is composed of many steps: generate or configure hard-processes, hadronize signal and minimum-bias (pileup) events, simulate energy deposition in the ATLAS detector, digitize electronics response, simulate triggers, reconstruct data, convert the reconstructed data into ROOT ntuples for physics analysis, etc. Outputs are merged and/or filtered as necessary to optimize the chain. The bi-level workflow manager - ProdSys2 - generates actual workflow tasks and their jobs are executed across more than a hundred distributed computing sites by PanDA - the ATLAS job-level workload management system. On the outer level, the Database Engine for Tasks (DEfT) empowers production managers with templated workflow definitions. On the next level, the Job Execution and Definition Interface (JEDI) is integrated with PanDA to provide dynamic job definition tailored to the sites capabilities. We report on scaling up the production system to accommodate a growing number of requirements from main ATLAS areas: Trigger, Physics and Data Preparation. (paper)

  11. The ATLAS Production System Evolution: New Data Processing and Analysis Paradigm for the LHC Run2 and High-Luminosity

    CERN Document Server

    AUTHOR|(SzGeCERN)677929; The ATLAS collaboration; Barreiro Megino, Fernando Harald; De, Kaushik; Golubkov, Dmitry; Klimentov, Alexei; Maeno, Tadashi; Mashinistov, Ruslan; Padolski, Siarhei; Wenaus, Torre

    2017-01-01

    The second generation of the ATLAS Production System called ProdSys2 is a distributed workload manager that runs daily hundreds of thousands of jobs, from dozens of different ATLAS specific workflows, across more than hundred heterogeneous sites. It achieves high utilization by combining dynamic job definition based on many criteria, such as input and output size, memory requirements and CPU consumption, with manageable scheduling policies and by supporting different kind of computational resources, such as GRID, clouds, supercomputers and volunteering computers. The system dynamically assigns a group of jobs (task) to a group of geographically distributed computing resources. Dynamic assignment and resources utilization is one of the major features of the system, it didn’t exist in the earliest versions of the production system where Grid resources topology has been predefined using national or/and geographical pattern. Production System has a sophisticated job fault-recovery mechanism, which efficiently a...

  12. Multiple brain atlas database and atlas-based neuroimaging system.

    Science.gov (United States)

    Nowinski, W L; Fang, A; Nguyen, B T; Raphel, J K; Jagannathan, L; Raghavan, R; Bryan, R N; Miller, G A

    1997-01-01

    For the purpose of developing multiple, complementary, fully labeled electronic brain atlases and an atlas-based neuroimaging system for analysis, quantification, and real-time manipulation of cerebral structures in two and three dimensions, we have digitized, enhanced, segmented, and labeled the following print brain atlases: Co-Planar Stereotaxic Atlas of the Human Brain by Talairach and Tournoux, Atlas for Stereotaxy of the Human Brain by Schaltenbrand and Wahren, Referentially Oriented Cerebral MRI Anatomy by Talairach and Tournoux, and Atlas of the Cerebral Sulci by Ono, Kubik, and Abernathey. Three-dimensional extensions of these atlases have been developed as well. All two- and three-dimensional atlases are mutually preregistered and may be interactively registered with an actual patient's data. An atlas-based neuroimaging system has been developed that provides support for reformatting, registration, visualization, navigation, image processing, and quantification of clinical data. The anatomical index contains about 1,000 structures and over 400 sulcal patterns. Several new applications of the brain atlas database also have been developed, supported by various technologies such as virtual reality, the Internet, and electronic publishing. Fusion of information from multiple atlases assists the user in comprehensively understanding brain structures and identifying and quantifying anatomical regions in clinical data. The multiple brain atlas database and atlas-based neuroimaging system have substantial potential impact in stereotactic neurosurgery and radiotherapy by assisting in visualization and real-time manipulation in three dimensions of anatomical structures, in quantitative neuroradiology by allowing interactive analysis of clinical data, in three-dimensional neuroeducation, and in brain function studies.

  13. The ATLAS Production System Evolution. New Data Processing and Analysis Paradigm for the LHC Run2 and High-Luminosity

    CERN Document Server

    Borodin, Mikhail; The ATLAS collaboration

    2016-01-01

    The second generation of the ATLAS production system called ProdSys2 is a distributed workload manager that runs daily hundreds of thousands of jobs, from dozens of different ATLAS specific workflows, across a more than hundred heterogeneous sites. It achieves high utilization by combining dynamic job definition based on many criterias, such as input and output size, memory requirements and CPU consumption with manageable scheduling policies and by supporting different kind of computational resources, such as GRID, clouds, supercomputers and volunteering computers. The system dynamically assigns a group of jobs (task) to a group of geographically distributed computing resources. Dynamic assignment and resources utilization is one of the major system’s features, it didn’t exist in the earliest versions of the production system, where Grid resources topology has been predefined using national or/and geographical pattern. Production System has a sophisticated job fault-recovery mechanism, which allows effici...

  14. Overview of the ATLAS distributed computing system

    CERN Document Server

    Elmsheuser, Johannes; The ATLAS collaboration

    2018-01-01

    The CERN ATLAS experiment successfully uses a worldwide computing infrastructure to support the physics program during LHC Run 2. The grid workflow system PanDA routinely manages 250 to 500 thousand concurrently running production and analysis jobs to process simulation and detector data. In total more than 300 PB of data is distributed over more than 150 sites in the WLCG and handled by the ATLAS data management system Rucio. To prepare for the ever growing LHC luminosity in future runs new developments are underway to even more efficiently use opportunistic resources such as HPCs and utilize new technologies. This presentation will review and explain the outline and the performance of the ATLAS distributed computing system and give an outlook to new workflow and data management ideas for the beginning of the LHC Run 3.

  15. The ATLAS Detector Control System

    International Nuclear Information System (INIS)

    Lantzsch, K; Braun, H; Hirschbuehl, D; Kersten, S; Arfaoui, S; Franz, S; Gutzwiller, O; Schlenker, S; Tsarouchas, C A; Mindur, B; Hartert, J; Zimmermann, S; Talyshev, A; Oliveira Damazio, D; Poblaguev, A; Martin, T; Thompson, P D; Caforio, D; Sbarra, C; Hoffmann, D

    2012-01-01

    The ATLAS experiment is one of the multi-purpose experiments at the Large Hadron Collider (LHC) at CERN, constructed to study elementary particle interactions in collisions of high-energy proton beams. Twelve different sub detectors as well as the common experimental infrastructure are controlled and monitored by the Detector Control System (DCS) using a highly distributed system of 140 server machines running the industrial SCADA product PVSS. Higher level control system layers allow for automatic control procedures, efficient error recognition and handling, manage the communication with external systems such as the LHC controls, and provide a synchronization mechanism with the ATLAS data acquisition system. Different databases are used to store the online parameters of the experiment, replicate a subset used for physics reconstruction, and store the configuration parameters of the systems. This contribution describes the computing architecture and software tools to handle this complex and highly interconnected control system.

  16. The ATLAS Detector Control System

    Science.gov (United States)

    Lantzsch, K.; Arfaoui, S.; Franz, S.; Gutzwiller, O.; Schlenker, S.; Tsarouchas, C. A.; Mindur, B.; Hartert, J.; Zimmermann, S.; Talyshev, A.; Oliveira Damazio, D.; Poblaguev, A.; Braun, H.; Hirschbuehl, D.; Kersten, S.; Martin, T.; Thompson, P. D.; Caforio, D.; Sbarra, C.; Hoffmann, D.; Nemecek, S.; Robichaud-Veronneau, A.; Wynne, B.; Banas, E.; Hajduk, Z.; Olszowska, J.; Stanecka, E.; Bindi, M.; Polini, A.; Deliyergiyev, M.; Mandic, I.; Ertel, E.; Marques Vinagre, F.; Ribeiro, G.; Santos, H. F.; Barillari, T.; Habring, J.; Huber, J.; Arabidze, G.; Boterenbrood, H.; Hart, R.; Iakovidis, G.; Karakostas, K.; Leontsinis, S.; Mountricha, E.; Ntekas, K.; Filimonov, V.; Khomutnikov, V.; Kovalenko, S.; Grassi, V.; Mitrevski, J.; Phillips, P.; Chekulaev, S.; D'Auria, S.; Nagai, K.; Tartarelli, G. F.; Aielli, G.; Marchese, F.; Lafarguette, P.; Brenner, R.

    2012-12-01

    The ATLAS experiment is one of the multi-purpose experiments at the Large Hadron Collider (LHC) at CERN, constructed to study elementary particle interactions in collisions of high-energy proton beams. Twelve different sub detectors as well as the common experimental infrastructure are controlled and monitored by the Detector Control System (DCS) using a highly distributed system of 140 server machines running the industrial SCADA product PVSS. Higher level control system layers allow for automatic control procedures, efficient error recognition and handling, manage the communication with external systems such as the LHC controls, and provide a synchronization mechanism with the ATLAS data acquisition system. Different databases are used to store the online parameters of the experiment, replicate a subset used for physics reconstruction, and store the configuration parameters of the systems. This contribution describes the computing architecture and software tools to handle this complex and highly interconnected control system.

  17. Evolution of the ATLAS Nightly Build System

    International Nuclear Information System (INIS)

    Undrus, A

    2012-01-01

    The ATLAS Nightly Build System is a major component in the ATLAS collaborative software organization, validation, and code approval scheme. For over 10 years of development it has evolved into a factory for automatic release production and grid distribution. The 50 multi-platform branches of ATLAS releases provide vast opportunities for testing new packages, verification of patches to existing software, and migration to new platforms and compilers for ATLAS code that currently contains 2200 packages with 4 million C++ and 1.4 million python scripting lines written by about 1000 developers. Recent development was focused on the integration of ATLAS Nightly Build and Installation systems. The nightly releases are distributed and validated and some are transformed into stable releases used for data processing worldwide. The ATLAS Nightly System is managed by the NICOS control tool on a computing farm with 50 powerful multiprocessor nodes. NICOS provides the fully automated framework for the release builds, testing, and creation of distribution kits. The ATN testing framework of the Nightly System runs unit and integration tests in parallel suites, fully utilizing the resources of multi-core machines, and provides the first results even before compilations complete. The NICOS error detection system is based on several techniques and classifies the compilation and test errors according to their severity. It is periodically tuned to place greater emphasis on certain software defects by highlighting the problems on NICOS web pages and sending automatic e-mail notifications to responsible developers. These and other recent developments will be presented and future plans will be described.

  18. Production of the Finnish Wind Atlas

    DEFF Research Database (Denmark)

    Tammelin, Bengt; Vihma, Timo; Atlaskin, Evgeny

    2013-01-01

    ) the parameterization method for gust factor was extended to be applicable at higher altitudes; and (vii) the dissemination of the Wind Atlas was based on new technical solutions. The AROME results were calculated for the heights of 50, 75, 100, 125, 150, 200, 300 and 400 m, and the WAsP results for the heights of 50......, 75, 100, 125 and 150 m. In addition to the wind speed, the results included the values of the Weibull distribution parameters, the gust factor, wind power content and the potential power production, which was calculated for three turbine sizes. The Wind Atlas data are available for each grid point......The Finnish Wind Atlas was prepared applying the mesoscale model AROME with 2.5 km horizontal resolution and the diagnostic downscaling method Wind Atlas Analysis and Application Programme (WAsP) with 250 m resolution. The latter was applied for areas most favourable for wind power production: a 30...

  19. ATLAS MPGD production status

    CERN Document Server

    Schioppa, Marco; The ATLAS collaboration

    2018-01-01

    Micromegas (MICRO MEsh GAseous Structure) chambers are Micro-Pattern Gaseous Detectors designed to provide a high spatial resolution and reasonable good time resolution in highly irradiated environments. In 2007 an ambitious long-term R\\&D activity was started in the context of the ATLAS experiment, at CERN: the Muon ATLAS Micromegas Activity (MAMMA). After years of tests on prototypes and technology breakthroughs, Micromegas chambers were chosen as tracking detectors for an upgrade of the ATLAS Muon Spectrometer. These novel detectors will be installed in 2020 at the end of the second long shutdown of the Large Hadron Collider, and will serve mainly as precision detectors in the innermost part of the forward ATLAS Muon Spectrometer. Four different types of Micromegas modules, eight layers each, up to $3 m^2$ area (of unprecedented size), will cover a surface of $150 m^2$ for a total active area of about $1200 m^2$. With this upgrade the ATLAS muon system will maintain the full acceptance of its excellent...

  20. Experience with the gLite workload management system in ATLAS Monte Carlo production on LCG

    International Nuclear Information System (INIS)

    Campana, S; Sciaba, A; Rebatto, D

    2008-01-01

    The ATLAS experiment has been running continuous simulated events production since more than two years. A considerable fraction of the jobs is daily submitted and handled via the gLite Workload Management System, which overcomes several limitations of the previous LCG Resource Broker. The gLite WMS has been tested very intensively for the LHC experiments use cases for more than six months, both in terms of performance and reliability. The tests were carried out by the LCG Experiment Integration Support team (in close contact with the experiments) together with the EGEE integration and certification team and the gLite middleware developers. A pragmatic iterative and interactive approach allowed a very quick rollout of fixes and their rapid deployment, together with new functionalities, for the ATLAS production activities. The same approach is being adopted for other middleware components like the gLite and CREAM Computing Elements. In this contribution we will summarize the learning from the gLite WMS testing activity, pointing out the most important achievements and the open issues. In addition, we will present the current situation of the ATLAS simulated event production activity on the EGEE infrastructure based on the gLite WMS, showing the main improvements and benefits from the new middleware. Finally, the gLite WMS is being used by many other VOs, including the LHC experiments. In particular, some statistics will be shown on the CMS experience running WMS user analysis via the WMS

  1. b-Production measurements with ATLAS

    International Nuclear Information System (INIS)

    Robins, Simon

    2001-01-01

    The Large Hadron Collider at CERN will produce b-hadrons with unprecedentedly high statistics, and with a higher energy reach than any previous machine. The ATLAS detector will offer excellent acceptance and performance for b physics studies and will make high statistics measurements of b-quark production to probe QCD. ATLAS can contribute to an understanding of the large discrepancy seen between the QCD prediction and Tevatron cross-section measurements. Measurements will be made of the absolute and differential production cross-sections. Measurements of bb-bar correlations will be made to test the NLO contributions to the cross-section and to place limits on non-perturbative effects

  2. Experiences with the new ATLAS Distributed Data Management System

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00214543; The ATLAS collaboration

    2016-01-01

    The ATLAS Distributed Data Management (DDM) system has evolved drastically in the last two years with the Rucio software fully replacing the previous system before the start of LHC Run-2. The ATLAS DDM system manages now more than 200 petabytes spread on 130 storage sites and can handle file transfer rates of up to 30Hz. In this talk, we discuss our experience acquired in developing, commissioning, running and maintaining such a large system. First, we describe the general architecture of the system, our integration with external services like the WLCG File Transfer Service and the evolution of the system over its first year of production. Then, we show the performance of the system, describe the integration of new technologies such as object stores, and outline future developments which mainly focus on performance and automation. Finally we discuss the long term evolution of ATLAS data management.

  3. AGIS: Evolution of Distributed Computing information system for ATLAS

    Science.gov (United States)

    Anisenkov, A.; Di Girolamo, A.; Alandes, M.; Karavakis, E.

    2015-12-01

    ATLAS, a particle physics experiment at the Large Hadron Collider at CERN, produces petabytes of data annually through simulation production and tens of petabytes of data per year from the detector itself. The ATLAS computing model embraces the Grid paradigm and a high degree of decentralization of computing resources in order to meet the ATLAS requirements of petabytes scale data operations. It has been evolved after the first period of LHC data taking (Run-1) in order to cope with new challenges of the upcoming Run- 2. In this paper we describe the evolution and recent developments of the ATLAS Grid Information System (AGIS), developed in order to integrate configuration and status information about resources, services and topology of the computing infrastructure used by the ATLAS Distributed Computing applications and services.

  4. The ATLAS PanDA Monitoring System and its Evolution

    CERN Document Server

    Klimentov, A; The ATLAS collaboration; Potekhin, M; Wenaus, T

    2011-01-01

    The PanDA (Production and Distributed Analysis) Workload Management System is used for ATLAS distributed production and analysis worldwide. The needs of ATLAS global computing imposed challenging requirements on PanDA design in areas such as scalability, robustness, automation, diagnostics, and usability for both production shifters and analysis users. Important to meeting these and other requirements is a comprehensive monitoring system. Through a system-wide job database, the PanDA monitor provides a comprehensive and coherent view of the system and job execution, from high level summaries to detailed drill-down job diagnostics. It is (like the rest of PanDA) an Apache-based Python application backed by Oracle. The presentation layer is HTML code generated on the fly in the Python application which is also responsible for managing database queries. However, this approach is lacking in user interface flexibility, simplicity of communication with external systems, and ease of maintenance. We decided to migrat...

  5. The ATLAS PanDA Monitoring System and its Evolution

    CERN Document Server

    Klimentov, A; The ATLAS collaboration; Potekhin, M; Wenaus, T

    2010-01-01

    The PanDA (Production and Distributed Analysis) Workload Management System is used for ATLAS distributed production and analysis worldwide. The needs of ATLAS global computing imposed challenging requirements on PanDA design in areas such as scalability, robustness, automation, diagnostics, and usability for both production shifters and analysis users. Important to meeting these and other requirements is a comprehensive monitoring system. Through a system-wide job database, the PanDA monitor provides a comprehensive and coherent view of the system and job execution, from high level summaries to detailed drill-down job diagnostics. It is (like the rest of PanDA) an Apache-based Python application backed by Oracle. The presentation layer is HTML code generated on the fly in the Python application which is also responsible for managing database queries. However, this approach is lacking in user interface flexibility, simplicity of communication with external systems, and ease of maintenance. We decided to migrat...

  6. Glance Information System for ATLAS Management

    International Nuclear Information System (INIS)

    Grael, F F; Maidantchik, C; Évora, L H R A; Karam, K; Moraes, L O F; Cirilli, M; Nessi, M; Pommès, K

    2011-01-01

    ATLAS Experiment is an international collaboration where more than 37 countries, 172 institutes and laboratories, 2900 physicists, engineers, and computer scientists plus 700 students participate. The management of this teamwork involves several aspects such as institute contribution, employment records, members' appointment, authors' list, preparation and publication of papers and speakers nomination. Previously, most of the information was accessible by a limited group and developers had to face problems such as different terminology, diverse data modeling, heterogeneous databases and unlike users needs. Moreover, the systems were not designed to handle new requirements. The maintenance has to be an easy task due to the long lifetime experiment and professionals turnover. The Glance system, a generic mechanism for accessing any database, acts as an intermediate layer isolating the user from the particularities of each database. It retrieves, inserts and updates the database independently of its technology and modeling. Relying on Glance, a group of systems were built to support the ATLAS management and operation aspects: ATLAS Membership, ATLAS Appointments, ATLAS Speakers, ATLAS Analysis Follow-Up, ATLAS Conference Notes, ATLAS Thesis, ATLAS Traceability and DSS Alarms Viewer. This paper presents the overview of the Glance information framework and describes the privilege mechanism developed to grant different level of access for each member and system.

  7. Glance Information System for ATLAS Management

    Science.gov (United States)

    Grael, F. F.; Maidantchik, C.; Évora, L. H. R. A.; Karam, K.; Moraes, L. O. F.; Cirilli, M.; Nessi, M.; Pommès, K.; ATLAS Collaboration

    2011-12-01

    ATLAS Experiment is an international collaboration where more than 37 countries, 172 institutes and laboratories, 2900 physicists, engineers, and computer scientists plus 700 students participate. The management of this teamwork involves several aspects such as institute contribution, employment records, members' appointment, authors' list, preparation and publication of papers and speakers nomination. Previously, most of the information was accessible by a limited group and developers had to face problems such as different terminology, diverse data modeling, heterogeneous databases and unlike users needs. Moreover, the systems were not designed to handle new requirements. The maintenance has to be an easy task due to the long lifetime experiment and professionals turnover. The Glance system, a generic mechanism for accessing any database, acts as an intermediate layer isolating the user from the particularities of each database. It retrieves, inserts and updates the database independently of its technology and modeling. Relying on Glance, a group of systems were built to support the ATLAS management and operation aspects: ATLAS Membership, ATLAS Appointments, ATLAS Speakers, ATLAS Analysis Follow-Up, ATLAS Conference Notes, ATLAS Thesis, ATLAS Traceability and DSS Alarms Viewer. This paper presents the overview of the Glance information framework and describes the privilege mechanism developed to grant different level of access for each member and system.

  8. Evolution of the ATLAS distributed computing system during the LHC long shutdown

    Science.gov (United States)

    Campana, S.; Atlas Collaboration

    2014-06-01

    The ATLAS Distributed Computing project (ADC) was established in 2007 to develop and operate a framework, following the ATLAS computing model, to enable data storage, processing and bookkeeping on top of the Worldwide LHC Computing Grid (WLCG) distributed infrastructure. ADC development has always been driven by operations and this contributed to its success. The system has fulfilled the demanding requirements of ATLAS, daily consolidating worldwide up to 1 PB of data and running more than 1.5 million payloads distributed globally, supporting almost one thousand concurrent distributed analysis users. Comprehensive automation and monitoring minimized the operational manpower required. The flexibility of the system to adjust to operational needs has been important to the success of the ATLAS physics program. The LHC shutdown in 2013-2015 affords an opportunity to improve the system in light of operational experience and scale it to cope with the demanding requirements of 2015 and beyond, most notably a much higher trigger rate and event pileup. We will describe the evolution of the ADC software foreseen during this period. This includes consolidating the existing Production and Distributed Analysis framework (PanDA) and ATLAS Grid Information System (AGIS), together with the development and commissioning of next generation systems for distributed data management (DDM/Rucio) and production (Prodsys-2). We will explain how new technologies such as Cloud Computing and NoSQL databases, which ATLAS investigated as R&D projects in past years, will be integrated in production. Finally, we will describe more fundamental developments such as breaking job-to-data locality by exploiting storage federations and caches, and event level (rather than file or dataset level) workload engines.

  9. Evolution of the ATLAS distributed computing system during the LHC long shutdown

    International Nuclear Information System (INIS)

    Campana, S

    2014-01-01

    The ATLAS Distributed Computing project (ADC) was established in 2007 to develop and operate a framework, following the ATLAS computing model, to enable data storage, processing and bookkeeping on top of the Worldwide LHC Computing Grid (WLCG) distributed infrastructure. ADC development has always been driven by operations and this contributed to its success. The system has fulfilled the demanding requirements of ATLAS, daily consolidating worldwide up to 1 PB of data and running more than 1.5 million payloads distributed globally, supporting almost one thousand concurrent distributed analysis users. Comprehensive automation and monitoring minimized the operational manpower required. The flexibility of the system to adjust to operational needs has been important to the success of the ATLAS physics program. The LHC shutdown in 2013-2015 affords an opportunity to improve the system in light of operational experience and scale it to cope with the demanding requirements of 2015 and beyond, most notably a much higher trigger rate and event pileup. We will describe the evolution of the ADC software foreseen during this period. This includes consolidating the existing Production and Distributed Analysis framework (PanDA) and ATLAS Grid Information System (AGIS), together with the development and commissioning of next generation systems for distributed data management (DDM/Rucio) and production (Prodsys-2). We will explain how new technologies such as Cloud Computing and NoSQL databases, which ATLAS investigated as R and D projects in past years, will be integrated in production. Finally, we will describe more fundamental developments such as breaking job-to-data locality by exploiting storage federations and caches, and event level (rather than file or dataset level) workload engines.

  10. arcControlTower: the System for Atlas Production and Analysis on ARC

    International Nuclear Information System (INIS)

    Filipcic, Andrej

    2011-01-01

    PanDA, the Atlas management and distribution system for production and analysis jobs on EGEE and OSG clusters, is based on pilot jobs to increase the throughput and stability of the job execution on grid. The ARC middleware uses a specific approach which tightly connects the job requirements with cluster capabilities like resource usage, software availability and caching of input files. The pilot concept renders the ARC features useless. The arcControlTower is the job submission system which merges the pilot benefits and ARC advantages. It takes the pilot payload from the panda server and submits the jobs to the Nordugrid ARC clusters as regular jobs, with all the job resources known in advance. All the pilot communication with the PanDA server is done by the arcControlTower, so it plays the role of a pilot factory and the pilot itself. There are several advantages to this approach: no grid middleware is needed on the worker nodes, the fair-share between the production and user jobs is tuned with the arcControlTower load parameters, the jobs can be controlled by ARC client tools. The system could be extended to other submission systems using central distribution.

  11. ATLAS silicon microstrip detector system (SCT)

    International Nuclear Information System (INIS)

    Unno, Y.

    2003-01-01

    The S CT together with the pixel and the transition radiation tracker systems and with a central solenoid forms the central tracking system of the ATLAS detector at LHC. Series production of SCT Silicon microstrip sensors is near completion. The sensors have been shown to be robust against high voltage operation to the 500 V required after fluences of 3x10 14 protons/cm 2 . SCT barrel modules are in series production. A low-noise CCD camera has been used to debug the onset of leakage currents

  12. AGIS: The ATLAS Grid Information System

    CERN Document Server

    Anisenkov, A; The ATLAS collaboration; Klimentov, A; Senchenko, A

    2012-01-01

    The ATLAS Computing model embraces the Grid paradigm and a high degree of decentralization and computing resources able to meet ATLAS requirements of petabytes scale data operations. In this paper we present ATLAS Grid Information System (AGIS) designed to integrate configuration and status information about resources, services and topology of whole ATLAS Grid needed by ATLAS Distributed Computing applications and services.

  13. Recent ATLAS Results in Charmonium Production

    CERN Document Server

    Reznicek, Pavel; The ATLAS collaboration

    2018-01-01

    Recent results of production of charmonium in pp and heavy-ion collisions at ATLAS. Production of $J/\\psi$ and $\\psi$(2S) in 7 and 8 TeV pp collisions. Production of $J/\\psi$, $\\psi$(2S) and $\\Upsilon$(nS) in 5.02 TeV pp and p+Pb collisions.

  14. ATLAS Review Office

    CERN Multimedia

    Szeless, B

    The ATLAS internal reviews, be it the mandatory Production Readiness Reviews, the now newly installed Production Advancement Reviews, or the more and more requested different Design Reviews, have become a part of our ATLAS culture over the past years. The Activity Systems Status Overviews are, for the time being, a one in time event and should be held for each system as soon as possible to have some meaning. There seems to a consensus that the reviews have become a useful project tool for the ATLAS management but even more so for the sub-systems themselves making achievements as well as possible shortcomings visible. One other recognized byproduct is the increasing cross talk between the systems, a very important ingredient to make profit all the systems from the large collective knowledge we dispose of in ATLAS. In the last two months, the first two PARs were organized for the MDT End Caps and the TRT Barrel Modules, both part of the US contribution to the ATLAS Project. Furthermore several different design...

  15. Alignment of the ATLAS Inner Detector Tracking System

    CERN Document Server

    Heller, C; The ATLAS collaboration

    2011-01-01

    ATLAS is one of the multipurpose experiments that records the products of the LHC proton-proton and heavy ion collisions. In order to reconstruct trajectories of charged particles produced in these collisions, ATLAS is equipped with a tracking system built using two different technologies, silicon planar sensors (pixel and microstrips) and drift-tube based detectors. Together they constitute the ATLAS Inner Detector, which is embedded in a 2 T axial field. Efficiently reconstructing tracks from charged particles traversing the detector, and precisely measure their momenta is of crucial importance for physics analyses. In order to achieve its scientific goals, an alignment of the ATLAS Inner Detector is required to accurately determine its more than 700,000 degrees of freedom. The goal of the alignment is set such that the limited knowledge of the sensor locations should not deteriorate the resolution of track parameters by more than 20% with respect to the intrinsic tracker resolution. The implementation of t...

  16. Production experience with the ATLAS Event Service

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00066086; The ATLAS collaboration; Calafiura, Paolo; Childers, John Taylor; De, Kaushik; Guan, Wen; Maeno, Tadashi; Nilsson, Paul; Tsulaia, Vakhtang; van Gemmeren, Peter; Wenaus, Torre

    2017-01-01

    The ATLAS Event Service (AES) has been designed and implemented for efficient running of ATLAS production workflows on a variety of computing platforms, ranging from conventional Grid sites to opportunistic, often short-lived resources, such as spot market commercial clouds, supercomputers and volunteer computing. The Event Service architecture allows real time delivery of fine grained workloads to running payload applications which process dispatched events or event ranges and immediately stream the outputs to highly scalable Object Stores. Thanks to its agile and flexible architecture the AES is currently being used by grid sites for assigning low priority workloads to otherwise idle computing resources; similarly harvesting HPC resources in an efficient back-fill mode; and massively scaling out to the 50-100k concurrent core level on the Amazon spot market to efficiently utilize those transient resources for peak production needs. Platform ports in development include ATLAS@Home (BOINC) and the Google Comp...

  17. Production Experience with the ATLAS Event Service

    CERN Document Server

    Benjamin, Douglas; The ATLAS collaboration

    2016-01-01

    The ATLAS Event Service (ES) has been designed and implemented for efficient running of ATLAS production workflows on a variety of computing platforms, ranging from conventional Grid sites to opportunistic, often short-lived resources, such as spot market commercial clouds, supercomputers and volunteer computing. The Event Service architecture allows real time delivery of fine grained workloads to running payload applications which process dispatched events or event ranges and immediately stream the outputs to highly scalable Object Stores. Thanks to its agile and flexible architecture the ES is currently being used by grid sites for assigning low priority workloads to otherwise idle computing resources; similarly harvesting HPC resources in an efficient back-fill mode; and massively scaling out to the 50-100k concurrent core level on the Amazon spot market to efficiently utilize those transient resources for peak production needs. Platform ports in development include ATLAS@Home (BOINC) and the Goggle Comput...

  18. A quality control atlas for scintillation camera systems

    International Nuclear Information System (INIS)

    Busemann Sokole, E.; Graham, L.S.; Todd-Pokropek, A.; Wegst, A.; Robilotta, C.C.

    2002-01-01

    Full text: The accurate interpretation of quality control and clinical nuclear medicine image data is coupled to an understanding of image patterns and quantitative results. Understanding is gained by learning from different examples, and knowledge of underlying principles of image production. An Atlas of examples has been created to assist with interpreting quality control tests and recognizing artifacts in clinical examples. The project was initiated and supported by the International Atomic Energy Agency (IAEA). The Atlas was developed and written by Busemann Sokole from image examples submitted from nuclear medicine users from around the world. The descriptive text was written in a consistent format to accompany each image or image set. Each example in the atlas finally consisted of the images; a brief description of the data acquisition, radionuclide/radiopharmaceutical, specific circumstances under which the image was produced; results describing the images and subsequent conclusions; comments, where appropriate, giving guidelines for follow-up strategies and trouble shooting; and occasional literature references. Hardcopy images required digitizing into JPEG format for inclusion into a digital document. Where possible, an example was contained on one page. The atlas was reviewed by an international group of experts. A total of about 250 examples were compiled into 6 sections: planar, SPECT, whole body, camera/computer interface, environment/radioactivity, and display/hardcopy. Subtle loss of image quality may be difficult to detect. SPECT examples, therefore, include simulations demonstrating effects of deterioration in camera performance (e.g. center-of-rotation offset, non-uniformity) or suboptimal clinical performance. The atlas includes normal results, results from poor adjustment of the camera system, poor results obtained at acceptance testing, artifacts due to system malfunction, and artifacts due to environmental situations. Some image patterns are

  19. RiceAtlas, a spatial database of global rice calendars and production.

    Science.gov (United States)

    Laborte, Alice G; Gutierrez, Mary Anne; Balanza, Jane Girly; Saito, Kazuki; Zwart, Sander J; Boschetti, Mirco; Murty, M V R; Villano, Lorena; Aunario, Jorrel Khalil; Reinke, Russell; Koo, Jawoo; Hijmans, Robert J; Nelson, Andrew

    2017-05-30

    Knowing where, when, and how much rice is planted and harvested is crucial information for understanding the effects of policy, trade, and global and technological change on food security. We developed RiceAtlas, a spatial database on the seasonal distribution of the world's rice production. It consists of data on rice planting and harvesting dates by growing season and estimates of monthly production for all rice-producing countries. Sources used for planting and harvesting dates include global and regional databases, national publications, online reports, and expert knowledge. Monthly production data were estimated based on annual or seasonal production statistics, and planting and harvesting dates. RiceAtlas has 2,725 spatial units. Compared with available global crop calendars, RiceAtlas is nearly ten times more spatially detailed and has nearly seven times more spatial units, with at least two seasons of calendar data, making RiceAtlas the most comprehensive and detailed spatial database on rice calendar and production.

  20. AGIS: The ATLAS Grid Information System

    CERN Document Server

    Anisenkov, A; The ATLAS collaboration; Klimentov, A; Oleynik, D; Petrosyan, A

    2014-01-01

    In this paper we describe ATLAS Grid Information System (AGIS), the system designed to integrate configuration and status information about resources, services and topology of the computing infrastructure used by ATLAS Distributed Computing applications and services.

  1. AGIS: The ATLAS Grid Information System

    OpenAIRE

    Anisenkov, A; Di Girolamo, A; Klimentov, A; Oleynik, D; Petrosyan, A

    2013-01-01

    In this paper we describe ATLAS Grid Information System (AGIS), the system designed to integrate configuration and status information about resources, services and topology of the computing infrastructure used by ATLAS Distributed Computing applications and services.

  2. The consistency service of the ATLAS Distributed Data Management system

    CERN Document Server

    Serfon, C; The ATLAS collaboration

    2011-01-01

    With the continuously increasing volume of data produced by ATLAS and stored on the WLCG sites, the probability of data corruption or data losses, due to software and hardware failures is increasing. In order to ensure the consistency of all data produced by ATLAS a Consistency Service has been developed as part of the DQ2 Distributed Data Management system. This service is fed by the different ATLAS tools, i.e. the analysis tools, production tools, DQ2 site services or by site administrators that report corrupted or lost files. It automatically corrects the errors reported and informs the users in case of irrecoverable file loss.

  3. The Consistency Service of the ATLAS Distributed Data Management system

    CERN Document Server

    Serfon, C; The ATLAS collaboration

    2010-01-01

    With the continuously increasing volume of data produced by ATLAS and stored on the WLCG sites, the probability of data corruption or data losses, due to software and hardware failure is increasing. In order to ensure the consistency of all data produced by ATLAS a Consistency Service has been developed as part of the DQ2 Distributed Data Management system. This service is fed by the different ATLAS tools, i.e. the analysis tools, production tools, DQ2 site services or by site administrators that report corrupted or lost files. It automatically correct the errors reported and informs the users in case of irrecoverable file loss.

  4. ATLAS measurements of vector boson production

    CERN Document Server

    Levchenko, M; The ATLAS collaboration

    2014-01-01

    ATLAS measurements of vector boson production with associated jets Productions of light and heavy-flavour jets in association with a W or a Z boson in proton-proton collisions are important processes to study QCD in multi-scale environments and the proton parton content. The cross section, differential in several kinematics variables, have been measured with the ATLAS detector in 7 TeV proton-proton collisions and compared to high-order QCD calculations and Monte Carlo simulations. The results demonstrate the need for the inclusion of high-multiplicity matrix elements in the calculations of high jet multiplicities. The ratio of (Z+jets)/(W+jets) provides a precise test of QCD due to the large cancellations of theoretical and experimental uncertainties. Measurement of W+c production cross section has a unique sensitivity to the strange-quark density, which is poorly known at low x. W or Z boson production in association with b-quark jets, on the other hand, probes the b-quark density in the proton and the b-qu...

  5. ATLAS Data Challenge 2 A massive Monte Carlo production on the GRID

    CERN Document Server

    González de la Hoz, S; Lozano, J; Salt, J; Fassi, F; March, L; Adams, D; Deng, W; Nevski, P; Smith, J; Yu, D; Zhao, X; Poulard, G; Goossens, L; Nairz, A; Branco, M; Benekos, N C; Sturrock, R; Walker, R; Vetterli, M; Chudoba, J; Tas, P; Duckeck, G; Kennedy, J; Nielsen, J; Wäänänen, A; Bernardet, K; Negri, G; Rebatto, D; De Salvo, A; Perini, L; Vaccarossa, L; Ould-Saada, F; Read, A; Merino, G; Smirnova, O G; Ellert, M; Quing, D; Brochu, F; Gieraltowski, J; Youssef, S; De, K; Oz-turk, N; Sosebee, M; Severini, H; Gardner, R; Mambeli, M; Smirnov, Y; European Grid Conference

    2005-01-01

    The study and validation of the ATLAS Computing Model started three years ago and will continue for few years in the context of the so-called Data Chal-lenges (DC). DC1 was conducted during 2002-03; the main goals achieved were to set up the simulation data production infrastructure in a real worldwide collaborative effort and to gain experience in exercising an ATLAS wide production model. DC2 (from May until December 2004) is divided into three phases: (i) generate Monte Carlo data using GEANT4 on three different Grid projects: LCG, GRID3 and NorduGrid; (ii) simulate the first pass recon-struction of real data expected in 2007, and (iii) test the Distributed Analysis model. Experience with the use of the system in world-wide DC2 production of ten million events will be presented. We also present how the three Grid fla-vours are operated. Finally we discuss the first prototypes of Distributed Analy-sis systems.

  6. gLExec Integration with the ATLAS PanDA Workload Management System

    CERN Document Server

    Edward Karavakis; The ATLAS collaboration; Campana, Simone; De, Kaushik; Di Girolamo, Alessandro; Maarten Litmaath; Maeno, Tadashi; Medrano Llamas, Ramon; Nilsson, Paul; Wenaus, Torre

    2015-01-01

    The ATLAS Experiment at the Large Hadron Collider has collected data during Run 1 and is ready to collect data in Run 2. The ATLAS data are distributed, processed and analysed at more than 130 grid and cloud sites across the world. At any given time, there are more than 150,000 concurrent jobs running and about a million jobs are submitted on a daily basis on behalf of thousands of physicists within the ATLAS collaboration. The Production and Distributed Analysis (PanDA) workload management system has proved to be a key component of ATLAS and plays a crucial role in the success of the large-scale distributed computing as it is the sole system for distributed processing of Grid jobs across the collaboration since October 2007. ATLAS user jobs are executed on worker nodes by pilots sent to the sites by pilot factories. This pilot architecture has greatly improved job reliability and although it has clear advantages, such as making the working environment homogeneous by hiding any potential heterogeneities, the ...

  7. Predictive analytics as an essential mechanism for situational awareness at the ATLAS Production System

    CERN Document Server

    Titov, Mikhail; The ATLAS collaboration

    2017-01-01

    The workflow management process should be under the control of the certain service that is able to forecast the processing time dynamically according to the status of the processing environment and workflow itself, and to react immediately on any abnormal behaviour of the execution process. Such situational awareness analytic service would provide the possibility to monitor the execution process, to detect the source of any malfunction, and to optimize the management process. The stated service for the second generation of the ATLAS Production System (ProdSys2, an automated scheduling system) is based on predictive analytics approach to estimate the duration of the data processings (in terms of ProdSys2, it is task and chain of tasks) with later usage in decision making processes. Machine learning ensemble methods are chosen to estimate completion time (i.e., “Time To Complete”, TTC) for every (production) task and chain of tasks, thus “abnormal” task processing times would warn about possible failure...

  8. Experiences with the new ATLAS Distributed Data Management System

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00214543; The ATLAS collaboration; Serfon, Cedric; Barisits, Martin-Stefan; Lassnig, Mario; Beermann, Thomas; Guan, Wen

    2017-01-01

    The ATLAS Distributed Data Management (DDM) system has evolved drastically in the last two years with the Rucio software fully replacing the previous system before the start of LHC Run-2. The ATLAS DDM system manages now more than 250 petabytes spread on 130 storage sites and can handle file transfer rates of up to 30Hz. In this paper, we discuss our experience acquired in developing, commissioning, running and maintaining such a large system. First, we describe the general architecture of the system, our integration with external services like the WLCG File Transfer Service and the evolution of the system over its first years of production. Then, we show the performance of the system, describe the integration of new technologies such as object stores, and outline some new developments, which mainly focus on performance and automation.

  9. Quality Assurance and Functionality Tests on Electrical Components during the ATLAS IBL Production

    CERN Document Server

    Jentzsch, J

    2013-01-01

    To improve performance of the ATLAS inner tracker, a fourth Pixel layer, called the Insertable B-layer (IBL), will be installed in 2014 on a new beam pipe. A new read out chip generation, FE-I4, has been developed and two different sensor designs, a rather conventional planar and a 3D design, have been flip chipped to these front ends. New staves holding new stave and module flex circuits have been developed as well. Therefore, a production QA test bench has been established to test all production staves before integration with the new beam pipe. This setup combines former ATLAS Pixel services and a new readout system, namely the RCE (Reconfigurable Cluster Element) system developed at SLAC. With this setup all production staves will be tested to ensure the installation of only those staves which fulfill the IBL criteria. Quality assurance measurements under cleanroom conditions, including temperature and humidity control, are performed on the individual components during the various production steps of the I...

  10. ATLAS TDAQ System Administration:

    CERN Document Server

    Lee, Christopher Jon; The ATLAS collaboration; Bogdanchikov, Alexander; Ballestrero, Sergio; Contescu, Alexandru Cristian; Dubrov, Sergei; Fazio, Daniel; Korol, Aleksandr; Scannicchio, Diana; Twomey, Matthew Shaun; Voronkov, Artem

    2015-01-01

    The ATLAS Trigger and Data Acquisition (TDAQ) system is responsible for the online processing of live data, streaming from the ATLAS experiment at the Large Hadron Collider (LHC) at CERN. The online farm is composed of ̃3000 servers, processing the data readout from ̃100 million detector channels through multiple trigger levels. During the two years of the first Long Shutdown (LS1) there has been a tremendous amount of work done by the ATLAS TDAQ System Administrators, implementing numerous new software applications, upgrading the OS and the hardware, changing some design philosophies and exploiting the High Level Trigger farm with different purposes. During the data taking only critical security updates are applied and broken hardware is replaced to ensure a stable operational environment. The LS1 provided an excellent opportunity to look into new technologies and applications that would help to improve and streamline the daily tasks of not only the System Administrators, but also of the scientists who wil...

  11. Production experience with the ATLAS Event Service

    Science.gov (United States)

    Benjamin, D.; Calafiura, P.; Childers, T.; De, K.; Guan, W.; Maeno, T.; Nilsson, P.; Tsulaia, V.; Van Gemmeren, P.; Wenaus, T.; ATLAS Collaboration

    2017-10-01

    The ATLAS Event Service (AES) has been designed and implemented for efficient running of ATLAS production workflows on a variety of computing platforms, ranging from conventional Grid sites to opportunistic, often short-lived resources, such as spot market commercial clouds, supercomputers and volunteer computing. The Event Service architecture allows real time delivery of fine grained workloads to running payload applications which process dispatched events or event ranges and immediately stream the outputs to highly scalable Object Stores. Thanks to its agile and flexible architecture the AES is currently being used by grid sites for assigning low priority workloads to otherwise idle computing resources; similarly harvesting HPC resources in an efficient back-fill mode; and massively scaling out to the 50-100k concurrent core level on the Amazon spot market to efficiently utilize those transient resources for peak production needs. Platform ports in development include ATLAS@Home (BOINC) and the Google Compute Engine, and a growing number of HPC platforms. After briefly reviewing the concept and the architecture of the Event Service, we will report the status and experience gained in AES commissioning and production operations on supercomputers, and our plans for extending ES application beyond Geant4 simulation to other workflows, such as reconstruction and data analysis.

  12. Argonne's atlas control system upgrade

    International Nuclear Information System (INIS)

    Munson, F.; Quock, D.; Chapin, B.; Figueroa, J.

    1999-01-01

    The ATLAS facility (Argonne Tandem-Linac Accelerator System) is located at the Argonne National Laboratory. The facility is a tool used in nuclear and atomic physics research, which focuses primarily on heavy-ion physics. The accelerator as well as its control system are evolutionary in nature, and consequently, continue to advance. In 1998 the most recent project to upgrade the ATLAS control system was completed. This paper briefly reviews the upgrade, and summarizes the configuration and features of the resulting control system

  13. Integration Of PanDA Workload Management System With Supercomputers for ATLAS

    CERN Document Server

    Oleynik, Danila; The ATLAS collaboration; De, Kaushik; Wenaus, Torre; Maeno, Tadashi; Barreiro Megino, Fernando Harald; Nilsson, Paul; Guan, Wen; Panitkin, Sergey

    2016-01-01

    The Large Hadron Collider (LHC), operating at the international CERN Laboratory in Geneva, Switzerland, is leading Big Data driven scientific explorations. Experiments at the LHC explore the fundamental nature of matter and the basic forces that shape our universe, and were recently credited for the discovery of a Higgs boson. ATLAS, one of the largest collaborations ever assembled in the sciences, is at the forefront of research at the LHC. To address an unprecedented multi-petabyte data processing challenge, the ATLAS experiment is relying on a heterogeneous distributed computational infrastructure. The ATLAS experiment uses PanDA (Production ANd Distributed Analysis system) Workload Management System for managing the workflow for all data processing on over 150 data centers. Through PanDA, ATLAS physicists see a single computing facility that enables rapid scientific breakthroughs for the experiment, even though the data centers are physically scattered all over the world. While PanDA currently uses more t...

  14. The ATLAS IBL CO2 Cooling System

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00237783; The ATLAS collaboration; Zwalinski, L.; Bortolin, C.; Vogt, S.; Godlewski, J.; Crespo-Lopez, O.; Van Overbeek, M.; Blaszcyk, T.

    2017-01-01

    The ATLAS Pixel detector has been equipped with an extra B-layer in the space obtained by a reduced beam pipe. This new pixel detector called the ATLAS Insertable B-Layer (IBL) is installed in 2014 and is operational in the current ATLAS data taking. The IBL detector is cooled with evaporative CO2 and is the first of its kind in ATLAS. The ATLAS IBL CO2 cooling system is designed for lower temperature operation (<-35⁰C) than the previous developed CO2 cooling systems in High Energy Physics experiments. The cold temperatures are required to protect the pixel sensors for the high expected radiation dose up to 550 fb^-1 integrated luminosity.

  15. The ATLAS IBL CO2 Cooling System

    CERN Document Server

    Verlaat, Bartholomeus; The ATLAS collaboration

    2016-01-01

    The Atlas Pixel detector has been equipped with an extra B-layer in the space obtained by a reduced beam pipe. This new pixel detector called the ATLAS Insertable B-Layer (IBL) is installed in 2014 and is operational in the current ATLAS data taking. The IBL detector is cooled with evaporative CO2 and is the first of its kind in ATLAS. The ATLAS IBL CO2 cooling system is designed for lower temperature operation (<-35⁰C) than the previous developed CO2 cooling systems in High Energy Physics experiments. The cold temperatures are required to protect the pixel sensors for the high expected radiation dose up to 550 fb^-1 integrated luminosity. This paper describes the design, development, construction and commissioning of the IBL CO2 cooling system. It describes the challenges overcome and the important lessons learned for the development of future systems which are now under design for the Phase-II upgrade detectors.

  16. Planetary Data Systems (PDS) Imaging Node Atlas II

    Science.gov (United States)

    Stanboli, Alice; McAuley, James M.

    2013-01-01

    The Planetary Image Atlas (PIA) is a Rich Internet Application (RIA) that serves planetary imaging data to the science community and the general public. PIA also utilizes the USGS Unified Planetary Coordinate system (UPC) and the on-Mars map server. The Atlas was designed to provide the ability to search and filter through greater than 8 million planetary image files. This software is a three-tier Web application that contains a search engine backend (MySQL, JAVA), Web service interface (SOAP) between server and client, and a GWT Google Maps API client front end. This application allows for the search, retrieval, and download of planetary images and associated meta-data from the following missions: 2001 Mars Odyssey, Cassini, Galileo, LCROSS, Lunar Reconnaissance Orbiter, Mars Exploration Rover, Mars Express, Magellan, Mars Global Surveyor, Mars Pathfinder, Mars Reconnaissance Orbiter, MESSENGER, Phoe nix, Viking Lander, Viking Orbiter, and Voyager. The Atlas utilizes the UPC to translate mission-specific coordinate systems into a unified coordinate system, allowing the end user to query across missions of similar targets. If desired, the end user can also use a mission-specific view of the Atlas. The mission-specific views rely on the same code base. This application is a major improvement over the initial version of the Planetary Image Atlas. It is a multi-mission search engine. This tool includes both basic and advanced search capabilities, providing a product search tool to interrogate the collection of planetary images. This tool lets the end user query information about each image, and ignores the data that the user has no interest in. Users can reduce the number of images to look at by defining an area of interest with latitude and longitude ranges.

  17. ATLAS Distributed Computing

    CERN Document Server

    Schovancova, J; The ATLAS collaboration

    2011-01-01

    The poster details the different aspects of the ATLAS Distributed Computing experience after the first year of LHC data taking. We describe the performance of the ATLAS distributed computing system and the lessons learned during the 2010 run, pointing out parts of the system which were in a good shape, and also spotting areas which required improvements. Improvements ranged from hardware upgrade on the ATLAS Tier-0 computing pools to improve data distribution rates, tuning of FTS channels between CERN and Tier-1s, and studying data access patterns for Grid analysis to improve the global processing rate. We show recent software development driven by operational needs with emphasis on data management and job execution in the ATLAS production system.

  18. The ATLAS PanDA Monitoring System and its Evolution

    Science.gov (United States)

    Klimentov, A.; Nevski, P.; Potekhin, M.; Wenaus, T.

    2011-12-01

    The PanDA (Production and Distributed Analysis) Workload Management System is used for ATLAS distributed production and analysis worldwide. The needs of ATLAS global computing imposed challenging requirements on the design of PanDA in areas such as scalability, robustness, automation, diagnostics, and usability for both production shifters and analysis users. Through a system-wide job database, the PanDA monitor provides a comprehensive and coherent view of the system and job execution, from high level summaries to detailed drill-down job diagnostics. It is (like the rest of PanDA) an Apache-based Python application backed by Oracle. The presentation layer is HTML code generated on the fly in the Python application which is also responsible for managing database queries. However, this approach is lacking in user interface flexibility, simplicity of communication with external systems, and ease of maintenance. A decision was therefore made to migrate the PanDA monitor server to Django Web Application Framework and apply JSON/AJAX technology in the browser front end. This allows us to greatly reduce the amount of application code, separate data preparation from presentation, leverage open source for tools such as authentication and authorization mechanisms, and provide a richer and more dynamic user experience. We describe our approach, design and initial experience with the migration process.

  19. The ATLAS PanDA Monitoring System and its Evolution

    International Nuclear Information System (INIS)

    Klimentov, A; Nevski, P; Wenaus, T; Potekhin, M

    2011-01-01

    The PanDA (Production and Distributed Analysis) Workload Management System is used for ATLAS distributed production and analysis worldwide. The needs of ATLAS global computing imposed challenging requirements on the design of PanDA in areas such as scalability, robustness, automation, diagnostics, and usability for both production shifters and analysis users. Through a system-wide job database, the PanDA monitor provides a comprehensive and coherent view of the system and job execution, from high level summaries to detailed drill-down job diagnostics. It is (like the rest of PanDA) an Apache-based Python application backed by Oracle. The presentation layer is HTML code generated on the fly in the Python application which is also responsible for managing database queries. However, this approach is lacking in user interface flexibility, simplicity of communication with external systems, and ease of maintenance. A decision was therefore made to migrate the PanDA monitor server to Django Web Application Framework and apply JSON/AJAX technology in the browser front end. This allows us to greatly reduce the amount of application code, separate data preparation from presentation, leverage open source for tools such as authentication and authorization mechanisms, and provide a richer and more dynamic user experience. We describe our approach, design and initial experience with the migration process.

  20. Bringing ATLAS production to HPC resources. A case study with SuperMuc and Hydra

    Energy Technology Data Exchange (ETDEWEB)

    Duckeck, Guenter; Walker, Rodney [LMU Muenchen (Germany); Kennedy, John; Mazzaferro, Luca [RZG Garching (Germany); Kluth, Stefan [Max-Planck-Institut fuer Physik, Muenchen (Germany); Collaboration: ATLAS-Collaboration

    2015-07-01

    The possible usage of Supercomputer systems or HPC resources by ATLAS is now becoming viable due to the changing nature of these systems and it is also very attractive due to the need for increasing amounts of simulated data. The ATLAS experiment at CERN will begin a period of high luminosity data taking in 2015. The corresponding need for simulated data might potentially exceed the capabilities of the current Grid infrastructure. ATLAS aims to address this need by opportunistically accessing resources such as cloud and HPC systems. This contribution presents the results of two projects undertaken by LMU/LRZ and MPP/RZG to use the supercomputer facilities SuperMuc (LRZ) and Hydra (RZG). Both are Linux based supercomputers in the 100 k CPU-core category. The integration of such HPC resources into the ATLAS production system poses many challenges. Firstly, established techniques and features of standard WLCG operation are prohibited or much restricted on HPC systems, e.g. Grid middleware, software installation, outside connectivity, etc. Secondly, efficient use of available resources requires massive multi-core jobs, back-fill submission and check-pointing. We discuss the customization of these components and the strategies for HPC usage as well as possibilities for future directions.

  1. AGIS: Evolution of Distributed Computing Information system for ATLAS

    CERN Document Server

    Anisenkov, Alexey; The ATLAS collaboration; Alandes Pradillo, Maria; Karavakis, Edward

    2015-01-01

    The variety of the ATLAS Computing Infrastructure requires a central information system to define the topology of computing resources and to store the different parameters and configuration data which are needed by the various ATLAS software components. The ATLAS Grid Information System is the system designed to integrate configuration and status information about resources, services and topology of the computing infrastructure used by ATLAS Distributed Computing applications and services.

  2. Development of the detector control system for the ATLAS Level-1 trigger and measurement of the single top production cross section

    CERN Document Server

    Curtis, Christopher J

    This thesis discusses the development of the Detector Control System (DCS) for the ATLAS Level-1 Trigger. Microcontroller code has been developed to read out slow controls data from the Level-1 Calorimeter Trigger modules into the wider DCS. Back-end software has been developed for archiving this data. A Finite State Machine (FSM) has also been developed to offer remote access to the L1 Trigger hardware from the ATLAS Control Room. This Thesis also discusses the discovery potential for electroweak single top production during early running. Using Monte Carlo data some of the major systematics are discussed. A potential upper limit on the production cross section is calculated to be 45.2 pb. If the Standard Model prediction is assumed, a measured signal could potentially have a significance of up to 2.23¾ using 200 pb−1 of data.

  3. ATLAS Distributed Computing Automation

    CERN Document Server

    Schovancova, J; The ATLAS collaboration; Borrego, C; Campana, S; Di Girolamo, A; Elmsheuser, J; Hejbal, J; Kouba, T; Legger, F; Magradze, E; Medrano Llamas, R; Negri, G; Rinaldi, L; Sciacca, G; Serfon, C; Van Der Ster, D C

    2012-01-01

    The ATLAS Experiment benefits from computing resources distributed worldwide at more than 100 WLCG sites. The ATLAS Grid sites provide over 100k CPU job slots, over 100 PB of storage space on disk or tape. Monitoring of status of such a complex infrastructure is essential. The ATLAS Grid infrastructure is monitored 24/7 by two teams of shifters distributed world-wide, by the ATLAS Distributed Computing experts, and by site administrators. In this paper we summarize automation efforts performed within the ATLAS Distributed Computing team in order to reduce manpower costs and improve the reliability of the system. Different aspects of the automation process are described: from the ATLAS Grid site topology provided by the ATLAS Grid Information System, via automatic site testing by the HammerCloud, to automatic exclusion from production or analysis activities.

  4. ATLAS Tile calorimeter calibration and monitoring systems

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00445232; The ATLAS collaboration

    2016-01-01

    The ATLAS Tile Calorimeter (TileCal) is the central section of the hadronic calorimeter of the ATLAS experiment and provides important information for reconstruction of hadrons, jets, hadronic decays of tau leptons and missing transverse energy. This sampling calorimeter uses steel plates as absorber and scintillating tiles as active medium. The light produced by the passage of charged particles is transmitted by wavelength shifting fibres to photomultiplier tubes (PMTs), located on the outside of the calorimeter. The readout is segmented into about 5000 cells (longitudinally and transversally), each of them being read out by two PMTs in parallel. To calibrate and monitor the stability and performance of each part of the readout chain during the data taking, a set of calibration systems is used. The TileCal calibration system comprises Cesium radioactive sources, laser and charge injection elements and it allows to monitor and equalize the calorimeter response at each stage of the signal production, from scin...

  5. ATLAS Tile calorimeter calibration and monitoring systems

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00445232; The ATLAS collaboration

    2017-01-01

    The ATLAS Tile Calorimeter (TileCal) is the central section of the hadronic calorimeter of the ATLAS experiment and provides important information for reconstruction of hadrons, jets, hadronic decays of tau leptons and missing transverse energy. This sampling calorimeter uses steel plates as absorber and scintillating tiles as active medium. The light produced by the passage of charged particles is transmitted by wavelength shifting fibres to photomultiplier tubes (PMTs), located on the outside of the calorimeter. The readout is segmented into about 5000 cells (longitudinally and transversally), each of them being read out by two PMTs in parallel. To calibrate and monitor the stability and performance of each part of the readout chain during the data taking, a set of calibration systems is used. The TileCal calibration system comprises cesium radioactive sources, Laser and charge injection elements, and allows for monitoring and equalization of the calorimeter response at each stage of the signal production, ...

  6. ATLAS reach for Quarkonium production and polarization measurements

    CERN Document Server

    Etzion, Erez; 8th International Conference on Hyperons, Charm and Beauty Hadrons

    2009-01-01

    The ATLAS detector at CERN's LHC is preparing to take data from the first proton-proton collisions expected in the next few months. We report on the analysis of simulated data samples for production of heavy Quarkonium states J/psi and Upsilon, corresponding to an integrated luminosity of 10 pb^-1 with center of mass energy of 14 TeV expected at the early ATLAS data. We review various aspects of prompt Quarkonium production at LHC: the accessible ranges in transverse momentum and pseudorapidity, spin alignment of vector states, separation of color octet and color singlet production mechanism and feasibility of observing radiative decays Xi_c and Xi_b decays. Strategies of various measurements are outlined and methods of separating promptly produced J/psi and Upsilon mesons from various backgrounds are discussed.

  7. A system for managing information at ATLAS

    International Nuclear Information System (INIS)

    Tilbrook, I.R.

    1993-01-01

    In response to a need for better management of maintenance and document information at the Argonne Tandem-Linear Accelerating System (ATLAS), the ATLAS Information Management System (AIMS) has been created. The system is based on the relational database model. The system's applications use the Alpha-4 relational database management system, a commercially available software package. The system's function and design are described

  8. Evolution of the Argonne Tandem Linear Accelerator System (ATLAS) control system

    International Nuclear Information System (INIS)

    Power, M.; Munson, F.

    2012-01-01

    Given that the Argonne Tandem Linear Accelerator System (ATLAS) recently celebrated its 25. anniversary, this paper will explore the past, present, and future of the ATLAS Control System, and how it has evolved along with the accelerator and control system technology. ATLAS as we know it today, originated with a Tandem Van de Graff in the sixties. With the addition of the Booster section in the late seventies, came the first computerized control. ATLAS itself was placed into service on June 25, 1985, and was the world's first superconducting linear accelerator for ions. Since its dedication as a National User Facility, more than a thousand experiments by more than 2,000 users worldwide, have taken advantage of the unique capabilities it provides. Today, ATLAS continues to be a user facility for physicists who study the particles that form the heart of atoms. Its most recent addition, CARIBU (Californium Rare Isotope Breeder Upgrade), creates special beams that feed into ATLAS. ATLAS is similar to a living organism, changing and responding to new technological challenges and research needs. As it continues to evolve, so does the control system: from the original days using a DEC PDP-11/34 computer and two CAMAC crates, to a DEC Alpha computer running Vsystem software and more than twenty CAMAC crates, to distributed computers and VME systems. Future upgrades are also in the planning stages that will continue to evolve the control system. (authors)

  9. Testbeam studies of production modules of the ATLAS Tile Calorimeter

    OpenAIRE

    Adragna, P.; Alexa, C.; Anderson, K.; Antonaki, A.; Arabidze, A.; Batkova, L.; Batusov, V.; Beck, H.P.; Bednar, P.; Bergeaas Kuutmann, E.; Biscarat, C.; Blanchot, G.; Bogush, A.; Bohm, C.; Boldea, V.

    2009-01-01

    We report test beam studies of {11\\,\\%} of the production ATLAS Tile Calorimeter modules. The modules were equipped with production front-end electronics and all the calibration systems planned for the final detector. The studies used muon, electron and hadron beams ranging in energy from 3~GeV to 350~GeV. Two independent studies showed that the light yield of the calorimeter was $\\sim 70$~pe/GeV, exceeding the design goal by {40\\,\\%}. Electron beams provided a calibration of the modules at t...

  10. The ATLAS beam pick-up based timing system

    International Nuclear Information System (INIS)

    Ohm, C.; Pauly, T.

    2010-01-01

    The ATLAS BPTX stations are composed of electrostatic button pick-up detectors, located 175 m away along the beam pipe on both sides of ATLAS. The pick-ups are installed as a part of the LHC beam instrumentation and used by ATLAS for timing purposes. The usage of the BPTX signals in ATLAS is twofold: they are used both in the trigger system and for LHC beam monitoring. The BPTX signals are discriminated with a constant-fraction discriminator to provide a Level-1 trigger when a bunch passes through ATLAS. Furthermore, the BPTX detectors are used by a stand-alone monitoring system for the LHC bunches and timing signals. The BPTX monitoring system measures the phase between collisions and clock with a precision better than 100 ps in order to guarantee a stable phase relationship for optimal signal sampling in the sub-detector front-end electronics. In addition to monitoring this phase, the properties of the individual bunches are measured and the structure of the beams is determined. On September 10, 2008, the first LHC beams reached the ATLAS experiment. During this period with beam, the ATLAS BPTX system was used extensively to time in the read-out of the sub-detectors. In this paper, we present the performance of the BPTX system and its measurements of the first LHC beams.

  11. Evolution of the ATLAS PanDA Workload Management System for Exascale Computational Science

    OpenAIRE

    Maeno, T; De, K; Klimentov, A; Nilsson, P; Oleynik, D; Panitkin, S; Petrosyan, A; Schovancova, J; Vaniachine, A; Wenaus, T; Yu, D

    2013-01-01

    An important foundation underlying the impressive success of data processing and analysis in the ATLAS experiment [1] at the LHC [2] is the Production and Distributed Analysis (PanDA) workload management system [3]. PanDA was designed specifically for ATLAS and proved to be highly successful in meeting all the distributed computing needs of the experiment. However, the core design of PanDA is not experiment specific. The PanDA workload management system is capable of meeting the needs of othe...

  12. Automated Loads Analysis System (ATLAS)

    Science.gov (United States)

    Gardner, Stephen; Frere, Scot; O’Reilly, Patrick

    2013-01-01

    ATLAS is a generalized solution that can be used for launch vehicles. ATLAS is used to produce modal transient analysis and quasi-static analysis results (i.e., accelerations, displacements, and forces) for the payload math models on a specific Shuttle Transport System (STS) flight using the shuttle math model and associated forcing functions. This innovation solves the problem of coupling of payload math models into a shuttle math model. It performs a transient loads analysis simulating liftoff, landing, and all flight events between liftoff and landing. ATLAS utilizes efficient and numerically stable algorithms available in MSC/NASTRAN.

  13. Virtualization of the ATLAS software environment on a shared HPC system

    CERN Document Server

    Schnoor, Ulrike; The ATLAS collaboration

    2017-01-01

    High-Performance Computing (HPC) and other research cluster computing resources provided by universities can be useful supplements to the collaboration’s own WLCG computing resources for data analysis and production of simulated event samples. The shared HPC cluster "NEMO" at the University of Freiburg has been made available to local ATLAS users through the provisioning of virtual machines incorporating the ATLAS software environment analogously to a WLCG center. The talk describes the concept and implementation of virtualizing the ATLAS software environment to run both data analysis and production on the HPC host system which is connected to the existing Tier-3 infrastructure. Main challenges include the integration into the NEMO and Tier-3 schedulers in a dynamic, on-demand way, the scalability of the OpenStack infrastructure, as well as the automatic generation of a fully functional virtual machine image providing access to the local user environment, the dCache storage element and the parallel file sys...

  14. High precision laser control of the ATLAS tile-calorimeter module mass production at JINR

    International Nuclear Information System (INIS)

    Batusov, V.; Budagov, Yu.; Flyagin, V.; Khubua, D.; Lomakin, Yu.; Lyablin, M.; Rusakovich, N.; Shabalin, D.; Topilin, N.; Nessi, M.

    2001-01-01

    We present a short description of our last few years experience in the quality control of the ATLAS hadron barrel tile-calorimeter module mass production at JINR. A Laser Measurement System (LMS) proposed and realized in Dubna guarantees a high-precision module assembly. The non-planarity of module size surfaces (1.9x5.6 m) controlled area is well within the required ±0.6 mm tolerance for each of JINR assembled modules. The module assembly technique achieved with the LMS system allows us to deliver to CERN one module every 2 weeks. This laser-based measurement system could be used in future for the control measurement of other large-scale units during the ATLAS assembly

  15. Heavy Flavour Production and Decay at ATLAS

    CERN Document Server

    Jones, RWL; The ATLAS collaboration

    2013-01-01

    ATLAS is taking advantage of its large integrated luminosity band sophisticated muon and dimuon triggers to make competitive measurements of heavy flavour production and decay. Inclusive production and heavy flavour jet production is discussed before turning to charm and onium production. The production and decay of individual B hadron species is then addressed, including the current best measurement of the Λb lifetime. A much improved analysis of CP related quantities in Bs decays is presented, before turning to recent results and prospects for rare B decays.

  16. Integration of Globus Online with the ATLAS PanDA Workload Management System

    CERN Document Server

    Contreras, C; The ATLAS collaboration; Maeno, T; Nilsson, P; Potekhin, M

    2012-01-01

    The PanDA Workload Management System is the basis for distributed production and analysis for the ATLAS experiment at the LHC. In this role, it relies on sophisticated dynamic data movement facilities developed in ATLAS. In certain scenarios, such as small research teams in ATLAS Tier-3 sites and non-ATLAS Virtual Organizations, the overhead of installation and operation of these components makes their use not very cost effective. Globus Online is an emerging new tool from the Globus Alliance, which already proved popular within the research community. It provides the users with fast and robust file transfer capabilities that can also be managed from a Web interface, and in addition to grid sites, can have individual workstations and laptops serving as data transmission endpoints. We will describe the integration of the Globus Online functionality into the PanDA suite of software, in order to give more flexibility in choosing the method of data transfer to ATLAS Tier-3 and OSG users.

  17. Integration of Globus Online with the ATLAS PanDA Workload Management System

    International Nuclear Information System (INIS)

    Contreras, C; Deng, W; Maeno, T; Potekhin, M; Nilsson, P

    2012-01-01

    The PanDA Workload Management System is the basis for distributed production and analysis for the ATLAS experiment at the LHC. In this role, it relies on sophisticated dynamic data movement facilities developed in ATLAS. In certain scenarios, such as small research teams in ATLAS Tier-3 sites and non-ATLAS Virtual Organizations, the overhead of installation and operation of these components makes their use not very cost effective. Globus Online is an emerging new tool from the Globus Alliance, which already proved popular within the research community. It provides the users with fast and robust file transfer capabilities that can also be managed from a Web interface, and in addition to grid sites, can have individual workstations and laptops serving as data transmission endpoints. We will describe the integration of the Globus Online functionality into the PanDA suite of software, in order to give more flexibility in choosing the method of data transfer to ATLAS Tier-3 and Open Science Grid (OSG) users.

  18. How to keep the Grid full and working with ATLAS production and physics jobs

    Science.gov (United States)

    Pacheco Pagés, A.; Barreiro Megino, F. H.; Cameron, D.; Fassi, F.; Filipcic, A.; Di Girolamo, A.; González de la Hoz, S.; Glushkov, I.; Maeno, T.; Walker, R.; Yang, W.; ATLAS Collaboration

    2017-10-01

    The ATLAS production system provides the infrastructure to process millions of events collected during the LHC Run 1 and the first two years of Run 2 using grid, clouds and high performance computing. We address in this contribution the strategies and improvements that have been implemented to the production system for optimal performance and to achieve the highest efficiency of available resources from operational perspective. We focus on the recent developments.

  19. How to keep the Grid full and working with ATLAS production and physics jobs

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00221495; The ATLAS collaboration; Barreiro Megino, Fernando Harald; Cameron, David; Fassi, Farida; Filipcic, Andrej; Di Girolamo, Alessandro; Gonzalez de la Hoz, Santiago; Glushkov, Ivan; Maeno, Tadashi; Walker, Rodney; Yang, Wei

    2017-01-01

    The ATLAS production system provides the infrastructure to process millions of events collected during the LHC Run 1 and the first two years of Run 2 using grid, clouds and high performance computing. We address in this contribution the strategies and improvements that have been implemented to the production system for optimal performance and to achieve the highest efficiency of available resources from operational perspective. We focus on the recent developments.

  20. Concepts and Plans towards fast large scale Monte Carlo production for the ATLAS Experiment

    Science.gov (United States)

    Ritsch, E.; Atlas Collaboration

    2014-06-01

    The huge success of the physics program of the ATLAS experiment at the Large Hadron Collider (LHC) during Run 1 relies upon a great number of simulated Monte Carlo events. This Monte Carlo production takes the biggest part of the computing resources being in use by ATLAS as of now. In this document we describe the plans to overcome the computing resource limitations for large scale Monte Carlo production in the ATLAS Experiment for Run 2, and beyond. A number of fast detector simulation, digitization and reconstruction techniques are being discussed, based upon a new flexible detector simulation framework. To optimally benefit from these developments, a redesigned ATLAS MC production chain is presented at the end of this document.

  1. Testbeam Studies of Production Modules of the ATLAS Tile Calorimeter

    CERN Document Server

    Adragna, P; Anderson, K; Antonaki, A; Arabidze, A; Batkova, L; Batusov, V; Beck, H P; Bednar, P; Bergeaas Kuutmann, E; Biscarat, C; Blanchot, G; Bogush, A; Bohm, C; Boldea, V; Bosman, M; Bromberg, C; Budagov, Yu A; Burckhart-Chromek, D; Caprini, M; Caloba, L; Calvet, D; Carli, T; Carvalho, J; Cascella, M; Castelo, J; Castillo, M V; Cavalli-Sforza, M; Cavasinni, V; Cerqueira, A S; Clément, C; Cobal, M; Cogswell, F; Constantinescu, S; Costanzo, D; Corso-Radu, A; Cuenca, C; Damazio, D O; David, M; Davidek, T; De, K; Del Prete, T; Di Girolamo, B; Dita, S; Djobava, T; Dobson, M; Dolejsi, J; Dolezal, Z; Dotti, A; Downing, R; Efthymiopoulos, I; Eriksson, D; Errede, D; Errede, S; Farbin, A; Fassouliotis, D; Febbraro, R; Fedorko, I; Fenyuk, A; Ferdi, C; Ferrer, A; Flaminio, V; Francis, D; Fullana, E; Gadomski, S; Gameiro, S; Garde, V; Gellerstedt, K; Giakoumopoulou, V; Gildemeister, O; Gilewsky, V; Giokaris, N; Gollub, N; Gomes, A; González, V; Gorini, B; Grenier, P; Gris, P; Gruwé, M; Guarino, V; Guicheney, C; Sen-Gupta, A; Haeberli, C; Hakobyan, H; Haney, M; Hellman, S; Henriques, A; Higón, E; Holmgren, S; Hurwitz, M; Huston, J; Iglesias, C; Isaev, A; Jen-La Plante, I; Jon-And, K; Joos, M; Junk, T; Karyukhin, A; Kazarov, A; Khandanyan, H; Khramov, J; Khubua, J; Kolos, S; Korolkov, I; Krivkova, P; Kulchitsky, Y; Kurochkin, Yu; Kuzhir, P; Le Compte, T; Lefèvre, R; Lehmann, G; Leitner, R; Lembesi, M; Lesser, J; Li, J; Liablin, M; Lokajícek, M; Lomakin, Y; Lupi, A; Maidantchik, C; Maio, A; Makouski, M; Maliukov, S; Manousakis, A; Mapelli, L; Marques, C; Marroquim, F; Martin, F; Mazzoni, E; Merritt, F S; Myagkov, A; Miller, R; Minashvili, I; Miralles, L; Montarou, G; Mosidze, M; Némécek, S; Nessi, M; Nodulman, L; Nordkvist, B; Norniella, O; Onofre, A; Oreglia, M; Pallin, D; Pantea, D; Petersen, J; Pilcher, J E; Pina, J; Pinhão, J; Podlyski, F; Portell, X; Poveda, J; Pribyl, L; Price, L E; Proudfoot, J; Ramstedt, M; Richards, R; Roda, C; Romanov, V; Rosnet, P; Roy, P; Ruiz, A; Rumiantsev, V; Russakovich, N; Salto, O; Salvachúa, B; Sanchis, E; Sanders, H; Santoni, C; Santos, J; Saraiva, J G; Sarri, F; Satsunkevitch, I; Says, L-P; Schlager, G; Schlereth, J L; Seixas, J M; Selldén, B; Shalanda, N; Shevtsov, P; Shochet, M; Silva, J; Da Silva, P; Simaitis, V; Simonyan, M; Sisakian, A; Sjölin, J; Solans, C; Solodkov, A; Soloviev, I; Solovyanov, O; Sosebee, M; Spanó, F; Stanek, R; Starchenko, E; Starovoitov, P; Stavina, P; Suk, M; Sykora, I; Tang, F; Tas, P; Teuscher, R; Tokar, S; Topilin, N; Torres, J; Tremblet, L; Tsiareshka, P; Tylmad, M; Underwood, D; Ünel, G; Usai, G; Valero, A; Valkár, S; Valls, J A; Vartapetian, A; Vazeille, F; Vichou, I; Vinogradov, V; Vivarelli, I; Volpi, M; White, A; Zaitsev, A; Zenine, A; Zenis, T

    2009-01-01

    We report test beam studies of {11\\,\\%} of the production ATLAS Tile Calorimeter modules. The modules were equipped with production front-end electronics and all the calibration systems planned for the final detector. The studies used muon, electron and hadron beams ranging in energy from 3~GeV to 350~GeV. Two independent studies showed that the light yield of the calorimeter was $\\sim 70$~pe/GeV, exceeding the design goal by {40\\,\\%}. Electron beams provided a calibration of the modules at the electromagnetic energy scale. Over 200~calorimeter cells the variation of the response was {2.4\\,\\%}. The linearity with energy was also measured. Muon beams provided an intercalibration of the response of all calorimeter cells. The response to muons entering in the ATLAS projective geometry showed an RMS variation of 2.5\\,\\% for 91~measurements over a range of rapidities and modules. The mean response to hadrons of fixed energy had an RMS variation of {1.4\\,\\%} for the modules and projective angles studied. The respon...

  2. Support system for ATLAS distributed computing operations

    CERN Document Server

    Kishimoto, Tomoe; The ATLAS collaboration

    2018-01-01

    The ATLAS distributed computing system has allowed the experiment to successfully meet the challenges of LHC Run 2. In order for distributed computing to operate smoothly and efficiently, several support teams are organized in the ATLAS experiment. The ADCoS (ATLAS Distributed Computing Operation Shifts) is a dedicated group of shifters who follow and report failing jobs, failing data transfers between sites, degradation of ATLAS central computing services, and more. The DAST (Distributed Analysis Support Team) provides user support to resolve issues related to running distributed analysis on the grid. The CRC (Computing Run Coordinator) maintains a global view of the day-to-day operations. In this presentation, the status and operational experience of the support system for ATLAS distributed computing in LHC Run 2 will be reported. This report also includes operations experience from the grid site point of view, and an analysis of the errors that create the biggest waste of wallclock time. The report of oper...

  3. Optimizing new components of PanDA for ATLAS production on HPC resources

    CERN Document Server

    Maeno, Tadashi; The ATLAS collaboration

    2017-01-01

    The Production and Distributed Analysis system (PanDA) has been used for workload management in the ATLAS Experiment for over a decade. It uses pilots to retrieve jobs from the PanDA server and execute them on worker nodes. While PanDA has been mostly used on Worldwide LHC Computing Grid (WLCG) resources for production operations, R&D work has been ongoing on cloud and HPC resources for many years. These efforts have led to the significant usage of large scale HPC resources in the past couple of years. In this talk we will describe the changes to the pilot which enabled the use of HPC sites by PanDA, specifically the Titan supercomputer at Oakridge National Laboratory. Furthermore, it was decided in 2016 to start a fresh redesign of the Pilot with a more modern approach to better serve present and future needs from ATLAS and other collaborations that are interested in using the PanDA System. Another new project for development of a resource oriented service, PanDA Harvester, was also launched in 2016. The...

  4. ATLAS Tile calorimeter calibration and monitoring systems

    Science.gov (United States)

    Chomont, Arthur; ATLAS Collaboration

    2017-11-01

    The ATLAS Tile Calorimeter (TileCal) is the central section of the hadronic calorimeter of the ATLAS experiment and provides important information for reconstruction of hadrons, jets, hadronic decays of tau leptons and missing transverse energy. This sampling calorimeter uses steel plates as absorber and scintillating tiles as active medium. The light produced by the passage of charged particles is transmitted by wavelength shifting fibres to photomultiplier tubes (PMTs), located on the outside of the calorimeter. The readout is segmented into about 5000 cells (longitudinally and transversally), each of them being read out by two PMTs in parallel. To calibrate and monitor the stability and performance of each part of the readout chain during the data taking, a set of calibration systems is used. The TileCal calibration system comprises cesium radioactive sources, Laser and charge injection elements, and allows for monitoring and equalization of the calorimeter response at each stage of the signal production, from scintillation light to digitization. Based on LHC Run 1 experience, several calibration systems were improved for Run 2. The lessons learned, the modifications, and the current LHC Run 2 performance are discussed.

  5. How to keep the Grid full and working with ATLAS production and physics jobs

    CERN Document Server

    Pacheco Pages, Andres; The ATLAS collaboration; Di Girolamo, Alessandro; Walker, Rodney; Filip\\v{c}i\\v{c}, Andrej; Cameron, David; Yang, Wei; Fassi, Farida; Glushkov, Ivan

    2016-01-01

    The ATLAS production system has provided the infrastructure to process of tens of thousand of events during LHC Run1 and the first years of the LHC Run2 using grid, clouds and high performance computing. We address in this contribution several strategies and improvements added to the production system to optimize its performance to get the maximum efficiency of available resources from operational perspective and focusing in detail in the recent developments

  6. Alignment of the ATLAS Inner Detector Tracking System

    CERN Document Server

    Moles-Valls, R

    2008-01-01

    The ATLAS experiment is equipped with a tracking system for c harged particles built on two technologies: silicon and drift tube base detectors. These kind of detectors compose the ATLAS Inner Detector (ID). The Alignment of the ATLAS ID tracking s ystem requires the determination of almost 36000 degrees of freedom. From the tracking point o f view, the alignment parameters should be know to a few microns precision. This permits to att ain optimal measurements of the parameters of the charged particles trajectories, thus ena bling ATLAS to achieve its physics goals. The implementation of the alignment software, its framewor k and the data flow will be discussed. Special attention will be paid to the recent challenges wher e large scale computing simulation of the ATLAS detector has been performed, mimicking the ATLAS o peration, which is going to be very important for the LHC startup scenario. The alignment r esult for several challenges (real cosmic ray data taking and computing system commissioning) will be...

  7. Concepts and Plans towards fast large scale Monte Carlo production for the ATLAS Experiment

    CERN Document Server

    Chapman, J; Duehrssen, M; Elsing, M; Froidevaux, D; Harrington, R; Jansky, R; Langenberg, R; Mandrysch, R; Marshall, Z; Ritsch, E; Salzburger, A

    2014-01-01

    The huge success of the physics program of the ATLAS experiment at the Large Hadron Collider (LHC) during run I relies upon a great number of simulated Monte Carlo events. This Monte Carlo production takes the biggest part of the computing resources being in use by ATLAS as of now. In this document we describe the plans to overcome the computing resource limitations for large scale Monte Carlo production in the ATLAS Experiment for run II, and beyond. A number of fast detector simulation, digitization and reconstruction techniques and are being discussed, based upon a new flexible detector simulation framework. To optimally benefit from these developments, a redesigned ATLAS MC production chain is presented at the end of this document.

  8. Completion of the ATLAS control system upgrade

    International Nuclear Information System (INIS)

    Munson, F. H.

    1998-01-01

    In the fall of 1992 at the SNEAP(Symposium of North Eastern Accelerator Personnel) a project to up grade the ATLAS (Argonne Tandem Linear Accelerator System) control system was first reported. Not unlike the accelerator it services the control system will continue to evolve. However, the first of this year has marked the completion of this most recent upgrade project. Since the control system upgrade took place during a period when ATLAS was operating at a record number of hours, special techniques were necessary to enable the development of the new control system ''on line'' while still saving the needs of normal operations. This paper reviews the techniques used for upgrading the ATLAS control system while the system was in use. In addition a summary of the upgrade project and final configuration, as well as some of the features of the new control system is provided

  9. Design and Implementation of the ATLAS Detector Control System

    CERN Document Server

    Boterenbrood, H; Cook, J; Filimonov, V; Hallgren, B I; Heubers, W P J; Khomoutnikov, V; Ryabov, Yu; Varela, F

    2004-01-01

    The overall dimensions of the ATLAS experiment and its harsh environment, due to radiation and magnetic field, represent new challenges for the implementation of the Detector Control System. It supervises all hardware of the ATLAS detector, monitors the infrastructure of the experiment, and provides information exchange with the LHC accelerator. The system must allow for the operation of the different ATLAS sub-detectors in stand-alone mode, as required for calibration and debugging, as well as the coherent and integrated operation of all sub-detectors for physics data taking. For this reason, the Detector Control System is logically arranged to map the hierarchical organization of the ATLAS detector. Special requirements are placed onto the ATLAS Detector Control System because of the large number of distributed I/O channels and of the inaccessibility of the equipment during operation. Standardization is a crucial issue for the design and implementation of the control system because of the large variety of e...

  10. The ATLAS Distributed Analysis System

    CERN Document Server

    Legger, F; The ATLAS collaboration; Pacheco Pages, A; Stradling, A

    2013-01-01

    In the LHC operations era, analysis of the multi-petabyte ATLAS data sample by globally distributed physicists is a challenging task. To attain the required scale the ATLAS Computing Model was designed around the concept of grid computing, realized in the Worldwide LHC Computing Grid (WLCG), the largest distributed computational resource existing in the sciences. The ATLAS experiment currently stores over 140 PB of data and runs about 140,000 concurrent jobs continuously at WLCG sites. During the first run of the LHC, the ATLAS Distributed Analysis (DA) service has operated stably and scaled as planned. More than 1600 users submitted jobs in 2012, with 2 million or more analysis jobs per week, peaking at about a million jobs per day. The system dynamically distributes popular data to expedite processing and maximally utilize resources. The reliability of the DA service is high but steadily improving; grid sites are continually validated against a set of standard tests, and a dedicated team of expert shifters ...

  11. The ATLAS Distributed Analysis System

    CERN Document Server

    Legger, F; The ATLAS collaboration

    2014-01-01

    In the LHC operations era, analysis of the multi-petabyte ATLAS data sample by globally distributed physicists is a challenging task. To attain the required scale the ATLAS Computing Model was designed around the concept of grid computing, realized in the Worldwide LHC Computing Grid (WLCG), the largest distributed computational resource existing in the sciences. The ATLAS experiment currently stores over 140 PB of data and runs about 140,000 concurrent jobs continuously at WLCG sites. During the first run of the LHC, the ATLAS Distributed Analysis (DA) service has operated stably and scaled as planned. More than 1600 users submitted jobs in 2012, with 2 million or more analysis jobs per week, peaking at about a million jobs per day. The system dynamically distributes popular data to expedite processing and maximally utilize resources. The reliability of the DA service is high but steadily improving; grid sites are continually validated against a set of standard tests, and a dedicated team of expert shifters ...

  12. ATLAS: A High-cadence All-sky Survey System

    Science.gov (United States)

    Tonry, J. L.; Denneau, L.; Heinze, A. N.; Stalder, B.; Smith, K. W.; Smartt, S. J.; Stubbs, C. W.; Weiland, H. J.; Rest, A.

    2018-06-01

    Technology has advanced to the point that it is possible to image the entire sky every night and process the data in real time. The sky is hardly static: many interesting phenomena occur, including variable stationary objects such as stars or QSOs, transient stationary objects such as supernovae or M dwarf flares, and moving objects such as asteroids and the stars themselves. Funded by NASA, we have designed and built a sky survey system for the purpose of finding dangerous near-Earth asteroids (NEAs). This system, the “Asteroid Terrestrial-impact Last Alert System” (ATLAS), has been optimized to produce the best survey capability per unit cost, and therefore is an efficient and competitive system for finding potentially hazardous asteroids (PHAs) but also for tracking variables and finding transients. While carrying out its NASA mission, ATLAS now discovers more bright (m day cadence. ATLAS discovered the afterglow of a gamma-ray burst independent of the high energy trigger and has released a variable star catalog of 5 × 106 sources. This is the first of a series of articles describing ATLAS, devoted to the design and performance of the ATLAS system. Subsequent articles will describe in more detail the software, the survey strategy, ATLAS-derived NEA population statistics, transient detections, and the first data release of variable stars and transient light curves.

  13. ATLAS job monitoring in the Dashboard Framework

    CERN Document Server

    Sargsyan, L; The ATLAS collaboration; Campana, S; Karavakis, E; Kokoszkiewicz, L; Saiz, P; Schovancova, J; Tuckett, D

    2012-01-01

    Monitoring of the large-scale data processing of the ATLAS experiment includes monitoring of production and user analysis jobs. The Experiment Dashboard provides a common job monitoring solution, which is shared by ATLAS and CMS experiments. This includes an accounting portal as well as real-time monitoring. Dashboard job monitoring for ATLAS combines information from PanDA job processing database, Production system database and monitoring information from jobs submitted through GANGA to Workload Management System (WMS) or local batch systems. Usage of Dashboard-based job monitoring applications will decrease load on the PanDA database and overcome scale limitations in PanDA monitoring caused by the short job rotation cycle in the PanDA database. Aggregation of the task/job metrics from different sources provides complete view of job processing activity in ATLAS scope.

  14. ATLAS job monitoring in the Dashboard Framework

    International Nuclear Information System (INIS)

    Andreeva, J; Campana, S; Karavakis, E; Kokoszkiewicz, L; Saiz, P; Tuckett, D; Sargsyan, L; Schovancova, J

    2012-01-01

    Monitoring of the large-scale data processing of the ATLAS experiment includes monitoring of production and user analysis jobs. The Experiment Dashboard provides a common job monitoring solution, which is shared by ATLAS and CMS experiments. This includes an accounting portal as well as real-time monitoring. Dashboard job monitoring for ATLAS combines information from the PanDA job processing database, Production system database and monitoring information from jobs submitted through GANGA to Workload Management System (WMS) or local batch systems. Usage of Dashboard-based job monitoring applications will decrease load on the PanDA database and overcome scale limitations in PanDA monitoring caused by the short job rotation cycle in the PanDA database. Aggregation of the task/job metrics from different sources provides complete view of job processing activity in ATLAS scope.

  15. Improving Security in the ATLAS PanDA System

    International Nuclear Information System (INIS)

    Caballero, J; Maeno, T; Potekhin, M; Wenaus, T; Nilsson, P; Stewart, G

    2011-01-01

    The security challenges faced by users of the grid are considerably different to those faced in previous environments. The adoption of pilot jobs systems by LHC experiments has mitigated many of the problems associated with the inhomogeneities found on the grid and has greatly improved job reliability; however, pilot jobs systems themselves must then address many security issues, including the execution of multiple users' code under a common 'grid' identity. In this paper we describe the improvements and evolution of the security model in the ATLAS PanDA (Production and Distributed Analysis) system. We describe the security in the PanDA server which is in place to ensure that only authorized members of the VO are allowed to submit work into the system and that jobs are properly audited and monitored. We discuss the security in place between the pilot code itself and the PanDA server, ensuring that only properly authenticated workload is delivered to the pilot for execution. When the code to be executed is from a 'normal' ATLAS user, as opposed to the production system or other privileged actor, then the pilot may use an EGEE developed identity switching tool called gLExec. This changes the grid proxy available to the job and also switches the UNIX user identity to protect the privileges of the pilot code proxy. We describe the problems in using this system and how they are overcome. Finally, we discuss security drills which have been run using PanDA and show how these improved our operational security procedures.

  16. Virtualization of the ATLAS software environment on a shared HPC system

    CERN Document Server

    Gamel, Anton Josef; The ATLAS collaboration

    2017-01-01

    The shared HPC cluster NEMO at the University of Freiburg has been made available to local ATLAS users through the provisioning of virtual machines incorporating the ATLAS software environment analogously to a WLCG center. This concept allows to run both data analysis and production on the HPC host system which is connected to the existing Tier2/Tier3 infrastructure. Schedulers of the two clusters were integrated in a dynamic, on-demand way. An automatically generated, fully functional virtual machine image provides access to the local user environment. The performance in the virtualized environment is evaluated for typical High-Energy Physics applications.

  17. Quality assurance and functionality tests on electrical components during the ATLAS IBL production

    International Nuclear Information System (INIS)

    Jentzsch, J

    2013-01-01

    To improve performance of the ATLAS inner tracker, a fourth Pixel layer, called the Insertable B-layer (IBL), will be installed in 2014 on a new beam pipe. A new read out chip generation, FE-I4, has been developed and two different sensor designs, a rather conventional planar and a 3D design, have been flip chipped to these front ends. New staves holding new stave and module flex circuits have been developed as well. Therefore, a production QA test bench has been established to test all production staves before integration with the new beam pipe. This setup combines former ATLAS Pixel services and a new readout system, namely the RCE (Reconfigurable Cluster Element) system developed at SLAC. With this setup all production staves will be tested to ensure the installation of only those staves which fulfill the IBL criteria. Quality assurance measurements under cleanroom conditions, including temperature and humidity control, are performed on the individual components during the various production steps of the IBL, namely connectivity as well as electrical tests and signal probing on individual parts and assembled subsystems. The pre-assembly QC procedures, the capabilities of the stave qualification setup, and recent results from testing a prototype stave are presented and discussed.

  18. Quality assurance and functionality tests on electrical components during the ATLAS IBL production

    Science.gov (United States)

    Jentzsch, J.

    2013-02-01

    To improve performance of the ATLAS inner tracker, a fourth Pixel layer, called the Insertable B-layer (IBL), will be installed in 2014 on a new beam pipe. A new read out chip generation, FE-I4, has been developed and two different sensor designs, a rather conventional planar and a 3D design, have been flip chipped to these front ends. New staves holding new stave and module flex circuits have been developed as well. Therefore, a production QA test bench has been established to test all production staves before integration with the new beam pipe. This setup combines former ATLAS Pixel services and a new readout system, namely the RCE (Reconfigurable Cluster Element) system developed at SLAC. With this setup all production staves will be tested to ensure the installation of only those staves which fulfill the IBL criteria. Quality assurance measurements under cleanroom conditions, including temperature and humidity control, are performed on the individual components during the various production steps of the IBL, namely connectivity as well as electrical tests and signal probing on individual parts and assembled subsystems. The pre-assembly QC procedures, the capabilities of the stave qualification setup, and recent results from testing a prototype stave are presented and discussed.

  19. Experience from a pilot based system for ATLAS

    International Nuclear Information System (INIS)

    Nilsson, P

    2008-01-01

    The PanDA software provides a highly performing distributed production and distributed analysis system. It is the first system in the ATLAS experiment to use a pilot based late job delivery technique. This paper describes the architecture of the pilot system used in PanDA. Unique features have been implemented for high reliability automation in a distributed environment. Performance of PanDA is analyzed from one and a half years of experience of performing distributed computing on the Open Science Grid (OSG) infrastructure. Experience with pilot delivery mechanism using Condor-G, and a glide-in factory developed under OSG will be described

  20. Real time physics analysis with the ATLAS tau trigger system

    International Nuclear Information System (INIS)

    Casado Lechuga, M. P.

    2009-01-01

    The scope of the ATLAS tau trigger system at the LHC is most ambitious. It aims at reconstructing in real time, a matter of seconds, a detailed picture of the high energy proton proton collisions at the LHC. Such system is mandatory in order to select efficiently data needed for discovery of new physics in a proton proton collision environment where the rates of jets observed in the detector are high and the tau identification is difficult. New physics scenarios targeted specifically by the the ATLAS tau trigger system are Standard Model or Supersymmetric Higgs production, and production of new exotic resonances. This contribution will detail how the analysis techniques developed offline for efficient data analysis have been implemented in the algorithms which run online at the trigger. In particular, the focus will be on how to satisfy the requirements imposed by the physics goals while addressing the limitations from the overall event rate and latency allowed. The prospects for early running during the first LHC collisions and trigger evolution from first collisions to stable running will be also summarized, following change of trigger goals from commissioning of detector to measurement of Standard Model physics and discoveries. (author)

  1. Upgrading the ATLAS control system

    International Nuclear Information System (INIS)

    Munson, F.H.; Ferraretto, M.

    1993-01-01

    Heavy-ion accelerators are tools used in the research of nuclear and atomic physics. The ATLAS facility at the Argonne National Laboratory is one such tool. The ATLAS control system serves as the primary operator interface to the accelerator. A project to upgrade the control system is presently in progress. Since this is an upgrade project and not a new installation, it was imperative that the development work proceed without interference to normal operations. An additional criteria for the development work was that the writing of additional ''in-house'' software should be kept to a minimum. This paper briefly describes the control system being upgraded, and explains some of the reasons for the decision to upgrade the control system. Design considerations and goals for the new system are described, and the present status of the upgrade is discussed

  2. Advanced Technology Lifecycle Analysis System (ATLAS)

    Science.gov (United States)

    O'Neil, Daniel A.; Mankins, John C.

    2004-01-01

    Developing credible mass and cost estimates for space exploration and development architectures require multidisciplinary analysis based on physics calculations, and parametric estimates derived from historical systems. Within the National Aeronautics and Space Administration (NASA), concurrent engineering environment (CEE) activities integrate discipline oriented analysis tools through a computer network and accumulate the results of a multidisciplinary analysis team via a centralized database or spreadsheet Each minute of a design and analysis study within a concurrent engineering environment is expensive due the size of the team and supporting equipment The Advanced Technology Lifecycle Analysis System (ATLAS) reduces the cost of architecture analysis by capturing the knowledge of discipline experts into system oriented spreadsheet models. A framework with a user interface presents a library of system models to an architecture analyst. The analyst selects models of launchers, in-space transportation systems, and excursion vehicles, as well as space and surface infrastructure such as propellant depots, habitats, and solar power satellites. After assembling the architecture from the selected models, the analyst can create a campaign comprised of missions spanning several years. The ATLAS controller passes analyst specified parameters to the models and data among the models. An integrator workbook calls a history based parametric analysis cost model to determine the costs. Also, the integrator estimates the flight rates, launched masses, and architecture benefits over the years of the campaign. An accumulator workbook presents the analytical results in a series of bar graphs. In no way does ATLAS compete with a CEE; instead, ATLAS complements a CEE by ensuring that the time of the experts is well spent Using ATLAS, an architecture analyst can perform technology sensitivity analysis, study many scenarios, and see the impact of design decisions. When the analyst is

  3. Heavy Flavour Production and Properties at CMS and ATLAS

    CERN Document Server

    Barton, Adam Edward; The ATLAS collaboration

    2018-01-01

    Measurements of heavy flavour properties and production are an important part of the physics program of the ATLAS and CMS experiments at LHC. They can potentially expose physics beyond the standard model, constrain supersymmetry and advance hadron spectroscopy and test QCD. In the past years, the two collaborations have published results in several different fields, such as rare decays, searches for new states, CP and P violation and quarkonia polarisation. In this note, some of the most recent results from ATLAS and CMS are summarised.

  4. Heavy Flavour Production and Properties at ATLAS and CMS

    CERN Document Server

    Barton, Adam Edward; The ATLAS collaboration

    2017-01-01

    Measurements of heavy flavour properties and production are an important part of the physics program of the ATLAS and CMS experiments at LHC. They can potentially expose physics beyond the standard model, constrain supersymmetry and advance hadron spectroscopy and test QCD. In the past years, the two collaborations have published results in several different fields, such as rare decays, searches for new states, CP and P violation and quarkonia polarization. In this note, some of the most recent results from ATLAS and CMS are summarized

  5. NODC Standard Product: World Ocean Atlas 1994 (11 disc set) (NODC Accession 0098057)

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — Observed and standard level profile data (along with quality control flags) used in the production of these atlases were made available in a World Ocean Atlas 1994...

  6. Alignment of the ATLAS Inner Detector Tracking System

    CERN Document Server

    Lacuesta, V; The ATLAS collaboration

    2010-01-01

    ATLAS is a multipurpose experiment that records the LHC collisions. To reconstruct trajectories of charged particles produced in these collisions, ATLAS tracking system is equipped with silicon planar sensors and drift‐tube based detectors. They constitute the ATLAS Inner Detector. In order to achieve its scientific goals, the alignment of the ATLAS tracking system requires the determine accurately its almost 36000 degrees of freedom. Thus the demanded precision for the alignment of the silicon sensors is below 10 micrometers. This implies to use a large sample of high momentum and isolated charge particle tracks. The high level trigger selects those tracks online. Then the raw data with the hits information of the triggered tracks is stored in a calibration stream. Tracks from cosmic trigger during empty LHC bunches are also used as input for the alignment. The implementation of the track based alignment within the ATLAS software framework unifies different alignment approaches and allows the alignment of ...

  7. Recent ATLAS Results in Charmonium Production

    CERN Document Server

    Reznicek, Pavel; The ATLAS collaboration

    2018-01-01

    Production of heavy flavoured hadrons can serve as a probe of QCD processes and, in the case of production in heavy ion collisions, also as a probe of the properties of deconfined quark-gluon plasma. This paper describes the latest results in the production of charmonium at the ATLAS experiment at the Large Hadron Collider. The presented analyses include production of $J/\\psi$ and $\\psi$(2S) in $pp$ collisions and $\\sqrt{s} =$7, 8 and 13 TeV, and production of $J/\\psi$, $\\psi$(2S) and $\\Upsilon$(nS) in $pp$ and $p$+Pb collisions at $\\sqrt{s} =$5.02 TeV. In both the analyses prompt and non-prompt production is disentangled. In the latter analysis production suppression in $p$+Pb with respect to $pp$ collisions is measured and emphasis is put on the collision centrality dependence.

  8. LASER monitoring system for the ATLAS Tile Calorimeter

    International Nuclear Information System (INIS)

    Viret, S.

    2010-01-01

    The ATLAS detector at the Large Hadron Collider (LHC) at CERN uses a scintillator-iron technique for its hadronic Tile Calorimeter (TileCal). Scintillating light is readout via 9852 photomultiplier tubes (PMTs). Calibration and monitoring of these PMTs are made using a LASER based system. Short light pulses are sent simultaneously into all the TileCal photomultiplier's tubes (PMTs) during ATLAS physics runs, thus providing essential information for ATLAS data quality and monitoring analyses. The experimental setup developed for this purpose is described as well as preliminary results obtained during ATLAS commissioning phase in 2008.

  9. Performance of the ATLAS Trigger System in 2010

    CERN Document Server

    Aad, Georges; Abdallah, Jalal; Abdelalim, Ahmed Ali; Abdesselam, Abdelouahab; Abdinov, Ovsat; Abi, Babak; Abolins, Maris; Abramowicz, Halina; Abreu, Henso; Acerbi, Emilio; Acharya, Bobby Samir; Adams, David; Addy, Tetteh; Adelman, Jahred; Aderholz, Michael; Adomeit, Stefanie; Adragna, Paolo; Adye, Tim; Aefsky, Scott; Aguilar-Saavedra, Juan Antonio; Aharrouche, Mohamed; Ahlen, Steven; Ahles, Florian; Ahmad, Ashfaq; Ahsan, Mahsana; Aielli, Giulio; Akdogan, Taylan; Åkesson, Torsten Paul Ake; Akimoto, Ginga; Akimov, Andrei; Akiyama, Kunihiro; Alam, Mohammad; Alam, Muhammad Aftab; Albrand, Solveig; Aleksa, Martin; Aleksandrov, Igor; Alessandria, Franco; Alexa, Calin; Alexander, Gideon; Alexandre, Gauthier; Alexopoulos, Theodoros; Alhroob, Muhammad; Aliev, Malik; Alimonti, Gianluca; Alison, John; Aliyev, Magsud; Allport, Phillip; Allwood-Spiers, Sarah; Almond, John; Aloisio, Alberto; Alon, Raz; Alonso, Alejandro; Alviggi, Mariagrazia; Amako, Katsuya; Amaral, Pedro; Amelung, Christoph; Ammosov, Vladimir; Amorim, Antonio; Amorós, Gabriel; Amram, Nir; Anastopoulos, Christos; Andari, Nansi; Andeen, Timothy; Anders, Christoph Falk; Anderson, Kelby; Andreazza, Attilio; Andrei, George Victor; Andrieux, Marie-Laure; Anduaga, Xabier; Angerami, Aaron; Anghinolfi, Francis; Anjos, Nuno; Annovi, Alberto; Antonaki, Ariadni; Antonelli, Mario; Antonelli, Stefano; Antonov, Alexey; Antos, Jaroslav; Anulli, Fabio; Aoun, Sahar; Aperio Bella, Ludovica; Apolle, Rudi; Arabidze, Giorgi; Aracena, Ignacio; Arai, Yasuo; Arce, Ayana; Archambault, John-Paul; Arfaoui, Samir; Arguin, Jean-Francois; Arik, Engin; Arik, Metin; Armbruster, Aaron James; Arnaez, Olivier; Arnault, Christian; Artamonov, Andrei; Artoni, Giacomo; Arutinov, David; Asai, Shoji; Asfandiyarov, Ruslan; Ask, Stefan; Åsman, Barbro; Asquith, Lily; Assamagan, Ketevi; Astbury, Alan; Astvatsatourov, Anatoli; Atoian, Grigor; Aubert, Bernard; Auerbach, Benjamin; Auge, Etienne; Augsten, Kamil; Aurousseau, Mathieu; Austin, Nicholas; Avolio, Giuseppe; Avramidou, Rachel Maria; Axen, David; Ay, Cano; Azuelos, Georges; Azuma, Yuya; Baak, Max; Baccaglioni, Giuseppe; Bacci, Cesare; Bach, Andre; Bachacou, Henri; Bachas, Konstantinos; Bachy, Gerard; Backes, Moritz; Backhaus, Malte; Badescu, Elisabeta; Bagnaia, Paolo; Bahinipati, Seema; Bai, Yu; Bailey, David; Bain, Travis; Baines, John; Baker, Oliver Keith; Baker, Mark; Baker, Sarah; Baltasar Dos Santos Pedrosa, Fernando; Banas, Elzbieta; Banerjee, Piyali; Banerjee, Swagato; Banfi, Danilo; Bangert, Andrea Michelle; Bansal, Vikas; Bansil, Hardeep Singh; Barak, Liron; Baranov, Sergei; Barashkou, Andrei; Barbaro Galtieri, Angela; Barber, Tom; Barberio, Elisabetta Luigia; Barberis, Dario; Barbero, Marlon; Bardin, Dmitri; Barillari, Teresa; Barisonzi, Marcello; Barklow, Timothy; Barlow, Nick; Barnett, Bruce; Barnett, Michael; Baroncelli, Antonio; Barone, Gaetano; Barr, Alan; Barreiro, Fernando; Barreiro Guimarães da Costa, João; Barrillon, Pierre; Bartoldus, Rainer; Barton, Adam Edward; Bartsch, Detlef; Bartsch, Valeria; Bates, Richard; Batkova, Lucia; Batley, Richard; Battaglia, Andreas; Battistin, Michele; Battistoni, Giuseppe; Bauer, Florian; Bawa, Harinder Singh; Beare, Brian; Beau, Tristan; Beauchemin, Pierre-Hugues; Beccherle, Roberto; Bechtle, Philip; Beck, Hans Peter; Beckingham, Matthew; Becks, Karl-Heinz; Beddall, Andrew; Beddall, Ayda; Bedikian, Sourpouhi; Bednyakov, Vadim; Bee, Christopher; Begel, Michael; Behar Harpaz, Silvia; Behera, Prafulla; Beimforde, Michael; Belanger-Champagne, Camille; Bell, Paul; Bell, William; Bella, Gideon; Bellagamba, Lorenzo; Bellina, Francesco; Bellomo, Massimiliano; Belloni, Alberto; Beloborodova, Olga; Belotskiy, Konstantin; Beltramello, Olga; Ben Ami, Sagi; Benary, Odette; Benchekroun, Driss; Benchouk, Chafik; Bendel, Markus; Benedict, Brian Hugues; Benekos, Nektarios; Benhammou, Yan; Benjamin, Douglas; Benoit, Mathieu; Bensinger, James; Benslama, Kamal; Bentvelsen, Stan; Berge, David; Bergeaas Kuutmann, Elin; Berger, Nicolas; Berghaus, Frank; Berglund, Elina; Beringer, Jürg; Bernardet, Karim; Bernat, Pauline; Bernhard, Ralf; Bernius, Catrin; Berry, Tracey; Bertin, Antonio; Bertinelli, Francesco; Bertolucci, Federico; Besana, Maria Ilaria; Besson, Nathalie; Bethke, Siegfried; Bhimji, Wahid; Bianchi, Riccardo-Maria; Bianco, Michele; Biebel, Otmar; Bieniek, Stephen Paul; Biesiada, Jed; Biglietti, Michela; Bilokon, Halina; Bindi, Marcello; Binet, Sebastien; Bingul, Ahmet; Bini, Cesare; Biscarat, Catherine; Bitenc, Urban; Black, Kevin; Blair, Robert; Blanchard, Jean-Baptiste; Blanchot, Georges; Blazek, Tomas; Blocker, Craig; Blocki, Jacek; Blondel, Alain; Blum, Walter; Blumenschein, Ulrike; Bobbink, Gerjan; Bobrovnikov, Victor; Bocchetta, Simona Serena; Bocci, Andrea; Boddy, Christopher Richard; Boehler, Michael; Boek, Jennifer; Boelaert, Nele; Böser, Sebastian; Bogaerts, Joannes Andreas; Bogdanchikov, Alexander; Bogouch, Andrei; Bohm, Christian; Boisvert, Veronique; Bold, Tomasz; Boldea, Venera; Bolnet, Nayanka Myriam; Bona, Marcella; Bondarenko, Valery; Boonekamp, Maarten; Boorman, Gary; Booth, Chris; Bordoni, Stefania; Borer, Claudia; Borisov, Anatoly; Borissov, Guennadi; Borjanovic, Iris; Borroni, Sara; Bos, Kors; Boscherini, Davide; Bosman, Martine; Boterenbrood, Hendrik; Botterill, David; Bouchami, Jihene; Boudreau, Joseph; Bouhova-Thacker, Evelina Vassileva; Boulahouache, Chaouki; Bourdarios, Claire; Bousson, Nicolas; Boveia, Antonio; Boyd, James; Boyko, Igor; Bozhko, Nikolay; Bozovic-Jelisavcic, Ivanka; Bracinik, Juraj; Braem, André; Branchini, Paolo; Brandenburg, George; Brandt, Andrew; Brandt, Gerhard; Brandt, Oleg; Bratzler, Uwe; Brau, Benjamin; Brau, James; Braun, Helmut; Brelier, Bertrand; Bremer, Johan; Brenner, Richard; Bressler, Shikma; Breton, Dominique; Britton, Dave; Brochu, Frederic; Brock, Ian; Brock, Raymond; Brodbeck, Timothy; Brodet, Eyal; Broggi, Francesco; Bromberg, Carl; Brooijmans, Gustaaf; Brooks, William; Brown, Gareth; Brown, Heather; Brubaker, Erik; Bruckman de Renstrom, Pawel; Bruncko, Dusan; Bruneliere, Renaud; Brunet, Sylvie; Bruni, Alessia; Bruni, Graziano; Bruschi, Marco; Buanes, Trygve; Bucci, Francesca; Buchanan, James; Buchanan, Norman; Buchholz, Peter; Buckingham, Ryan; Buckley, Andrew; Buda, Stelian Ioan; Budagov, Ioulian; Budick, Burton; Büscher, Volker; Bugge, Lars; Buira-Clark, Daniel; Bulekov, Oleg; Bunse, Moritz; Buran, Torleiv; Burckhart, Helfried; Burdin, Sergey; Burgess, Thomas; Burke, Stephen; Busato, Emmanuel; Bussey, Peter; Buszello, Claus-Peter; Butin, François; Butler, Bart; Butler, John; Buttar, Craig; Butterworth, Jonathan; Buttinger, William; Byatt, Tom; Cabrera Urbán, Susana; Caforio, Davide; Cakir, Orhan; Calafiura, Paolo; Calderini, Giovanni; Calfayan, Philippe; Calkins, Robert; Caloba, Luiz; Caloi, Rita; Calvet, David; Calvet, Samuel; Camacho Toro, Reina; Camard, Arnaud; Camarri, Paolo; Cambiaghi, Mario; Cameron, David; Cammin, Jochen; Campana, Simone; Campanelli, Mario; Canale, Vincenzo; Canelli, Florencia; Canepa, Anadi; Cantero, Josu; Capasso, Luciano; Capeans Garrido, Maria Del Mar; Caprini, Irinel; Caprini, Mihai; Capriotti, Daniele; Capua, Marcella; Caputo, Regina; Caramarcu, Costin; Cardarelli, Roberto; Carli, Tancredi; Carlino, Gianpaolo; Carminati, Leonardo; Caron, Bryan; Caron, Sascha; Carrillo Montoya, German D; Carter, Antony; Carter, Janet; Carvalho, João; Casadei, Diego; Casado, Maria Pilar; Cascella, Michele; Caso, Carlo; Castaneda Hernandez, Alfredo Martin; Castaneda-Miranda, Elizabeth; Castillo Gimenez, Victoria; Castro, Nuno Filipe; Cataldi, Gabriella; Cataneo, Fernando; Catinaccio, Andrea; Catmore, James; Cattai, Ariella; Cattani, Giordano; Caughron, Seth; Cauz, Diego; Cavalleri, Pietro; Cavalli, Donatella; Cavalli-Sforza, Matteo; Cavasinni, Vincenzo; Cazzato, Antonio; Ceradini, Filippo; Santiago Cerqueira, Augusto; Cerri, Alessandro; Cerrito, Lucio; Cerutti, Fabio; Cetin, Serkant Ali; Cevenini, Francesco; Chafaq, Aziz; Chakraborty, Dhiman; Chan, Kevin; Chapleau, Bertrand; Chapman, John Derek; Chapman, John Wehrley; Chareyre, Eve; Charlton, Dave; Chavda, Vikash; Cheatham, Susan; Chekanov, Sergei; Chekulaev, Sergey; Chelkov, Gueorgui; Chelstowska, Magda Anna; Chen, Chunhui; Chen, Hucheng; Chen, Li; Chen, Shenjian; Chen, Tingyang; Chen, Xin; Cheng, Shaochen; Cheplakov, Alexander; Chepurnov, Vladimir; Cherkaoui El Moursli, Rajaa; Chernyatin, Valeriy; Cheu, Elliott; Cheung, Sing-Leung; Chevalier, Laurent; Chiefari, Giovanni; Chikovani, Leila; Childers, John Taylor; Chilingarov, Alexandre; Chiodini, Gabriele; Chislett, Rebecca Thalatta; Chizhov, Mihail; Choudalakis, Georgios; Chouridou, Sofia; Christidi, Illectra-Athanasia; Christov, Asen; Chromek-Burckhart, Doris; Chu, Ming-Lee; Chudoba, Jiri; Ciapetti, Guido; Ciba, Krzysztof; Ciftci, Abbas Kenan; Ciftci, Rena; Cinca, Diane; Cindro, Vladimir; Ciobotaru, Matei Dan; Ciocca, Claudia; Ciocio, Alessandra; Cirilli, Manuela; Ciubancan, Mihai; Clark, Allan G; Clark, Philip; Cleland, Bill; Clemens, Jean-Claude; Clement, Benoit; Clement, Christophe; Clifft, Roger; Coadou, Yann; Cobal, Marina; Coccaro, Andrea; Cochran, James H; Coe, Paul; Cogan, Joshua Godfrey; Coggeshall, James; Cogneras, Eric; Cojocaru, Claudiu; Colas, Jacques; Colijn, Auke-Pieter; Collard, Caroline; Collins, Neil; Collins-Tooth, Christopher; Collot, Johann; Colon, German; Conde Muiño, Patricia; Coniavitis, Elias; Conidi, Maria Chiara; Consonni, Michele; Consorti, Valerio; Constantinescu, Serban; Conta, Claudio; Conventi, Francesco; Cook, James; Cooke, Mark; Cooper, Ben; Cooper-Sarkar, Amanda; Cooper-Smith, Neil; Copic, Katherine; Cornelissen, Thijs; Corradi, Massimo; Corriveau, Francois; Cortes-Gonzalez, Arely; Cortiana, Giorgio; Costa, Giuseppe; Costa, María José; Costanzo, Davide; Costin, Tudor; Côté, David; Coura Torres, Rodrigo; Courneyea, Lorraine; Cowan, Glen; Cowden, Christopher; Cox, Brian; Cranmer, Kyle; Crescioli, Francesco; Cristinziani, Markus; Crosetti, Giovanni; Crupi, Roberto; Crépé-Renaudin, Sabine; Cuciuc, Constantin-Mihai; Cuenca Almenar, Cristóbal; Cuhadar Donszelmann, Tulay; Cuneo, Stefano; Curatolo, Maria; Curtis, Chris; Cwetanski, Peter; Czirr, Hendrik; Czyczula, Zofia; D'Auria, Saverio; D'Onofrio, Monica; D'Orazio, Alessia; Da Rocha Gesualdi Mello, Aline; Da Silva, Paulo Vitor; Da Via, Cinzia; Dabrowski, Wladyslaw; Dahlhoff, Andrea; Dai, Tiesheng; Dallapiccola, Carlo; Dam, Mogens; Dameri, Mauro; Damiani, Daniel; Danielsson, Hans Olof; Dannheim, Dominik; Dao, Valerio; Darbo, Giovanni; Darlea, Georgiana Lavinia; Daum, Cornelis; Dauvergne, Jean-Pierre; Davey, Will; Davidek, Tomas; Davidson, Nadia; Davidson, Ruth; Davies, Eleanor; Davies, Merlin; Davison, Adam; Davygora, Yuriy; Dawe, Edmund; Dawson, Ian; Dawson, John; Daya, Rozmin; De, Kaushik; de Asmundis, Riccardo; De Castro, Stefano; De Castro Faria Salgado, Pedro; De Cecco, Sandro; de Graat, Julien; De Groot, Nicolo; de Jong, Paul; De La Taille, Christophe; De la Torre, Hector; De Lotto, Barbara; De Mora, Lee; De Nooij, Lucie; De Oliveira Branco, Miguel; De Pedis, Daniele; de Saintignon, Paul; De Salvo, Alessandro; De Sanctis, Umberto; De Santo, Antonella; De Vivie De Regie, Jean-Baptiste; Dean, Simon; Dedovich, Dmitri; Degenhardt, James; Dehchar, Mohamed; Deile, Mario; Del Papa, Carlo; Del Peso, Jose; Del Prete, Tarcisio; Dell'Acqua, Andrea; Dell'Asta, Lidia; Della Pietra, Massimo; della Volpe, Domenico; Delmastro, Marco; Delpierre, Pierre; Delruelle, Nicolas; Delsart, Pierre-Antoine; Deluca, Carolina; Demers, Sarah; Demichev, Mikhail; Demirkoz, Bilge; Deng, Jianrong; Denisov, Sergey; Derendarz, Dominik; Derkaoui, Jamal Eddine; Derue, Frederic; Dervan, Paul; Desch, Klaus Kurt; Devetak, Erik; Deviveiros, Pier-Olivier; Dewhurst, Alastair; DeWilde, Burton; Dhaliwal, Saminder; Dhullipudi, Ramasudhakar; Di Ciaccio, Anna; Di Ciaccio, Lucia; Di Girolamo, Alessandro; Di Girolamo, Beniamino; Di Luise, Silvestro; Di Mattia, Alessandro; Di Micco, Biagio; Di Nardo, Roberto; Di Simone, Andrea; Di Sipio, Riccardo; Diaz, Marco Aurelio; Diblen, Faruk; Diehl, Edward; Dieli, Michele Vincenzo; Dietl, Hans; Dietrich, Janet; Dietzsch, Thorsten; Diglio, Sara; Dindar Yagci, Kamile; Dingfelder, Jochen; Dionisi, Carlo; Dita, Petre; Dita, Sanda; Dittus, Fridolin; Djama, Fares; Djilkibaev, Rashid; Djobava, Tamar; Barros do Vale, Maria Aline; Do Valle Wemans, André; Doan, Thi Kieu Oanh; Dobbs, Matt; Dobinson, Robert; Dobos, Daniel; Dobson, Ellie; Dobson, Marc; Dodd, Jeremy; Dogan, Ozgen Berkol; Doglioni, Caterina; Doherty, Tom; Doi, Yoshikuni; Dolejsi, Jiri; Dolenc, Irena; Dolezal, Zdenek; Dolgoshein, Boris; Dohmae, Takeshi; Donadelli, Marisilvia; Donega, Mauro; Donini, Julien; Dopke, Jens; Doria, Alessandra; Dos Anjos, Andre; Dosil, Mireia; Dotti, Andrea; Dova, Maria-Teresa; Dowell, John; Doxiadis, Alexander; Doyle, Tony; Drasal, Zbynek; Drees, Jürgen; Dressnandt, Nandor; Drevermann, Hans; Driouichi, Chafik; Dris, Manolis; Dubbert, Jörg; Dubbs, Tim; Dube, Sourabh; Duchovni, Ehud; Duckeck, Guenter; Dudarev, Alexey; Dudziak, Fanny; Dührssen, Michael; Duerdoth, Ian; Duflot, Laurent; Dufour, Marc-Andre; Dunford, Monica; Duran Yildiz, Hatice; Duxfield, Robert; Dwuznik, Michal; Dydak, Friedrich; Dzahini, Daniel; Düren, Michael; Ebenstein, William; Ebke, Johannes; Eckert, Simon; Eckweiler, Sebastian; Edmonds, Keith; Edwards, Clive; Edwards, Nicholas Charles; Ehrenfeld, Wolfgang; Ehrich, Thies; Eifert, Till; Eigen, Gerald; Einsweiler, Kevin; Eisenhandler, Eric; Ekelof, Tord; El Kacimi, Mohamed; Ellert, Mattias; Elles, Sabine; Ellinghaus, Frank; Ellis, Katherine; Ellis, Nicolas; Elmsheuser, Johannes; Elsing, Markus; Ely, Robert; Emeliyanov, Dmitry; Engelmann, Roderich; Engl, Albert; Epp, Brigitte; Eppig, Andrew; Erdmann, Johannes; Ereditato, Antonio; Eriksson, Daniel; Ernst, Jesse; Ernst, Michael; Ernwein, Jean; Errede, Deborah; Errede, Steven; Ertel, Eugen; Escalier, Marc; Escobar, Carlos; Espinal Curull, Xavier; Esposito, Bellisario; Etienne, Francois; Etienvre, Anne-Isabelle; Etzion, Erez; Evangelakou, Despoina; Evans, Hal; Fabbri, Laura; Fabre, Caroline; Fakhrutdinov, Rinat; Falciano, Speranza; Falou, Alain; Fang, Yaquan; Fanti, Marcello; Farbin, Amir; Farilla, Addolorata; Farley, Jason; Farooque, Trisha; Farrington, Sinead; Farthouat, Philippe; Fassnacht, Patrick; Fassouliotis, Dimitrios; Fatholahzadeh, Baharak; Favareto, Andrea; Fayard, Louis; Fazio, Salvatore; Febbraro, Renato; Federic, Pavol; Fedin, Oleg; Fedorko, Ivan; Fedorko, Woiciech; Fehling-Kaschek, Mirjam; Feligioni, Lorenzo; Fellmann, Denis; Felzmann, Ulrich; Feng, Cunfeng; Feng, Eric; Fenyuk, Alexander; Ferencei, Jozef; Ferland, Jonathan; Fernando, Waruna; Ferrag, Samir; Ferrando, James; Ferrara, Valentina; Ferrari, Arnaud; Ferrari, Pamela; Ferrari, Roberto; Ferrer, Antonio; Ferrer, Maria Lorenza; Ferrere, Didier; Ferretti, Claudio; Ferretto Parodi, Andrea; Fiascaris, Maria; Fiedler, Frank; Filipčič, Andrej; Filippas, Anastasios; Filthaut, Frank; Fincke-Keeler, Margret; Fiolhais, Miguel; Fiorini, Luca; Firan, Ana; Fischer, Gordon; Fischer, Peter; Fisher, Matthew; Fisher, Steve; Flechl, Martin; Fleck, Ivor; Fleckner, Johanna; Fleischmann, Philipp; Fleischmann, Sebastian; Flick, Tobias; Flores Castillo, Luis; Flowerdew, Michael; Föhlisch, Florian; Fokitis, Manolis; Fonseca Martin, Teresa; Forbush, David Alan; Formica, Andrea; Forti, Alessandra; Fortin, Dominique; Foster, Joe; Fournier, Daniel; Foussat, Arnaud; Fowler, Andrew; Fowler, Ken; Fox, Harald; Francavilla, Paolo; Franchino, Silvia; Francis, David; Frank, Tal; Franklin, Melissa; Franz, Sebastien; Fraternali, Marco; Fratina, Sasa; French, Sky; Froeschl, Robert; Froidevaux, Daniel; Frost, James; Fukunaga, Chikara; Fullana Torregrosa, Esteban; Fuster, Juan; Gabaldon, Carolina; Gabizon, Ofir; Gadfort, Thomas; Gadomski, Szymon; Gagliardi, Guido; Gagnon, Pauline; Galea, Cristina; Gallas, Elizabeth; Gallas, Manuel; Gallo, Valentina Santina; Gallop, Bruce; Gallus, Petr; Galyaev, Eugene; Gan, KK; Gao, Yongsheng; Gapienko, Vladimir; Gaponenko, Andrei; Garberson, Ford; Garcia-Sciveres, Maurice; García, Carmen; García Navarro, José Enrique; Gardner, Robert; Garelli, Nicoletta; Garitaonandia, Hegoi; Garonne, Vincent; Garvey, John; Gatti, Claudio; Gaudio, Gabriella; Gaumer, Olivier; Gaur, Bakul; Gauthier, Lea; Gavrilenko, Igor; Gay, Colin; Gaycken, Goetz; Gayde, Jean-Christophe; Gazis, Evangelos; Ge, Peng; Gee, Norman; Geerts, Daniël Alphonsus Adrianus; Geich-Gimbel, Christoph; Gellerstedt, Karl; Gemme, Claudia; Gemmell, Alistair; Genest, Marie-Hélène; Gentile, Simonetta; George, Matthias; George, Simon; Gerlach, Peter; Gershon, Avi; Geweniger, Christoph; Ghazlane, Hamid; Ghez, Philippe; Ghodbane, Nabil; Giacobbe, Benedetto; Giagu, Stefano; Giakoumopoulou, Victoria; Giangiobbe, Vincent; Gianotti, Fabiola; Gibbard, Bruce; Gibson, Adam; Gibson, Stephen; Gilbert, Laura; Gilchriese, Murdock; Gilewsky, Valentin; Gillberg, Dag; Gillman, Tony; Gingrich, Douglas; Ginzburg, Jonatan; Giokaris, Nikos; Giordano, Raffaele; Giorgi, Francesco Michelangelo; Giovannini, Paola; Giraud, Pierre-Francois; Giugni, Danilo; Giunta, Michele; Giusti, Paolo; Gjelsten, Børge Kile; Gladilin, Leonid; Glasman, Claudia; Glatzer, Julian; Glazov, Alexandre; Glitza, Karl-Walter; Glonti, George; Godfrey, Jennifer; Godlewski, Jan; Goebel, Martin; Göpfert, Thomas; Goeringer, Christian; Gössling, Claus; Göttfert, Tobias; Goldfarb, Steven; Goldin, Daniel; Golling, Tobias; Golovnia, Serguei; Gomes, Agostinho; Gomez Fajardo, Luz Stella; Gonçalo, Ricardo; Goncalves Pinto Firmino Da Costa, Joao; Gonella, Laura; Gonidec, Allain; Gonzalez, Saul; González de la Hoz, Santiago; Gonzalez Silva, Laura; Gonzalez-Sevilla, Sergio; Goodson, Jeremiah Jet; Goossens, Luc; Gorbounov, Petr Andreevich; Gordon, Howard; Gorelov, Igor; Gorfine, Grant; Gorini, Benedetto; Gorini, Edoardo; Gorišek, Andrej; Gornicki, Edward; Gorokhov, Serguei; Goryachev, Vladimir; Gosdzik, Bjoern; Gosselink, Martijn; Gostkin, Mikhail Ivanovitch; Gouanère, Michel; Gough Eschrich, Ivo; Gouighri, Mohamed; Goujdami, Driss; Goulette, Marc Phillippe; Goussiou, Anna; Goy, Corinne; Grabowska-Bold, Iwona; Grabski, Varlen; Grafström, Per; Grah, Christian; Grahn, Karl-Johan; Grancagnolo, Francesco; Grancagnolo, Sergio; Grassi, Valerio; Gratchev, Vadim; Grau, Nathan; Gray, Heather; Gray, Julia Ann; Graziani, Enrico; Grebenyuk, Oleg; Greenfield, Debbie; Greenshaw, Timothy; Greenwood, Zeno Dixon; Gregor, Ingrid-Maria; Grenier, Philippe; Griesmayer, Erich; Griffiths, Justin; Grigalashvili, Nugzar; Grillo, Alexander; Grinstein, Sebastian; Grishkevich, Yaroslav; Grivaz, Jean-Francois; Grognuz, Joel; Groh, Manfred; Gross, Eilam; Grosse-Knetter, Joern; Groth-Jensen, Jacob; Grybel, Kai; Guarino, Victor; Guest, Daniel; Guicheney, Christophe; Guida, Angelo; Guillemin, Thibault; Guindon, Stefan; Guler, Hulya; Gunther, Jaroslav; Guo, Bin; Guo, Jun; Gupta, Ambreesh; Gusakov, Yury; Gushchin, Vladimir; Gutierrez, Andrea; Gutierrez, Phillip; Guttman, Nir; Gutzwiller, Olivier; Guyot, Claude; Gwenlan, Claire; Gwilliam, Carl; Haas, Andy; Haas, Stefan; Haber, Carl; Hackenburg, Robert; Hadavand, Haleh Khani; Hadley, David; Haefner, Petra; Hahn, Ferdinand; Haider, Stefan; Hajduk, Zbigniew; Hakobyan, Hrachya; Haller, Johannes; Hamacher, Klaus; Hamal, Petr; Hamilton, Andrew; Hamilton, Samuel; Han, Hongguang; Han, Liang; Hanagaki, Kazunori; Hance, Michael; Handel, Carsten; Hanke, Paul; Hansen, John Renner; Hansen, Jørgen Beck; Hansen, Jorn Dines; Hansen, Peter Henrik; Hansson, Per; Hara, Kazuhiko; Hare, Gabriel; Harenberg, Torsten; Harkusha, Siarhei; Harper, Devin; Harrington, Robert; Harris, Orin; Harrison, Karl; Hartert, Jochen; Hartjes, Fred; Haruyama, Tomiyoshi; Harvey, Alex; Hasegawa, Satoshi; Hasegawa, Yoji; Hassani, Samira; Hatch, Mark; Hauff, Dieter; Haug, Sigve; Hauschild, Michael; Hauser, Reiner; Havranek, Miroslav; Hawes, Brian; Hawkes, Christopher; Hawkings, Richard John; Hawkins, Donovan; Hayakawa, Takashi; Hayden, Daniel; Hayward, Helen; Haywood, Stephen; Hazen, Eric; He, Mao; Head, Simon; Hedberg, Vincent; Heelan, Louise; Heim, Sarah; Heine, Kristin; Heinemann, Beate; Heisterkamp, Simon; Helary, Louis; Heldmann, Michael; Heller, Mathieu; Hellman, Sten; Helsens, Clement; Henderson, Robert; Henke, Michael; Henrichs, Anna; Henriques Correia, Ana Maria; Henrot-Versille, Sophie; Henry-Couannier, Frédéric; Hensel, Carsten; Henß, Tobias; Medina Hernandez, Carlos; Hernández Jiménez, Yesenia; Herrberg, Ruth; Hershenhorn, Alon David; Herten, Gregor; Hertenberger, Ralf; Hervas, Luis; Hessey, Nigel; Hidvegi, Attila; Higón-Rodriguez, Emilio; Hill, Daniel; Hill, John; Hill, Norman; Hiller, Karl Heinz; Hillert, Sonja; Hillier, Stephen; Hinchliffe, Ian; Hines, Elizabeth; Hirose, Minoru; Hirsch, Florian; Hirschbuehl, Dominic; Hobbs, John; Hod, Noam; Hodgkinson, Mark; Hodgson, Paul; Hoecker, Andreas; Hoeferkamp, Martin; Hoffman, Julia; Hoffmann, Dirk; Hohlfeld, Marc; Holder, Martin; Holmes, Alan; Holmgren, Sven-Olof; Holy, Tomas; Holzbauer, Jenny; Homma, Yasuhiro; Hong, Tae Min; Hooft van Huysduynen, Loek; Horazdovsky, Tomas; Horn, Claus; Horner, Stephan; Horton, Katherine; Hostachy, Jean-Yves; Hou, Suen; Houlden, Michael; Hoummada, Abdeslam; Howarth, James; Howell, David; Hristova, Ivana; Hrivnac, Julius; Hruska, Ivan; Hryn'ova, Tetiana; Hsu, Pai-hsien Jennifer; Hsu, Shih-Chieh; Huang, Guang Shun; Hubacek, Zdenek; Hubaut, Fabrice; Huegging, Fabian; Huffman, Todd Brian; Hughes, Emlyn; Hughes, Gareth; Hughes-Jones, Richard; Huhtinen, Mika; Hurst, Peter; Hurwitz, Martina; Husemann, Ulrich; Huseynov, Nazim; Huston, Joey; Huth, John; Iacobucci, Giuseppe; Iakovidis, Georgios; Ibbotson, Michael; Ibragimov, Iskander; Ichimiya, Ryo; Iconomidou-Fayard, Lydia; Idarraga, John; Idzik, Marek; Iengo, Paolo; Igonkina, Olga; Ikegami, Yoichi; Ikeno, Masahiro; Ilchenko, Yuri; Iliadis, Dimitrios; Imbault, Didier; Imhaeuser, Martin; Imori, Masatoshi; Ince, Tayfun; Inigo-Golfin, Joaquin; Ioannou, Pavlos; Iodice, Mauro; Ionescu, Gelu; Irles Quiles, Adrian; Ishii, Koji; Ishikawa, Akimasa; Ishino, Masaya; Ishmukhametov, Renat; Issever, Cigdem; Istin, Serhat; Itoh, Yuki; Ivashin, Anton; Iwanski, Wieslaw; Iwasaki, Hiroyuki; Izen, Joseph; Izzo, Vincenzo; Jackson, Brett; Jackson, John; Jackson, Paul; Jaekel, Martin; Jain, Vivek; Jakobs, Karl; Jakobsen, Sune; Jakubek, Jan; Jana, Dilip; Jankowski, Ernest; Jansen, Eric; Jantsch, Andreas; Janus, Michel; Jarlskog, Göran; Jeanty, Laura; Jelen, Kazimierz; Jen-La Plante, Imai; Jenni, Peter; Jeremie, Andrea; Jež, Pavel; Jézéquel, Stéphane; Jha, Manoj Kumar; Ji, Haoshuang; Ji, Weina; Jia, Jiangyong; Jiang, Yi; Jimenez Belenguer, Marcos; Jin, Ge; Jin, Shan; Jinnouchi, Osamu; Joergensen, Morten Dam; Joffe, David; Johansen, Lars; Johansen, Marianne; Johansson, Erik; Johansson, Per; Johnert, Sebastian; Johns, Kenneth; Jon-And, Kerstin; Jones, Graham; Jones, Roger; Jones, Tegid; Jones, Tim; Jonsson, Ove; Joram, Christian; Jorge, Pedro; Joseph, John; Ju, Xiangyang; Juranek, Vojtech; Jussel, Patrick; Kabachenko, Vasily; Kabana, Sonja; Kaci, Mohammed; Kaczmarska, Anna; Kadlecik, Peter; Kado, Marumi; Kagan, Harris; Kagan, Michael; Kaiser, Steffen; Kajomovitz, Enrique; Kalinin, Sergey; Kalinovskaya, Lidia; Kama, Sami; Kanaya, Naoko; Kaneda, Michiru; Kanno, Takayuki; Kantserov, Vadim; Kanzaki, Junichi; Kaplan, Benjamin; Kapliy, Anton; Kaplon, Jan; Kar, Deepak; Karagoz, Muge; Karnevskiy, Mikhail; Karr, Kristo; Kartvelishvili, Vakhtang; Karyukhin, Andrey; Kashif, Lashkar; Kasmi, Azzedine; Kass, Richard; Kastanas, Alex; Kataoka, Mayuko; Kataoka, Yousuke; Katsoufis, Elias; Katzy, Judith; Kaushik, Venkatesh; Kawagoe, Kiyotomo; Kawamoto, Tatsuo; Kawamura, Gen; Kayl, Manuel; Kazanin, Vassili; Kazarinov, Makhail; Keates, James Robert; Keeler, Richard; Kehoe, Robert; Keil, Markus; Kekelidze, George; Kelly, Marc; Kennedy, John; Kenney, Christopher John; Kenyon, Mike; Kepka, Oldrich; Kerschen, Nicolas; Kerševan, Borut Paul; Kersten, Susanne; Kessoku, Kohei; Ketterer, Christian; Keung, Justin; Khakzad, Mohsen; Khalil-zada, Farkhad; Khandanyan, Hovhannes; Khanov, Alexander; Kharchenko, Dmitri; Khodinov, Alexander; Kholodenko, Anatoli; Khomich, Andrei; Khoo, Teng Jian; Khoriauli, Gia; Khoroshilov, Andrey; Khovanskiy, Nikolai; Khovanskiy, Valery; Khramov, Evgeniy; Khubua, Jemal; Kim, Hyeon Jin; Kim, Min Suk; Kim, Peter; Kim, Shinhong; Kimura, Naoki; Kind, Oliver; King, Barry; King, Matthew; King, Robert Steven Beaufoy; Kirk, Julie; Kirsch, Guillaume; Kirsch, Lawrence; Kiryunin, Andrey; Kisielewska, Danuta; Kittelmann, Thomas; Kiver, Andrey; Kiyamura, Hironori; Kladiva, Eduard; Klaiber-Lodewigs, Jonas; Klein, Max; Klein, Uta; Kleinknecht, Konrad; Klemetti, Miika; Klier, Amit; Klimentov, Alexei; Klingenberg, Reiner; Klinkby, Esben; Klioutchnikova, Tatiana; Klok, Peter; Klous, Sander; Kluge, Eike-Erik; Kluge, Thomas; Kluit, Peter; Kluth, Stefan; Kneringer, Emmerich; Knobloch, Juergen; Knoops, Edith; Knue, Andrea; Ko, Byeong Rok; Kobayashi, Tomio; Kobel, Michael; Kocian, Martin; Kocnar, Antonin; Kodys, Peter; Köneke, Karsten; König, Adriaan; Koenig, Sebastian; Köpke, Lutz; Koetsveld, Folkert; Koevesarki, Peter; Koffas, Thomas; Koffeman, Els; Kohn, Fabian; Kohout, Zdenek; Kohriki, Takashi; Koi, Tatsumi; Kokott, Thomas; Kolachev, Guennady; Kolanoski, Hermann; Kolesnikov, Vladimir; Koletsou, Iro; Koll, James; Kollar, Daniel; Kollefrath, Michael; Kolya, Scott; Komar, Aston; Komaragiri, Jyothsna Rani; Komori, Yuto; Kondo, Takahiko; Kono, Takanori; Kononov, Anatoly; Konoplich, Rostislav; Konstantinidis, Nikolaos; Kootz, Andreas; Koperny, Stefan; Kopikov, Sergey; Korcyl, Krzysztof; Kordas, Kostantinos; Koreshev, Victor; Korn, Andreas; Korol, Aleksandr; Korolkov, Ilya; Korolkova, Elena; Korotkov, Vladislav; Kortner, Oliver; Kortner, Sandra; Kostyukhin, Vadim; Kotamäki, Miikka Juhani; Kotov, Sergey; Kotov, Vladislav; Kotwal, Ashutosh; Kourkoumelis, Christine; Kouskoura, Vasiliki; Koutsman, Alex; Kowalewski, Robert Victor; Kowalski, Tadeusz; Kozanecki, Witold; Kozhin, Anatoly; Kral, Vlastimil; Kramarenko, Viktor; Kramberger, Gregor; Krasel, Olaf; Krasny, Mieczyslaw Witold; Krasznahorkay, Attila; Kraus, James; Kreisel, Arik; Krejci, Frantisek; Kretzschmar, Jan; Krieger, Nina; Krieger, Peter; Kroeninger, Kevin; Kroha, Hubert; Kroll, Joe; Kroseberg, Juergen; Krstic, Jelena; Kruchonak, Uladzimir; Krüger, Hans; Kruker, Tobias; Krumshteyn, Zinovii; Kruth, Andre; Kubota, Takashi; Kuehn, Susanne; Kugel, Andreas; Kuhl, Thorsten; Kuhn, Dietmar; Kukhtin, Victor; Kulchitsky, Yuri; Kuleshov, Sergey; Kummer, Christian; Kuna, Marine; Kundu, Nikhil; Kunkle, Joshua; Kupco, Alexander; Kurashige, Hisaya; Kurata, Masakazu; Kurochkin, Yurii; Kus, Vlastimil; Kuykendall, William; Kuze, Masahiro; Kuzhir, Polina; Kvasnicka, Ondrej; Kvita, Jiri; Kwee, Regina; La Rosa, Alessandro; La Rotonda, Laura; Labarga, Luis; Labbe, Julien; Lablak, Said; Lacasta, Carlos; Lacava, Francesco; Lacker, Heiko; Lacour, Didier; Lacuesta, Vicente Ramón; Ladygin, Evgueni; Lafaye, Rémi; Laforge, Bertrand; Lagouri, Theodota; Lai, Stanley; Laisne, Emmanuel; Lamanna, Massimo; Lampen, Caleb; Lampl, Walter; Lancon, Eric; Landgraf, Ulrich; Landon, Murrough; Landsman, Hagar; Lane, Jenna; Lange, Clemens; Lankford, Andrew; Lanni, Francesco; Lantzsch, Kerstin; Lapin, Vladimir; Laplace, Sandrine; Lapoire, Cecile; Laporte, Jean-Francois; Lari, Tommaso; Larionov, Anatoly; Larner, Aimee; Lasseur, Christian; Lassnig, Mario; Lau, Wing; Laurelli, Paolo; Lavorato, Antonia; Lavrijsen, Wim; Laycock, Paul; Lazarev, Alexandre; Lazzaro, Alfio; Le Dortz, Olivier; Le Guirriec, Emmanuel; Le Maner, Christophe; Le Menedeu, Eve; Lebedev, Alexander; Lebel, Céline; LeCompte, Thomas; Ledroit-Guillon, Fabienne Agnes Marie; Lee, Hurng-Chun; Lee, Jason; Lee, Shih-Chang; Lee, Lawrence; Lefebvre, Michel; Legendre, Marie; Leger, Annie; LeGeyt, Benjamin; Legger, Federica; Leggett, Charles; Lehmacher, Marc; Lehmann Miotto, Giovanna; Lei, Xiaowen; Leite, Marco Aurelio Lisboa; Leitner, Rupert; Lellouch, Daniel; Leltchouk, Mikhail; Lendermann, Victor; Leney, Katharine; Lenz, Tatiana; Lenzen, Georg; Lenzi, Bruno; Leonhardt, Kathrin; Leontsinis, Stefanos; Leroy, Claude; Lessard, Jean-Raphael; Lesser, Jonas; Lester, Christopher; Leung Fook Cheong, Annabelle; Levêque, Jessica; Levin, Daniel; Levinson, Lorne; Levitski, Mikhail; Lewandowska, Marta; Lewis, Adrian; Lewis, George; Leyko, Agnieszka; Leyton, Michael; Li, Bo; Li, Haifeng; Li, Shu; Li, Xuefei; Liang, Zhihua; Liang, Zhijun; Liberti, Barbara; Lichard, Peter; Lichtnecker, Markus; Lie, Ki; Liebig, Wolfgang; Lifshitz, Ronen; Lilley, Joseph; Limbach, Christian; Limosani, Antonio; Limper, Maaike; Lin, Simon; Linde, Frank; Linnemann, James; Lipeles, Elliot; Lipinsky, Lukas; Lipniacka, Anna; Liss, Tony; Lissauer, David; Lister, Alison; Litke, Alan; Liu, Chuanlei; Liu, Dong; Liu, Hao; Liu, Jianbei; Liu, Minghui; Liu, Shengli; Liu, Yanwen; Livan, Michele; Livermore, Sarah; Lleres, Annick; Llorente Merino, Javier; Lloyd, Stephen; Lobodzinska, Ewelina; Loch, Peter; Lockman, William; Lockwitz, Sarah; Loddenkoetter, Thomas; Loebinger, Fred; Loginov, Andrey; Loh, Chang Wei; Lohse, Thomas; Lohwasser, Kristin; Lokajicek, Milos; Loken, James; Lombardo, Vincenzo Paolo; Long, Robin Eamonn; Lopes, Lourenco; Lopez Mateos, David; Losada, Marta; Loscutoff, Peter; Lo Sterzo, Francesco; Losty, Michael; Lou, Xinchou; Lounis, Abdenour; Loureiro, Karina; Love, Jeremy; Love, Peter; Lowe, Andrew; Lu, Feng; Lu, Liang; Lubatti, Henry; Luci, Claudio; Lucotte, Arnaud; Ludwig, Andreas; Ludwig, Dörthe; Ludwig, Inga; Ludwig, Jens; Luehring, Frederick; Luijckx, Guy; Lumb, Debra; Luminari, Lamberto; Lund, Esben; Lund-Jensen, Bengt; Lundberg, Björn; Lundberg, Johan; Lundquist, Johan; Lungwitz, Matthias; Lupi, Anna; Lutz, Gerhard; Lynn, David; Lys, Jeremy; Lytken, Else; Ma, Hong; Ma, Lian Liang; Macana Goia, Jorge Andres; Maccarrone, Giovanni; Macchiolo, Anna; Maček, Boštjan; Machado Miguens, Joana; Mackeprang, Rasmus; Madaras, Ronald; Mader, Wolfgang; Maenner, Reinhard; Maeno, Tadashi; Mättig, Peter; Mättig, Stefan; Magalhaes Martins, Paulo Jorge; Magnoni, Luca; Magradze, Erekle; Mahalalel, Yair; Mahboubi, Kambiz; Mahout, Gilles; Maiani, Camilla; Maidantchik, Carmen; Maio, Amélia; Majewski, Stephanie; Makida, Yasuhiro; Makovec, Nikola; Mal, Prolay; Malecki, Pawel; Malecki, Piotr; Maleev, Victor; Malek, Fairouz; Mallik, Usha; Malon, David; Maltezos, Stavros; Malyshev, Vladimir; Malyukov, Sergei; Mameghani, Raphael; Mamuzic, Judita; Manabe, Atsushi; Mandelli, Luciano; Mandić, Igor; Mandrysch, Rocco; Maneira, José; Mangeard, Pierre-Simon; Manjavidze, Ioseb; Mann, Alexander; Manning, Peter; Manousakis-Katsikakis, Arkadios; Mansoulie, Bruno; Manz, Andreas; Mapelli, Alessandro; Mapelli, Livio; March, Luis; Marchand, Jean-Francois; Marchese, Fabrizio; Marchiori, Giovanni; Marcisovsky, Michal; Marin, Alexandru; Marino, Christopher; Marroquim, Fernando; Marshall, Robin; Marshall, Zach; Martens, Kalen; Marti-Garcia, Salvador; Martin, Andrew; Martin, Brian; Martin, Brian Thomas; Martin, Franck Francois; Martin, Jean-Pierre; Martin, Philippe; Martin, Tim; Martin dit Latour, Bertrand; Martinez, Mario; Martinez Outschoorn, Verena; Martyniuk, Alex; Marx, Marilyn; Marzano, Francesco; Marzin, Antoine; Masetti, Lucia; Mashimo, Tetsuro; Mashinistov, Ruslan; Masik, Jiri; Maslennikov, Alexey; Maß, Martin; Massa, Ignazio; Massaro, Graziano; Massol, Nicolas; Mastrandrea, Paolo; Mastroberardino, Anna; Masubuchi, Tatsuya; Mathes, Markus; Matricon, Pierre; Matsumoto, Hiroshi; Matsunaga, Hiroyuki; Matsushita, Takashi; Mattravers, Carly; Maugain, Jean-Marie; Maxfield, Stephen; Maximov, Dmitriy; May, Edward; Mayne, Anna; Mazini, Rachid; Mazur, Michael; Mazzanti, Marcello; Mazzoni, Enrico; Mc Kee, Shawn Patrick; McCarn, Allison; McCarthy, Robert; McCarthy, Tom; McCubbin, Norman; McFarlane, Kenneth; Mcfayden, Josh; McGlone, Helen; Mchedlidze, Gvantsa; McLaren, Robert Andrew; Mclaughlan, Tom; McMahon, Steve; McPherson, Robert; Meade, Andrew; Mechnich, Joerg; Mechtel, Markus; Medinnis, Mike; Meera-Lebbai, Razzak; Meguro, Tatsuma; Mehdiyev, Rashid; Mehlhase, Sascha; Mehta, Andrew; Meier, Karlheinz; Meinhardt, Jens; Meirose, Bernhard; Melachrinos, Constantinos; Mellado Garcia, Bruce Rafael; Mendoza Navas, Luis; Meng, Zhaoxia; Mengarelli, Alberto; Menke, Sven; Menot, Claude; Meoni, Evelin; Mercurio, Kevin Michael; Mermod, Philippe; Merola, Leonardo; Meroni, Chiara; Merritt, Frank; Messina, Andrea; Metcalfe, Jessica; Mete, Alaettin Serhan; Meuser, Stefan; Meyer, Carsten; Meyer, Jean-Pierre; Meyer, Jochen; Meyer, Joerg; Meyer, Thomas Christian; Meyer, W Thomas; Miao, Jiayuan; Michal, Sebastien; Micu, Liliana; Middleton, Robin; Miele, Paola; Migas, Sylwia; Mijović, Liza; Mikenberg, Giora; Mikestikova, Marcela; Mikuž, Marko; Miller, David; Miller, Robert; Mills, Bill; Mills, Corrinne; Milov, Alexander; Milstead, David; Milstein, Dmitry; Minaenko, Andrey; Miñano, Mercedes; Minashvili, Irakli; Mincer, Allen; Mindur, Bartosz; Mineev, Mikhail; Ming, Yao; Mir, Lluisa-Maria; Mirabelli, Giovanni; Miralles Verge, Lluis; Misiejuk, Andrzej; Mitrevski, Jovan; Mitrofanov, Gennady; Mitsou, Vasiliki A; Mitsui, Shingo; Miyagawa, Paul; Miyazaki, Kazuki; Mjörnmark, Jan-Ulf; Moa, Torbjoern; Mockett, Paul; Moed, Shulamit; Moeller, Victoria; Mönig, Klaus; Möser, Nicolas; Mohapatra, Soumya; Mohn, Bjarte; Mohr, Wolfgang; Mohrdieck-Möck, Susanne; Moisseev, Artemy; Moles-Valls, Regina; Molina-Perez, Jorge; Monk, James; Monnier, Emmanuel; Montesano, Simone; Monticelli, Fernando; Monzani, Simone; Moore, Roger; Moorhead, Gareth; Mora Herrera, Clemencia; Moraes, Arthur; Morais, Antonio; Morange, Nicolas; Morel, Julien; Morello, Gianfranco; Moreno, Deywis; Moreno Llácer, María; Morettini, Paolo; Morii, Masahiro; Morin, Jerome; Morita, Youhei; Morley, Anthony Keith; Mornacchi, Giuseppe; Morone, Maria-Christina; Morozov, Sergey; Morris, John; Morvaj, Ljiljana; Moser, Hans-Guenther; Mosidze, Maia; Moss, Josh; Mount, Richard; Mountricha, Eleni; Mouraviev, Sergei; Moyse, Edward; Mudrinic, Mihajlo; Mueller, Felix; Mueller, James; Mueller, Klemens; Müller, Thomas; Muenstermann, Daniel; Muijs, Sandra; Muir, Alex; Munwes, Yonathan; Murakami, Koichi; Murray, Bill; Mussche, Ido; Musto, Elisa; Myagkov, Alexey; Myska, Miroslav; Nadal, Jordi; Nagai, Koichi; Nagano, Kunihiro; Nagasaka, Yasushi; Nairz, Armin Michael; Nakahama, Yu; Nakamura, Koji; Nakano, Itsuo; Nanava, Gizo; Napier, Austin; Nash, Michael; Nation, Nigel; Nattermann, Till; Naumann, Thomas; Navarro, Gabriela; Neal, Homer; Nebot, Eduardo; Nechaeva, Polina; Negri, Andrea; Negri, Guido; Nektarijevic, Snezana; Nelson, Andrew; Nelson, Silke; Nelson, Timothy Knight; Nemecek, Stanislav; Nemethy, Peter; Nepomuceno, Andre Asevedo; Nessi, Marzio; Nesterov, Stanislav; Neubauer, Mark; Neusiedl, Andrea; Neves, Ricardo; Nevski, Pavel; Newman, Paul; Nguyen Thi Hong, Van; Nickerson, Richard; Nicolaidou, Rosy; Nicolas, Ludovic; Nicquevert, Bertrand; Niedercorn, Francois; Nielsen, Jason; Niinikoski, Tapio; Nikiforov, Andriy; Nikolaenko, Vladimir; Nikolaev, Kirill; Nikolic-Audit, Irena; Nikolopoulos, Konstantinos; Nilsen, Henrik; Nilsson, Paul; Ninomiya, Yoichi; Nisati, Aleandro; Nishiyama, Tomonori; Nisius, Richard; Nodulman, Lawrence; Nomachi, Masaharu; Nomidis, Ioannis; Nomoto, Hiroshi; Nordberg, Markus; Nordkvist, Bjoern; Norton, Peter; Novakova, Jana; Nozaki, Mitsuaki; Nožička, Miroslav; Nozka, Libor; Nugent, Ian Michael; Nuncio-Quiroz, Adriana-Elizabeth; Nunes Hanninger, Guilherme; Nunnemann, Thomas; Nurse, Emily; Nyman, Tommi; O'Brien, Brendan Joseph; O'Neale, Steve; O'Neil, Dugan; O'Shea, Val; Oakham, Gerald; Oberlack, Horst; Ocariz, Jose; Ochi, Atsuhiko; Oda, Susumu; Odaka, Shigeru; Odier, Jerome; Ogren, Harold; Oh, Alexander; Oh, Seog; Ohm, Christian; Ohshima, Takayoshi; Ohshita, Hidetoshi; Ohska, Tokio Kenneth; Ohsugi, Takashi; Okada, Shogo; Okawa, Hideki; Okumura, Yasuyuki; Okuyama, Toyonobu; Olcese, Marco; Olchevski, Alexander; Oliveira, Miguel Alfonso; Oliveira Damazio, Denis; Oliver Garcia, Elena; Olivito, Dominick; Olszewski, Andrzej; Olszowska, Jolanta; Omachi, Chihiro; Onofre, António; Onyisi, Peter; Oram, Christopher; Oreglia, Mark; Oren, Yona; Orestano, Domizia; Orlov, Iliya; Oropeza Barrera, Cristina; Orr, Robert; Ortega, Eduardo; Osculati, Bianca; Ospanov, Rustem; Osuna, Carlos; Otero y Garzon, Gustavo; Ottersbach, John; Ouchrif, Mohamed; Ould-Saada, Farid; Ouraou, Ahmimed; Ouyang, Qun; Owen, Mark; Owen, Simon; Øye, Ola; Ozcan, Veysi Erkcan; Ozturk, Nurcan; Pacheco Pages, Andres; Padilla Aranda, Cristobal; Paganis, Efstathios; Paige, Frank; Pajchel, Katarina; Palestini, Sandro; Pallin, Dominique; Palma, Alberto; Palmer, Jody; Pan, Yibin; Panagiotopoulou, Evgenia; Panes, Boris; Panikashvili, Natalia; Panitkin, Sergey; Pantea, Dan; Panuskova, Monika; Paolone, Vittorio; Papadelis, Aras; Papadopoulou, Theodora; Paramonov, Alexander; Park, Woochun; Parker, Andy; Parodi, Fabrizio; Parsons, John; Parzefall, Ulrich; Pasqualucci, Enrico; Passeri, Antonio; Pastore, Fernanda; Pastore, Francesca; Pásztor, Gabriella; Pataraia, Sophio; Patel, Nikhul; Pater, Joleen; Patricelli, Sergio; Pauly, Thilo; Pecsy, Martin; Pedraza Morales, Maria Isabel; Peleganchuk, Sergey; Peng, Haiping; Pengo, Ruggero; Penson, Alexander; Penwell, John; Perantoni, Marcelo; Perez, Kerstin; Perez Cavalcanti, Tiago; Perez Codina, Estel; Pérez García-Estañ, María Teresa; Perez Reale, Valeria; Peric, Ivan; Perini, Laura; Pernegger, Heinz; Perrino, Roberto; Perrodo, Pascal; Persembe, Seda; Peshekhonov, Vladimir; Peters, Onne; Petersen, Brian; Petersen, Jorgen; Petersen, Troels; Petit, Elisabeth; Petridis, Andreas; Petridou, Chariclia; Petrolo, Emilio; Petrucci, Fabrizio; Petschull, Dennis; Petteni, Michele; Pezoa, Raquel; Phan, Anna; Phillips, Alan; Phillips, Peter William; Piacquadio, Giacinto; Piccaro, Elisa; Piccinini, Maurizio; Pickford, Andrew; Piec, Sebastian Marcin; Piegaia, Ricardo; Pilcher, James; Pilkington, Andrew; Pina, João Antonio; Pinamonti, Michele; Pinder, Alex; Pinfold, James; Ping, Jialun; Pinto, Belmiro; Pirotte, Olivier; Pizio, Caterina; Placakyte, Ringaile; Plamondon, Mathieu; Plano, Will; Pleier, Marc-Andre; Pleskach, Anatoly; Poblaguev, Andrei; Poddar, Sahill; Podlyski, Fabrice; Poggioli, Luc; Poghosyan, Tatevik; Pohl, Martin; Polci, Francesco; Polesello, Giacomo; Policicchio, Antonio; Polini, Alessandro; Poll, James; Polychronakos, Venetios; Pomarede, Daniel Marc; Pomeroy, Daniel; Pommès, Kathy; Pontecorvo, Ludovico; Pope, Bernard; Popeneciu, Gabriel Alexandru; Popovic, Dragan; Poppleton, Alan; Portell Bueso, Xavier; Porter, Robert; Posch, Christoph; Pospelov, Guennady; Pospisil, Stanislav; Potrap, Igor; Potter, Christina; Potter, Christopher; Poulard, Gilbert; Poveda, Joaquin; Prabhu, Robindra; Pralavorio, Pascal; Prasad, Srivas; Pravahan, Rishiraj; Prell, Soeren; Pretzl, Klaus Peter; Pribyl, Lukas; Price, Darren; Price, Lawrence; Price, Michael John; Prichard, Paul; Prieur, Damien; Primavera, Margherita; Prokofiev, Kirill; Prokoshin, Fedor; Protopopescu, Serban; Proudfoot, James; Prudent, Xavier; Przysiezniak, Helenka; Psoroulas, Serena; Ptacek, Elizabeth; Purdham, John; Purohit, Milind; Puzo, Patrick; Pylypchenko, Yuriy; Qian, Jianming; Qian, Zuxuan; Qin, Zhonghua; Quadt, Arnulf; Quarrie, David; Quayle, William; Quinonez, Fernando; Raas, Marcel; Radescu, Voica; Radics, Balint; Rador, Tonguc; Ragusa, Francesco; Rahal, Ghita; Rahimi, Amir; Rahm, David; Rajagopalan, Srinivasan; Rammensee, Michael; Rammes, Marcus; Ramstedt, Magnus; Randrianarivony, Koloina; Ratoff, Peter; Rauscher, Felix; Rauter, Emanuel; Raymond, Michel; Read, Alexander Lincoln; Rebuzzi, Daniela; Redelbach, Andreas; Redlinger, George; Reece, Ryan; Reeves, Kendall; Reichold, Armin; Reinherz-Aronis, Erez; Reinsch, Andreas; Reisinger, Ingo; Reljic, Dusan; Rembser, Christoph; Ren, Zhongliang; Renaud, Adrien; Renkel, Peter; Rensch, Bertram; Rescigno, Marco; Resconi, Silvia; Resende, Bernardo; Reznicek, Pavel; Rezvani, Reyhaneh; Richards, Alexander; Richter, Robert; Richter-Was, Elzbieta; Ridel, Melissa; Rieke, Stefan; Rijpstra, Manouk; Rijssenbeek, Michael; Rimoldi, Adele; Rinaldi, Lorenzo; Rios, Ryan Randy; Riu, Imma; Rivoltella, Giancesare; Rizatdinova, Flera; Rizvi, Eram; Robertson, Steven; Robichaud-Veronneau, Andree; Robinson, Dave; Robinson, James; Robinson, Mary; Robson, Aidan; Rocha de Lima, Jose Guilherme; Roda, Chiara; Roda Dos Santos, Denis; Rodier, Stephane; Rodriguez, Diego; Rodriguez Garcia, Yohany; Roe, Adam; Roe, Shaun; Røhne, Ole; Rojo, Victoria; Rolli, Simona; Romaniouk, Anatoli; Romanov, Victor; Romeo, Gaston; Romero Maltrana, Diego; Roos, Lydia; Ros, Eduardo; Rosati, Stefano; Rosbach, Kilian; Rose, Matthew; Rosenbaum, Gabriel; Rosenberg, Eli; Rosendahl, Peter Lundgaard; Rosselet, Laurent; Rossetti, Valerio; Rossi, Elvira; Rossi, Leonardo Paolo; Rossi, Lucio; Rotaru, Marina; Roth, Itamar; Rothberg, Joseph; Rousseau, David; Royon, Christophe; Rozanov, Alexander; Rozen, Yoram; Ruan, Xifeng; Rubinskiy, Igor; Ruckert, Benjamin; Ruckstuhl, Nicole; Rud, Viacheslav; Rudolph, Gerald; Rühr, Frederik; Ruggieri, Federico; Ruiz-Martinez, Aranzazu; Rulikowska-Zarebska, Elzbieta; Rumiantsev, Viktor; Rumyantsev, Leonid; Runge, Kay; Runolfsson, Ogmundur; Rurikova, Zuzana; Rusakovich, Nikolai; Rust, Dave; Rutherfoord, John; Ruwiedel, Christoph; Ruzicka, Pavel; Ryabov, Yury; Ryadovikov, Vasily; Ryan, Patrick; Rybar, Martin; Rybkin, Grigori; Ryder, Nick; Rzaeva, Sevda; Saavedra, Aldo; Sadeh, Iftach; Sadrozinski, Hartmut; Sadykov, Renat; Safai Tehrani, Francesco; Sakamoto, Hiroshi; Salamanna, Giuseppe; Salamon, Andrea; Saleem, Muhammad; Salihagic, Denis; Salnikov, Andrei; Salt, José; Salvachua Ferrando, Belén; Salvatore, Daniela; Salvatore, Pasquale Fabrizio; Salvucci, Antonio; Salzburger, Andreas; Sampsonidis, Dimitrios; Samset, Björn Hallvard; Sandaker, Heidi; Sander, Heinz Georg; Sanders, Michiel; Sandhoff, Marisa; Sandoval, Tanya; Sandstroem, Rikard; Sandvoss, Stephan; Sankey, Dave; Sansoni, Andrea; Santamarina Rios, Cibran; Santoni, Claudio; Santonico, Rinaldo; Santos, Helena; Saraiva, João; Sarangi, Tapas; Sarkisyan-Grinbaum, Edward; Sarri, Francesca; Sartisohn, Georg; Sasaki, Osamu; Sasaki, Takashi; Sasao, Noboru; Satsounkevitch, Igor; Sauvage, Gilles; Sauvan, Jean-Baptiste; Savard, Pierre; Savinov, Vladimir; Savu, Dan Octavian; Savva, Panagiota; Sawyer, Lee; Saxon, David; Says, Louis-Pierre; Sbarra, Carla; Sbrizzi, Antonio; Scallon, Olivia; Scannicchio, Diana; Scarcella, Mark; Schaarschmidt, Jana; Schacht, Peter; Schäfer, Uli; Schaepe, Steffen; Schaetzel, Sebastian; Schaffer, Arthur; Schaile, Dorothee; Schamberger, R. Dean; Schamov, Andrey; Scharf, Veit; Schegelsky, Valery; Scheirich, Daniel; Schernau, Michael; Scherzer, Max; Schiavi, Carlo; Schieck, Jochen; Schioppa, Marco; Schlenker, Stefan; Schlereth, James; Schmidt, Evelyn; Schmidt, Michael; Schmieden, Kristof; Schmitt, Christian; Schmitt, Sebastian; Schmitz, Martin; Schöning, André; Schott, Matthias; Schouten, Doug; Schovancova, Jaroslava; Schram, Malachi; Schroeder, Christian; Schroer, Nicolai; Schuh, Silvia; Schuler, Georges; Schultes, Joachim; Schultz-Coulon, Hans-Christian; Schulz, Holger; Schumacher, Jan; Schumacher, Markus; Schumm, Bruce; Schune, Philippe; Schwanenberger, Christian; Schwartzman, Ariel; Schwemling, Philippe; Schwienhorst, Reinhard; Schwierz, Rainer; Schwindling, Jerome; Scott, Bill; Searcy, Jacob; Sedykh, Evgeny; Segura, Ester; Seidel, Sally; Seiden, Abraham; Seifert, Frank; Seixas, José; Sekhniaidze, Givi; Seliverstov, Dmitry; Sellden, Bjoern; Sellers, Graham; Seman, Michal; Semprini-Cesari, Nicola; Serfon, Cedric; Serin, Laurent; Seuster, Rolf; Severini, Horst; Sevior, Martin; Sfyrla, Anna; Shabalina, Elizaveta; Shamim, Mansoora; Shan, Lianyou; Shank, James; Shao, Qi Tao; Shapiro, Marjorie; Shatalov, Pavel; Shaver, Leif; Shaw, Christian; Shaw, Kate; Sherman, Daniel; Sherwood, Peter; Shibata, Akira; Shichi, Hideharu; Shimizu, Shima; Shimojima, Makoto; Shin, Taeksu; Shmeleva, Alevtina; Shochet, Mel; Short, Daniel; Shupe, Michael; Sicho, Petr; Sidoti, Antonio; Siebel, Anca-Mirela; Siegert, Frank; Siegrist, James; Sijacki, Djordje; Silbert, Ohad; Silva, José; Silver, Yiftah; Silverstein, Daniel; Silverstein, Samuel; Simak, Vladislav; Simard, Olivier; Simic, Ljiljana; Simion, Stefan; Simmons, Brinick; Simonyan, Margar; Sinervo, Pekka; Sinev, Nikolai; Sipica, Valentin; Siragusa, Giovanni; Sisakyan, Alexei; Sivoklokov, Serguei; Sjölin, Jörgen; Sjursen, Therese; Skinnari, Louise Anastasia; Skovpen, Kirill; Skubic, Patrick; Skvorodnev, Nikolai; Slater, Mark; Slavicek, Tomas; Sliwa, Krzysztof; Sloan, Terrence; Sloper, John erik; Smakhtin, Vladimir; Smirnov, Sergei; Smirnova, Lidia; Smirnova, Oxana; Smith, Ben Campbell; Smith, Douglas; Smith, Kenway; Smizanska, Maria; Smolek, Karel; Snesarev, Andrei; Snow, Steve; Snow, Joel; Snuverink, Jochem; Snyder, Scott; Soares, Mara; Sobie, Randall; Sodomka, Jaromir; Soffer, Abner; Solans, Carlos; Solar, Michael; Solc, Jaroslav; Soldatov, Evgeny; Soldevila, Urmila; Solfaroli Camillocci, Elena; Solodkov, Alexander; Solovyanov, Oleg; Sondericker, John; Soni, Nitesh; Sopko, Vit; Sopko, Bruno; Sorbi, Massimo; Sosebee, Mark; Soukharev, Andrey; Spagnolo, Stefania; Spanò, Francesco; Spighi, Roberto; Spigo, Giancarlo; Spila, Federico; Spiriti, Eleuterio; Spiwoks, Ralf; Spousta, Martin; Spreitzer, Teresa; Spurlock, Barry; St Denis, Richard Dante; Stahl, Thorsten; Stahlman, Jonathan; Stamen, Rainer; Stanecka, Ewa; Stanek, Robert; Stanescu, Cristian; Stapnes, Steinar; Starchenko, Evgeny; Stark, Jan; Staroba, Pavel; Starovoitov, Pavel; Staude, Arnold; Stavina, Pavel; Stavropoulos, Georgios; Steele, Genevieve; Steinbach, Peter; Steinberg, Peter; Stekl, Ivan; Stelzer, Bernd; Stelzer, Harald Joerg; Stelzer-Chilton, Oliver; Stenzel, Hasko; Stevenson, Kyle; Stewart, Graeme; Stillings, Jan Andre; Stockmanns, Tobias; Stockton, Mark; Stoerig, Kathrin; Stoicea, Gabriel; Stonjek, Stefan; Strachota, Pavel; Stradling, Alden; Straessner, Arno; Strandberg, Jonas; Strandberg, Sara; Strandlie, Are; Strang, Michael; Strauss, Emanuel; Strauss, Michael; Strizenec, Pavol; Ströhmer, Raimund; Strom, David; Strong, John; Stroynowski, Ryszard; Strube, Jan; Stugu, Bjarne; Stumer, Iuliu; Stupak, John; Sturm, Philipp; Soh, Dart-yin; Su, Dong; Subramania, Halasya Siva; Succurro, Antonella; Sugaya, Yorihito; Sugimoto, Takuya; Suhr, Chad; Suita, Koichi; Suk, Michal; Sulin, Vladimir; Sultansoy, Saleh; Sumida, Toshi; Sun, Xiaohu; Sundermann, Jan Erik; Suruliz, Kerim; Sushkov, Serge; Susinno, Giancarlo; Sutton, Mark; Suzuki, Yu; Svatos, Michal; Sviridov, Yuri; Swedish, Stephen; Sykora, Ivan; Sykora, Tomas; Szeless, Balazs; Sánchez, Javier; Ta, Duc; Tackmann, Kerstin; Taffard, Anyes; Tafirout, Reda; Taga, Adrian; Taiblum, Nimrod; Takahashi, Yuta; Takai, Helio; Takashima, Ryuichi; Takeda, Hiroshi; Takeshita, Tohru; Talby, Mossadek; Talyshev, Alexey; Tamsett, Matthew; Tanaka, Junichi; Tanaka, Reisaburo; Tanaka, Satoshi; Tanaka, Shuji; Tanaka, Yoshito; Tani, Kazutoshi; Tannoury, Nancy; Tappern, Geoffrey; Tapprogge, Stefan; Tardif, Dominique; Tarem, Shlomit; Tarrade, Fabien; Tartarelli, Giuseppe Francesco; Tas, Petr; Tasevsky, Marek; Tassi, Enrico; Tatarkhanov, Mous; Tayalati, Yahya; Taylor, Christopher; Taylor, Frank; Taylor, Geoffrey; Taylor, Wendy; Teixeira Dias Castanheira, Matilde; Teixeira-Dias, Pedro; Temming, Kim Katrin; Ten Kate, Herman; Teng, Ping-Kun; Terada, Susumu; Terashi, Koji; Terron, Juan; Terwort, Mark; Testa, Marianna; Teuscher, Richard; Thadome, Jocelyn; Therhaag, Jan; Theveneaux-Pelzer, Timothée; Thioye, Moustapha; Thoma, Sascha; Thomas, Juergen; Thompson, Emily; Thompson, Paul; Thompson, Peter; Thompson, Stan; Thomson, Evelyn; Thomson, Mark; Thun, Rudolf; Tic, Tomáš; Tikhomirov, Vladimir; Tikhonov, Yury; Timmermans, Charles; Tipton, Paul; Tique Aires Viegas, Florbela De Jes; Tisserant, Sylvain; Tobias, Jürgen; Toczek, Barbara; Todorov, Theodore; Todorova-Nova, Sharka; Toggerson, Brokk; Tojo, Junji; Tokár, Stanislav; Tokunaga, Kaoru; Tokushuku, Katsuo; Tollefson, Kirsten; Tomoto, Makoto; Tompkins, Lauren; Toms, Konstantin; Tong, Guoliang; Tonoyan, Arshak; Topfel, Cyril; Topilin, Nikolai; Torchiani, Ingo; Torrence, Eric; Torró Pastor, Emma; Toth, Jozsef; Touchard, Francois; Tovey, Daniel; Traynor, Daniel; Trefzger, Thomas; Treis, Johannes; Tremblet, Louis; Tricoli, Alesandro; Trigger, Isabel Marian; Trincaz-Duvoid, Sophie; Trinh, Thi Nguyet; Tripiana, Martin; Trischuk, William; Trivedi, Arjun; Trocmé, Benjamin; Troncon, Clara; Trottier-McDonald, Michel; Trzupek, Adam; Tsarouchas, Charilaos; Tseng, Jeffrey; Tsiakiris, Menelaos; Tsiareshka, Pavel; Tsionou, Dimitra; Tsipolitis, Georgios; Tsiskaridze, Vakhtang; Tskhadadze, Edisher; Tsukerman, Ilya; Tsulaia, Vakhtang; Tsung, Jieh-Wen; Tsuno, Soshi; Tsybychev, Dmitri; Tua, Alan; Tuggle, Joseph; Turala, Michal; Turecek, Daniel; Turk Cakir, Ilkay; Turlay, Emmanuel; Turra, Ruggero; Tuts, Michael; Tykhonov, Andrii; Tylmad, Maja; Tyndel, Mike; Tyrvainen, Harri; Tzanakos, George; Uchida, Kirika; Ueda, Ikuo; Ueno, Ryuichi; Ugland, Maren; Uhlenbrock, Mathias; Uhrmacher, Michael; Ukegawa, Fumihiko; Unal, Guillaume; Underwood, David; Undrus, Alexander; Unel, Gokhan; Unno, Yoshinobu; Urbaniec, Dustin; Urkovsky, Evgeny; Urrejola, Pedro; Usai, Giulio; Uslenghi, Massimiliano; Vacavant, Laurent; Vacek, Vaclav; Vachon, Brigitte; Vahsen, Sven; Valenta, Jan; Valente, Paolo; Valentinetti, Sara; Valkar, Stefan; Valladolid Gallego, Eva; Vallecorsa, Sofia; Valls Ferrer, Juan Antonio; van der Graaf, Harry; van der Kraaij, Erik; Van Der Leeuw, Robin; van der Poel, Egge; van der Ster, Daniel; Van Eijk, Bob; van Eldik, Niels; van Gemmeren, Peter; van Kesteren, Zdenko; van Vulpen, Ivo; Vandelli, Wainer; Vandoni, Giovanna; Vaniachine, Alexandre; Vankov, Peter; Vannucci, Francois; Varela Rodriguez, Fernando; Vari, Riccardo; Varnes, Erich; Varouchas, Dimitris; Vartapetian, Armen; Varvell, Kevin; Vassilakopoulos, Vassilios; Vazeille, Francois; Vegni, Guido; Veillet, Jean-Jacques; Vellidis, Constantine; Veloso, Filipe; Veness, Raymond; Veneziano, Stefano; Ventura, Andrea; Ventura, Daniel; Venturi, Manuela; Venturi, Nicola; Vercesi, Valerio; Verducci, Monica; Verkerke, Wouter; Vermeulen, Jos; Vest, Anja; Vetterli, Michel; Vichou, Irene; Vickey, Trevor; Viehhauser, Georg; Viel, Simon; Villa, Mauro; Villaplana Perez, Miguel; Vilucchi, Elisabetta; Vincter, Manuella; Vinek, Elisabeth; Vinogradov, Vladimir; Virchaux, Marc; Viret, Sébastien; Virzi, Joseph; Vitale, Antonio; Vitells, Ofer; Viti, Michele; Vivarelli, Iacopo; Vives Vaque, Francesc; Vlachos, Sotirios; Vlasak, Michal; Vlasov, Nikolai; Vogel, Adrian; Vokac, Petr; Volpi, Guido; Volpi, Matteo; Volpini, Giovanni; von der Schmitt, Hans; von Loeben, Joerg; von Radziewski, Holger; von Toerne, Eckhard; Vorobel, Vit; Vorobiev, Alexander; Vorwerk, Volker; Vos, Marcel; Voss, Rudiger; Voss, Thorsten Tobias; Vossebeld, Joost; Vranjes, Nenad; Vranjes Milosavljevic, Marija; Vrba, Vaclav; Vreeswijk, Marcel; Vu Anh, Tuan; Vuillermet, Raphael; Vukotic, Ilija; Wagner, Wolfgang; Wagner, Peter; Wahlen, Helmut; Wakabayashi, Jun; Walbersloh, Jorg; Walch, Shannon; Walder, James; Walker, Rodney; Walkowiak, Wolfgang; Wall, Richard; Waller, Peter; Wang, Chiho; Wang, Haichen; Wang, Hulin; Wang, Jike; Wang, Jin; Wang, Joshua C; Wang, Rui; Wang, Song-Ming; Warburton, Andreas; Ward, Patricia; Warsinsky, Markus; Watkins, Peter; Watson, Alan; Watson, Miriam; Watts, Gordon; Watts, Stephen; Waugh, Anthony; Waugh, Ben; Weber, Jens; Weber, Marc; Weber, Michele; Weber, Pavel; Weidberg, Anthony; Weigell, Philipp; Weingarten, Jens; Weiser, Christian; Wellenstein, Hermann; Wells, Phillippa; Wen, Mei; Wenaus, Torre; Wendler, Shanti; Weng, Zhili; Wengler, Thorsten; Wenig, Siegfried; Wermes, Norbert; Werner, Matthias; Werner, Per; Werth, Michael; Wessels, Martin; Weydert, Carole; Whalen, Kathleen; Wheeler-Ellis, Sarah Jane; Whitaker, Scott; White, Andrew; White, Martin; White, Sebastian; Whitehead, Samuel Robert; Whiteson, Daniel; Whittington, Denver; Wicek, Francois; Wicke, Daniel; Wickens, Fred; Wiedenmann, Werner; Wielers, Monika; Wienemann, Peter; Wiglesworth, Craig; Wiik, Liv Antje Mari; Wijeratne, Peter Alexander; Wildauer, Andreas; Wildt, Martin Andre; Wilhelm, Ivan; Wilkens, Henric George; Will, Jonas Zacharias; Williams, Eric; Williams, Hugh; Willis, William; Willocq, Stephane; Wilson, John; Wilson, Michael Galante; Wilson, Alan; Wingerter-Seez, Isabelle; Winkelmann, Stefan; Winklmeier, Frank; Wittgen, Matthias; Wolter, Marcin Wladyslaw; Wolters, Helmut; Wooden, Gemma; Wosiek, Barbara; Wotschack, Jorg; Woudstra, Martin; Wraight, Kenneth; Wright, Catherine; Wrona, Bozydar; Wu, Sau Lan; Wu, Xin; Wu, Yusheng; Wulf, Evan; Wunstorf, Renate; Wynne, Benjamin; Xaplanteris, Leonidas; Xella, Stefania; Xie, Song; Xie, Yigang; Xu, Chao; Xu, Da; Xu, Guofa; Yabsley, Bruce; Yamada, Miho; Yamamoto, Akira; Yamamoto, Kyoko; Yamamoto, Shimpei; Yamamura, Taiki; Yamaoka, Jared; Yamazaki, Takayuki; Yamazaki, Yuji; Yan, Zhen; Yang, Haijun; Yang, Un-Ki; Yang, Yi; Yang, Yi; Yang, Zhaoyu; Yanush, Serguei; Yao, Weiming; Yao, Yushu; Yasu, Yoshiji; Ybeles Smit, Gabriel Valentijn; Ye, Jingbo; Ye, Shuwei; Yilmaz, Metin; Yoosoofmiya, Reza; Yorita, Kohei; Yoshida, Riktura; Young, Charles; Youssef, Saul; Yu, Dantong; Yu, Jaehoon; Yu, Jie; Yuan, Li; Yurkewicz, Adam; Zaets, Vassilli; Zaidan, Remi; Zaitsev, Alexander; Zajacova, Zuzana; Zalite, Youris; Zanello, Lucia; Zarzhitsky, Pavel; Zaytsev, Alexander; Zeitnitz, Christian; Zeller, Michael; Zemla, Andrzej; Zendler, Carolin; Zenin, Anton; Zenin, Oleg; Ženiš, Tibor; Zenonos, Zenonas; Zenz, Seth; Zerwas, Dirk; Zevi della Porta, Giovanni; Zhan, Zhichao; Zhang, Dongliang; Zhang, Huaqiao; Zhang, Jinlong; Zhang, Xueyao; Zhang, Zhiqing; Zhao, Long; Zhao, Tianchi; Zhao, Zhengguo; Zhemchugov, Alexey; Zheng, Shuchen; Zhong, Jiahang; Zhou, Bing; Zhou, Ning; Zhou, Yue; Zhu, Cheng Guang; Zhu, Hongbo; Zhu, Yingchun; Zhuang, Xuai; Zhuravlov, Vadym; Zieminska, Daria; Zimmermann, Robert; Zimmermann, Simone; Zimmermann, Stephanie; Ziolkowski, Michael; Zitoun, Robert; Živković, Lidija; Zmouchko, Viatcheslav; Zobernig, Georg; Zoccoli, Antonio; Zolnierowski, Yves; Zsenei, Andras; zur Nedden, Martin; Zutshi, Vishnu; Zwalinski, Lukasz

    2012-01-03

    Proton-proton collisions at $\\sqrt{s}=7$ TeV and heavy ion collisions at $\\sqrt{s_{NN}}$=2.76 TeV were produced by the LHC and recorded using the ATLAS experiment's trigger system in 2010. The LHC is designed with a maximum bunch crossing rate of 40 MHz and the ATLAS trigger system is designed to record approximately 200 of these per second. The trigger system selects events by rapidly identifying signatures of muon, electron, photon, tau lepton, jet, and B meson candidates, as well as using global event signatures, such as missing transverse energy. An overview of the ATLAS trigger system, the evolution of the system during 2010 and the performance of the trigger system components and selections based on the 2010 collision data are shown. A brief outline of plans for the trigger system in 2011 is presented

  10. ATLAS Grid Workflow Performance Optimization

    CERN Document Server

    Elmsheuser, Johannes; The ATLAS collaboration

    2018-01-01

    The CERN ATLAS experiment grid workflow system manages routinely 250 to 500 thousand concurrently running production and analysis jobs to process simulation and detector data. In total more than 300 PB of data is distributed over more than 150 sites in the WLCG. At this scale small improvements in the software and computing performance and workflows can lead to significant resource usage gains. ATLAS is reviewing together with CERN IT experts several typical simulation and data processing workloads for potential performance improvements in terms of memory and CPU usage, disk and network I/O. All ATLAS production and analysis grid jobs are instrumented to collect many performance metrics for detailed statistical studies using modern data analytics tools like ElasticSearch and Kibana. This presentation will review and explain the performance gains of several ATLAS simulation and data processing workflows and present analytics studies of the ATLAS grid workflows.

  11. The ATLAS distributed analysis system

    International Nuclear Information System (INIS)

    Legger, F

    2014-01-01

    In the LHC operations era, analysis of the multi-petabyte ATLAS data sample by globally distributed physicists is a challenging task. To attain the required scale the ATLAS Computing Model was designed around the concept of Grid computing, realized in the Worldwide LHC Computing Grid (WLCG), the largest distributed computational resource existing in the sciences. The ATLAS experiment currently stores over 140 PB of data and runs about 140,000 concurrent jobs continuously at WLCG sites. During the first run of the LHC, the ATLAS Distributed Analysis (DA) service has operated stably and scaled as planned. More than 1600 users submitted jobs in 2012, with 2 million or more analysis jobs per week, peaking at about a million jobs per day. The system dynamically distributes popular data to expedite processing and maximally utilize resources. The reliability of the DA service is high and steadily improving; Grid sites are continually validated against a set of standard tests, and a dedicated team of expert shifters provides user support and communicates user problems to the sites. Both the user support techniques and the direct feedback of users have been effective in improving the success rate and user experience when utilizing the distributed computing environment. In this contribution a description of the main components, activities and achievements of ATLAS distributed analysis is given. Several future improvements being undertaken will be described.

  12. The ATLAS distributed analysis system

    Science.gov (United States)

    Legger, F.; Atlas Collaboration

    2014-06-01

    In the LHC operations era, analysis of the multi-petabyte ATLAS data sample by globally distributed physicists is a challenging task. To attain the required scale the ATLAS Computing Model was designed around the concept of Grid computing, realized in the Worldwide LHC Computing Grid (WLCG), the largest distributed computational resource existing in the sciences. The ATLAS experiment currently stores over 140 PB of data and runs about 140,000 concurrent jobs continuously at WLCG sites. During the first run of the LHC, the ATLAS Distributed Analysis (DA) service has operated stably and scaled as planned. More than 1600 users submitted jobs in 2012, with 2 million or more analysis jobs per week, peaking at about a million jobs per day. The system dynamically distributes popular data to expedite processing and maximally utilize resources. The reliability of the DA service is high and steadily improving; Grid sites are continually validated against a set of standard tests, and a dedicated team of expert shifters provides user support and communicates user problems to the sites. Both the user support techniques and the direct feedback of users have been effective in improving the success rate and user experience when utilizing the distributed computing environment. In this contribution a description of the main components, activities and achievements of ATLAS distributed analysis is given. Several future improvements being undertaken will be described.

  13. System administration of ATLAS TDAQ computing environment

    Science.gov (United States)

    Adeel-Ur-Rehman, A.; Bujor, F.; Benes, J.; Caramarcu, C.; Dobson, M.; Dumitrescu, A.; Dumitru, I.; Leahu, M.; Valsan, L.; Oreshkin, A.; Popov, D.; Unel, G.; Zaytsev, A.

    2010-04-01

    This contribution gives a thorough overview of the ATLAS TDAQ SysAdmin group activities which deals with administration of the TDAQ computing environment supporting High Level Trigger, Event Filter and other subsystems of the ATLAS detector operating on LHC collider at CERN. The current installation consists of approximately 1500 netbooted nodes managed by more than 60 dedicated servers, about 40 multi-screen user interface machines installed in the control rooms and various hardware and service monitoring machines as well. In the final configuration, the online computer farm will be capable of hosting tens of thousands applications running simultaneously. The software distribution requirements are matched by the two level NFS based solution. Hardware and network monitoring systems of ATLAS TDAQ are based on NAGIOS and MySQL cluster behind it for accounting and storing the monitoring data collected, IPMI tools, CERN LANDB and the dedicated tools developed by the group, e.g. ConfdbUI. The user management schema deployed in TDAQ environment is founded on the authentication and role management system based on LDAP. External access to the ATLAS online computing facilities is provided by means of the gateways supplied with an accounting system as well. Current activities of the group include deployment of the centralized storage system, testing and validating hardware solutions for future use within the ATLAS TDAQ environment including new multi-core blade servers, developing GUI tools for user authentication and roles management, testing and validating 64-bit OS, and upgrading the existing TDAQ hardware components, authentication servers and the gateways.

  14. Advanced Alignment of the ATLAS Tracking System

    CERN Document Server

    Pedraza Lopez, S; The ATLAS collaboration

    2012-01-01

    In order to reconstruct trajectories of charged particles, ATLAS is equipped with a tracking system built using different technologies embedded in a 2T solenoidal magnetic field. ATLAS physics goals require high resolution, unbiased measurement of all charged particle kinematic parameters in order to assure accurate invariant mass reconstruction and interaction and decay vertex finding. These critically depend on the systematic effects related to the alignment of the tracking system. In order to eliminate malicious systematic deformations, various advanced tools and techniques have been put in place. These include information from known mass resonances, energy of electrons and positrons measured by the electromagnetic calorimeters, etc. Despite being stable under normal running conditions, ATLAS tracking system responses to sudden environ-mental changes (temperature, magnetic field) by small collective deformations. These have to be identified and corrected in order to assure uniform, highest quality tracking...

  15. ATLAS TileCal Read Out Driver production

    International Nuclear Information System (INIS)

    Valero, A; Abdallah, J; Castillo, V; Cuenca, C; Ferrer, A; Fullana, E; Gonzalez, V; Higon, E; Poveda, J; Ruiz-MartInez, A; Saez, M A; Salvachua, B; SanchIs, E; Solans, C; Valls, J A

    2007-01-01

    The production tests of the 38 ATLAS TileCal Read Out Drivers (RODs) are presented in this paper. The hardware specifications and firmware functionality of the RODs modules, the test-bench and the test procedure to qualify the boards are described. Finally the performance results, the temperature studies and high rate tests are shown and discussed

  16. Migration of ATLAS PanDA to CERN

    International Nuclear Information System (INIS)

    Stewart, Graeme Andrew; Klimentov, Alexei; Maeno, Tadashi; Nevski, Pavel; Nowak, Marcin; De Castro Faria Salgado, Pedro Emanuel; Wenaus, Torre; Koblitz, Birger; Lamanna, Massimo

    2010-01-01

    The ATLAS Production and Distributed Analysis System (PanDA) is a key component of the ATLAS distributed computing infrastructure. All ATLAS production jobs, and a substantial amount of user and group analysis jobs, pass through the PanDA system, which manages their execution on the grid. PanDA also plays a key role in production task definition and the data set replication request system. PanDA has recently been migrated from Brookhaven National Laboratory (BNL) to the European Organization for Nuclear Research (CERN), a process we describe here. We discuss how the new infrastructure for PanDA, which relies heavily on services provided by CERN IT, was introduced in order to make the service as reliable as possible and to allow it to be scaled to ATLAS's increasing need for distributed computing. The migration involved changing the backend database for PanDA from MySQL to Oracle, which impacted upon the database schemas. The process by which the client code was optimised for the new database backend is discussed. We describe the procedure by which the new database infrastructure was tested and commissioned for production use. Operations during the migration had to be planned carefully to minimise disruption to ongoing ATLAS offline computing. All parts of the migration were fully tested before commissioning the new infrastructure and the gradual migration of computing resources to the new system allowed any problems of scaling to be addressed.

  17. The ATLAS Eventlndex: data flow and inclusion of other metadata

    Science.gov (United States)

    Barberis, D.; Cárdenas Zárate, S. E.; Favareto, A.; Fernandez Casani, A.; Gallas, E. J.; Garcia Montoro, C.; Gonzalez de la Hoz, S.; Hrivnac, J.; Malon, D.; Prokoshin, F.; Salt, J.; Sanchez, J.; Toebbicke, R.; Yuan, R.; ATLAS Collaboration

    2016-10-01

    The ATLAS EventIndex is the catalogue of the event-related metadata for the information collected from the ATLAS detector. The basic unit of this information is the event record, containing the event identification parameters, pointers to the files containing this event as well as trigger decision information. The main use case for the EventIndex is event picking, as well as data consistency checks for large production campaigns. The EventIndex employs the Hadoop platform for data storage and handling, as well as a messaging system for the collection of information. The information for the EventIndex is collected both at Tier-0, when the data are first produced, and from the Grid, when various types of derived data are produced. The EventIndex uses various types of auxiliary information from other ATLAS sources for data collection and processing: trigger tables from the condition metadata database (COMA), dataset information from the data catalogue AMI and the Rucio data management system and information on production jobs from the ATLAS production system. The ATLAS production system is also used for the collection of event information from the Grid jobs. EventIndex developments started in 2012 and in the middle of 2015 the system was commissioned and started collecting event metadata, as a part of ATLAS Distributed Computing operations.

  18. Testbeam studies of production modules of the ATLAS Tile Calorimeter

    International Nuclear Information System (INIS)

    Adragna, P.; Alexa, C.; Anderson, K.; Antonaki, A.; Arabidze, A.; Batkova, L.; Batusov, V.; Beck, H.P.; Bednar, P.; Bergeaas Kuutmann, E.; Biscarat, C.; Blanchot, G.; Bogush, A.; Bohm, C.; Boldea, V.; Bosman, M.; Bromberg, C.; Budagov, J.; Burckhart-Chromek, D.; Caprini, M.

    2009-01-01

    We report test beam studies of 11% of the production ATLAS Tile Calorimeter modules. The modules were equipped with production front-end electronics and all the calibration systems planned for the final detector. The studies used muon, electron and hadron beams ranging in energy from 3 to 350 GeV. Two independent studies showed that the light yield of the calorimeter was ∼70pe/GeV, exceeding the design goal by 40%. Electron beams provided a calibration of the modules at the electromagnetic energy scale. Over 200 calorimeter cells the variation of the response was 2.4%. The linearity with energy was also measured. Muon beams provided an intercalibration of the response of all calorimeter cells. The response to muons entering in the ATLAS projective geometry showed an RMS variation of 2.5% for 91 measurements over a range of rapidities and modules. The mean response to hadrons of fixed energy had an RMS variation of 1.4% for the modules and projective angles studied. The response to hadrons normalized to incident beam energy showed an 8% increase between 10 and 350 GeV, fully consistent with expectations for a noncompensating calorimeter. The measured energy resolution for hadrons of σ/E=52.9%/√(E)+5.7% was also consistent with expectations. Other auxiliary studies were made of saturation recovery of the readout system, the time resolution of the calorimeter and the performance of the trigger signals from the calorimeter.

  19. Testbeam studies of production modules of the ATLAS Tile Calorimeter

    Energy Technology Data Exchange (ETDEWEB)

    Adragna, P [Pisa University and INFN, Pisa (Italy); Alexa, C [National Institute for Physics and Nuclear Engineering, Bucharest (Romania); Anderson, K [University of Chicago, Chicago, Illinois (United States); Antonaki, A; Arabidze, A [University of Athens, Athens (Greece); Batkova, L [Comenius University, Bratislava (Slovakia); Batusov, V [JINR, Dubna (Russian Federation); Beck, H P [Laboratory for High Energy Physics, University of Bern (Switzerland); Bednar, P [Comenius University, Bratislava (Slovakia); Bergeaas Kuutmann, E [Stockholm University, Stockholm (Sweden); Biscarat, C [LPC Clermont-Ferrand, Universite Blaise Pascal, Clermont-Ferrand (France); Blanchot, G [Institut de Fisica d' Altes Energies, Universitat Autonoma de Barcelona, Barcelona (Spain); Bogush, A [Institute of Physics, National Academy of Sciences, Minsk (Belarus); Bohm, C [Stockholm University, Stockholm (Sweden); Boldea, V [National Institute for Physics and Nuclear Engineering, Bucharest (Romania); Bosman, M [Institut de Fisica d' Altes Energies, Universitat Autonoma de Barcelona, Barcelona (Spain); Bromberg, C [Michigan State University, East Lansing, Michigan (United States); Budagov, J [JINR, Dubna (Russian Federation); Burckhart-Chromek, D [CERN, Geneva (Switzerland); Caprini, M [National Institute for Physics and Nuclear Engineering, Bucharest (Romania)

    2009-07-21

    We report test beam studies of 11% of the production ATLAS Tile Calorimeter modules. The modules were equipped with production front-end electronics and all the calibration systems planned for the final detector. The studies used muon, electron and hadron beams ranging in energy from 3 to 350 GeV. Two independent studies showed that the light yield of the calorimeter was {approx}70pe/GeV, exceeding the design goal by 40%. Electron beams provided a calibration of the modules at the electromagnetic energy scale. Over 200 calorimeter cells the variation of the response was 2.4%. The linearity with energy was also measured. Muon beams provided an intercalibration of the response of all calorimeter cells. The response to muons entering in the ATLAS projective geometry showed an RMS variation of 2.5% for 91 measurements over a range of rapidities and modules. The mean response to hadrons of fixed energy had an RMS variation of 1.4% for the modules and projective angles studied. The response to hadrons normalized to incident beam energy showed an 8% increase between 10 and 350 GeV, fully consistent with expectations for a noncompensating calorimeter. The measured energy resolution for hadrons of {sigma}/E=52.9%/{radical}(E)+5.7% was also consistent with expectations. Other auxiliary studies were made of saturation recovery of the readout system, the time resolution of the calorimeter and the performance of the trigger signals from the calorimeter.

  20. Integrating Networking into ATLAS

    CERN Document Server

    Mc Kee, Shawn Patrick; The ATLAS collaboration

    2018-01-01

    Networking is foundational to the ATLAS distributed infrastructure and there are many ongoing activities related to networking both within and outside of ATLAS. We will report on the progress in a number of areas exploring ATLAS's use of networking and our ability to monitor the network, analyze metrics from the network, and tune and optimize application and end-host parameters to make the most effective use of the network. Specific topics will include work on Open vSwitch for production systems, network analytics, FTS testing and tuning, and network problem alerting and alarming.

  1. ATLAS + CMS top production and properties: Run 1 legacy

    CERN Document Server

    AUTHOR|(SzGeCERN)641905; The ATLAS collaboration

    2015-01-01

    The large Run 1 data sample of top-quark events collected at the Large Hadron Collider allows a variety of measurements to analyse the production and properties of the top quark. Measurements of top-quark production cross sections and top-quark properties in proton-proton collisions with the ATLAS and CMS detectors at the LHC are presented.

  2. Production and integration of the ATLAS Insertable B-Layer

    Science.gov (United States)

    Abbott, B.; Albert, J.; Alberti, F.; Alex, M.; Alimonti, G.; Alkire, S.; Allport, P.; Altenheiner, S.; Ancu, L. S.; Anderssen, E.; Andreani, A.; Andreazza, A.; Axen, B.; Arguin, J.; Backhaus, M.; Balbi, G.; Ballansat, J.; Barbero, M.; Barbier, G.; Bassalat, A.; Bates, R.; Baudin, P.; Battaglia, M.; Beau, T.; Beccherle, R.; Bell, A.; Benoit, M.; Bermgan, A.; Bertsche, C.; Bertsche, D.; Bilbao de Mendizabal, J.; Bindi, F.; Bomben, M.; Borri, M.; Bortolin, C.; Bousson, N.; Boyd, R. G.; Breugnon, P.; Bruni, G.; Brossamer, J.; Bruschi, M.; Buchholz, P.; Budun, E.; Buttar, C.; Cadoux, F.; Calderini, G.; Caminada, L.; Capeans, M.; Carney, R.; Casse, G.; Catinaccio, A.; Cavalli-Sforza, M.; Červ, M.; Cervelli, A.; Chau, C. C.; Chauveau, J.; Chen, S. P.; Chu, M.; Ciapetti, M.; Cindro, V.; Citterio, M.; Clark, A.; Cobal, M.; Coelli, S.; Collot, J.; Crespo-Lopez, O.; Dalla Betta, G. F.; Daly, C.; D'Amen, G.; Dann, N.; Dao, V.; Darbo, G.; DaVia, C.; David, P.; Debieux, S.; Delebecque, P.; De Lorenzi, F.; de Oliveira, R.; Dette, K.; Dietsche, W.; Di Girolamo, B.; Dinu, N.; Dittus, F.; Diyakov, D.; Djama, F.; Dobos, D.; Dondero, P.; Doonan, K.; Dopke, J.; Dorholt, O.; Dube, S.; Dzahini, D.; Egorov, K.; Ehrmann, O.; Einsweiler, K.; Elles, S.; Elsing, M.; Eraud, L.; Ereditato, A.; Eyring, A.; Falchieri, D.; Falou, A.; Fausten, C.; Favareto, A.; Favre, Y.; Feigl, S.; Fernandez Perez, S.; Ferrere, D.; Fleury, J.; Flick, T.; Forshaw, D.; Fougeron, D.; Franconi, L.; Gabrielli, A.; Gaglione, R.; Gallrapp, C.; Gan, K. K.; Garcia-Sciveres, M.; Gariano, G.; Gastaldi, T.; Gavrilenko, I.; Gaudiello, A.; Geffroy, N.; Gemme, C.; Gensolen, F.; George, M.; Ghislain, P.; Giangiacomi, N.; Gibson, S.; Giordani, M. P.; Giugni, D.; Gjersdal, H.; Glitza, K. W.; Gnani, D.; Godlewski, J.; Gonella, L.; Gonzalez-Sevilla, S.; Gorelov, I.; Gorišek, A.; Gössling, C.; Grancagnolo, S.; Gray, H.; Gregor, I.; Grenier, P.; Grinstein, S.; Gris, A.; Gromov, V.; Grondin, D.; Grosse-Knetter, J.; Guescini, F.; Guido, E.; Gutierrez, P.; Hallewell, G.; Hartman, N.; Hauck, S.; Hasi, J.; Hasib, A.; Hegner, F.; Heidbrink, S.; Heim, T.; Heinemann, B.; Hemperek, T.; Hessey, N. P.; Hetmánek, M.; Hinman, R. R.; Hoeferkamp, M.; Holmes, T.; Hostachy, J.; Hsu, S. C.; Hügging, F.; Husi, C.; Iacobucci, G.; Ibragimov, I.; Idarraga, J.; Ikegami, Y.; Ince, T.; Ishmukhametov, R.; Izen, J. M.; Janoška, Z.; Janssen, J.; Jansen, L.; Jeanty, L.; Jensen, F.; Jentzsch, J.; Jezequel, S.; Joseph, J.; Kagan, H.; Kagan, M.; Karagounis, M.; Kass, R.; Kastanas, A.; Kenney, C.; Kersten, S.; Kind, P.; Klein, M.; Klingenberg, R.; Kluit, R.; Kocian, M.; Koffeman, E.; Korchak, O.; Korolkov, I.; Kostyukhina-Visoven, I.; Kovalenko, S.; Kretz, M.; Krieger, N.; Krüger, H.; Kruth, A.; Kugel, A.; Kuykendall, W.; La Rosa, A.; Lai, C.; Lantzsch, K.; Lapoire, C.; Laporte, D.; Lari, T.; Latorre, S.; Leyton, M.; Lindquist, B.; Looper, K.; Lopez, I.; Lounis, A.; Lu, Y.; Lubatti, H. J.; Maeland, S.; Maier, A.; Mallik, U.; Manca, F.; Mandelli, B.; Mandić, I.; Marchand, D.; Marchiori, G.; Marx, M.; Massol, N.; Mättig, P.; Mayer, J.; McGoldrick, G.; Mekkaoui, A.; Menouni, M.; Menu, J.; Meroni, C.; Mesa, J.; Michal, S.; Miglioranzi, S.; Mikuž, M.; Miucci, A.; Mochizuki, K.; Monti, M.; Moore, J.; Morettini, P.; Morley, A.; Moss, J.; Muenstermann, D.; Murray, P.; Nakamura, K.; Nellist, C.; Nelson, D.; Nessi, M.; Nisius, R.; Nordberg, M.; Nuiry, F.; Obermann, T.; Ockenfels, W.; Oide, H.; Oriunno, M.; Ould-Saada, F.; Padilla, C.; Pangaud, P.; Parker, S.; Pelleriti, G.; Pernegger, H.; Piacquadio, G.; Picazio, A.; Pohl, D.; Polini, A.; Pons, X.; Popule, J.; Portell Bueso, X.; Potamianos, K.; Povoli, M.; Puldon, D.; Pylypchenko, Y.; Quadt, A.; Quayle, B.; Rarbi, F.; Ragusa, F.; Rambure, T.; Richards, E.; Riegel, C.; Ristic, B.; Rivière, F.; Rizatdinova, F.; RØhne, O.; Rossi, C.; Rossi, L. P.; Rovani, A.; Rozanov, A.; Rubinskiy, I.; Rudolph, M. S.; Rummler, A.; Ruscino, E.; Sabatini, F.; Salek, D.; Salzburger, A.; Sandaker, H.; Sannino, M.; Sanny, B.; Scanlon, T.; Schipper, J.; Schmidt, U.; Schneider, B.; Schorlemmer, A.; Schroer, N.; Schwemling, P.; Sciuccati, A.; Seidel, S.; Seiden, A.; Šícho, P.; Skubic, P.; Sloboda, M.; Smith, D. S.; Smith, M.; Sood, A.; Spencer, E.; Stramaglia, M.; Strauss, M.; Stucci, S.; Stugu, B.; Stupak, J.; Styles, N.; Su, D.; Takubo, Y.; Tassan, J.; Teng, P.; Teixeira, A.; Terzo, S.; Therry, X.; Todorov, T.; Tomášek, M.; Toms, K.; Travaglini, R.; Trischuk, W.; Troncon, C.; Troska, G.; Tsiskaridze, S.; Tsurin, I.; Tsybychev, D.; Unno, Y.; Vacavant, L.; Verlaat, B.; Vigeolas, E.; Vogt, M.; Vrba, V.; Vuillermet, R.; Wagner, W.; Walkowiak, W.; Wang, R.; Watts, S.; Weber, M. S.; Weber, M.; Weingarten, J.; Welch, S.; Wenig, S.; Wensing, M.; Wermes, N.; Wittig, T.; Wittgen, M.; Yildizkaya, T.; Yang, Y.; Yao, W.; Yi, Y.; Zaman, A.; Zaidan, R.; Zeitnitz, C.; Ziolkowski, M.; Zivkovic, V.; Zoccoli, A.; Zwalinski, L.

    2018-05-01

    During the shutdown of the CERN Large Hadron Collider in 2013-2014, an additional pixel layer was installed between the existing Pixel detector of the ATLAS experiment and a new, smaller radius beam pipe. The motivation for this new pixel layer, the Insertable B-Layer (IBL), was to maintain or improve the robustness and performance of the ATLAS tracking system, given the higher instantaneous and integrated luminosities realised following the shutdown. Because of the extreme radiation and collision rate environment, several new radiation-tolerant sensor and electronic technologies were utilised for this layer. This paper reports on the IBL construction and integration prior to its operation in the ATLAS detector.

  3. ATLAS TDAQ System Administration: an overview and evolution

    CERN Document Server

    LEE, CJ; The ATLAS collaboration; BOGDANCHIKOV, A; BRASOLIN, F; CONTESCU, AC; DARLEA, GL; KOROL, A; SCANNICCHIO, DA; TWOMEY, M; VALSAN, ML

    2013-01-01

    The ATLAS Trigger and Data Acquisition (TDAQ) system is responsible for the online processing of live data streaming from the ATLAS experiment at the Large Hadron Collider (LHC) at CERN. The system processes the direct data readout from ~100 million channels on the detector through three trigger levels, selecting interesting events for analysis with a factor of 10^7 reduction on the data rate with a latency of less than a few seconds. Most of the functionality is implemented on ~3000 servers composing the online farm. Due to the critical functionality of the system a sophisticated computing environment is maintained, covering the online farm and ATLAS control rooms, as well as a number of development and testing labs. The specificity of the system required the development of dedicated applications (e.g. ConfDB, BWM) for system configuration and maintenance; in parallel other Open Source tools (Puppet and Quattor) are used to centrally configure the operating systems. The health monitoring of the TDAQ system h...

  4. Test system for the production of the Atlas Tile Calorimeter front-end electronics

    International Nuclear Information System (INIS)

    Calvet, David

    2004-01-01

    The Atlas hadronic Tile Calorimeter front-end electronics is fully included in the so-called 'super-drawers'. The 256 super-drawers needed for the entire calorimeter are assembled and extensively tested in Clermont-Ferrand before being sent to CERN to be inserted in the calorimeter modules. A mobile system has been developed to perform a complete test of the super-drawers during their insertion

  5. Scaling up ATLAS Event Service to production levels on opportunistic computing platforms

    CERN Document Server

    Benjamin, Douglas; The ATLAS collaboration; Ernst, Michael; Guan, Wen; Hover, John; Lesny, David; Maeno, Tadashi; Nilsson, Paul; Tsulaia, Vakhtang; van Gemmeren, Peter; Vaniachine, Alexandre; Wang, Fuquan; Wenaus, Torre

    2016-01-01

    Continued growth in public cloud and HPC resources is on track to overcome the dedicated resources available for ATLAS on the WLCG. Example of such platforms are Amazon AWS EC2 Spot Instances, Edison Cray XC30 supercomputer, backfill at the Tier-2 and Tier-3 sites, opportunistic resources at the Open Science Grid, and ATLAS High Level Trigger farm between the data taking periods. Because of opportunistic resources specifics such as preemptive job scheduling and data I/O, their efficient usage requires workflow innovations provided by the ATLAS Event Service. Thanks to the finer granularity of the Event Service data processing workflow, the opportunistic resources are used more efficiently. We report on our progress in scaling opportunistic resource usage to double-digit levels in ATLAS production.

  6. ATLAS TDAQ System Administration: an overview and evolution

    CERN Document Server

    LEE, CJ; The ATLAS collaboration; BOGDANCHIKOV, A; BRASOLIN, F; CONTESCU, AC; DARLEA, G-L; KOROL, A; SCANNICCHIO, DA; TWOMEY, M; VALSAN, ML

    2013-01-01

    The ATLAS Trigger and Data Acquisition (TDAQ) system is responsible for the online processing of live data streaming from the ATLAS experiment at the Large Hadron Collider (LHC) at CERN. The system processes the direct data readout from ~100 million channels on the detector through multiple trigger levels, selecting interesting events for analysis with a factor of $10^{7}$ reduction on the data rate with a latency of less than a few seconds. Most of the functionality is implemented on ~3000 servers composing the online farm. Due to the critical functionality of the system a sophisticated computing environment is maintained, covering the online farm and ATLAS control rooms, as well as a number of development and testing labs. The specificity of the system required the development of dedicated applications (e.g. ConfDB, BWM) for system configuration and maintenance; in parallel other Open Source tools (Puppet and Quattor) are used to centrally configure the operating systems. The health monitoring of the TDAQ s...

  7. Boosted top production in ATLAS and CMS

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00237277; The ATLAS collaboration

    2017-01-01

    An overview of the boosted top production analyses using data collected by the ATLAS and CMS experiments at $\\sqrt{s}=$' 8 TeV and 13 TeV of proton-proton collisions at the LHC is presented. These analyses use techniques for the reconstruction of boosted objects to measure the production of top quarks at high transverse momenta. The measurements are optimized for the different final states and for different ranges of the transverse momenta of the particles involved, improving on measurements with traditional objects reconstruction based on the combination of resolved objects.

  8. AGIS: Integration of new technologies used in ATLAS Distributed Computing

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00291854; The ATLAS collaboration; Di Girolamo, Alessandro; Alandes Pradillo, Maria

    2017-01-01

    The variety of the ATLAS Distributed Computing infrastructure requires a central information system to define the topology of computing resources and to store different parameters and configuration data which are needed by various ATLAS software components. The ATLAS Grid Information System (AGIS) is the system designed to integrate configuration and status information about resources, services and topology of the computing infrastructure used by ATLAS Distributed Computing applications and services. Being an intermediate middleware system between clients and external information sources (like central BDII, GOCDB, MyOSG), AGIS defines the relations between experiment specific used resources and physical distributed computing capabilities. Being in production during LHC Runl AGIS became the central information system for Distributed Computing in ATLAS and it is continuously evolving to fulfil new user requests, enable enhanced operations and follow the extension of the ATLAS Computing model. The ATLAS Computin...

  9. Multi-Threaded Evolution of the Data-Logging System of the ATLAS Experiment at CERN

    CERN Document Server

    Colombo, T; The ATLAS collaboration

    2011-01-01

    The ATLAS experiment is currently observing proton-proton collisions delivered by the LHC accelerator at a centre of mass energy of 7 TeV with a peak luminosity of ~1033 cm-2 s-1. The ATLAS Trigger and Data Acquisition (TDAQ) system selects interesting events on-line in a three-level trigger system in order to store them at a budgeted rate of ~200 Hz for an event size of ~1.5 MB. This paper focuses on the TDAQ data-logging system. Its purpose is to receive events from the third level trigger, process them and stream the results into different raw data files according to the trigger decision. The data files are subsequently moved to the central mass storage facility at CERN. The system currently in production has been commissioned in 2007 and has been working smoothly since then. It is however based on an essentially single-threaded design that is anticipated not to cope with the increase in event rate and event size that is foreseen as part of the ATLAS and LHC upgrade programs. This design also severely limi...

  10. Multi-Threaded Evolution of the Data-Logging System of the ATLAS Experiment at CERN

    CERN Document Server

    Colombo, T; The ATLAS collaboration

    2011-01-01

    The ATLAS experiment observes proton-proton collisions delivered by the LHC accelerator at a centre of mass energy of 7 TeV with a peak luminosity of ~ 10^33 cm^-2 s^-1 in 2011. The ATLAS Trigger and Data Acquisition (TDAQ) system selects interesting events on-line in a three-level trigger system in order to store them at a budgeted average rate of ~ 400 Hz for an event size of ~1.2 MB. This paper focuses on the TDAQ data-logging system. Its purpose is to receive events from the third level trigger, process them and stream the data into different raw files according to the trigger decision. The system currently in production is based on an essentially single-threaded design that is anticipated not to cope with the increase in event rate and event size foreseen as part of the ATLAS and LHC upgrade programs. This design also severely limits the possibility of performing additional CPU-intensive tasks. Therefore, a novel design able to exploit the full power of multi-core architecture is needed. The main challen...

  11. Overview of ATLAS PanDA Workload Management

    Science.gov (United States)

    Maeno, T.; De, K.; Wenaus, T.; Nilsson, P.; Stewart, G. A.; Walker, R.; Stradling, A.; Caballero, J.; Potekhin, M.; Smith, D.; ATLAS Collaboration

    2011-12-01

    The Production and Distributed Analysis System (PanDA) plays a key role in the ATLAS distributed computing infrastructure. All ATLAS Monte-Carlo simulation and data reprocessing jobs pass through the PanDA system. We will describe how PanDA manages job execution on the grid using dynamic resource estimation and data replication together with intelligent brokerage in order to meet the scaling and automation requirements of ATLAS distributed computing. PanDA is also the primary ATLAS system for processing user and group analysis jobs, bringing further requirements for quick, flexible adaptation to the rapidly evolving analysis use cases of the early datataking phase, in addition to the high reliability, robustness and usability needed to provide efficient and transparent utilization of the grid for analysis users. We will describe how PanDA meets ATLAS requirements, the evolution of the system in light of operational experience, how the system has performed during the first LHC data-taking phase and plans for the future.

  12. Overview of ATLAS PanDA Workload Management

    International Nuclear Information System (INIS)

    Maeno, T.; De, K.; Wenaus, T.; Nilsson, P.; Stewart, G.A.; Walker, R.; Stradling, A.; Caballero, J.; Potekhin, M.; Smith, D.

    2011-01-01

    The Production and Distributed Analysis System (PanDA) plays a key role in the ATLAS distributed computing infrastructure. All ATLAS Monte-Carlo simulation and data reprocessing jobs pass through the PanDA system. We will describe how PanDA manages job execution on the grid using dynamic resource estimation and data replication together with intelligent brokerage in order to meet the scaling and automation requirements of ATLAS distributed computing. PanDA is also the primary ATLAS system for processing user and group analysis jobs, bringing further requirements for quick, flexible adaptation to the rapidly evolving analysis use cases of the early datataking phase, in addition to the high reliability, robustness and usability needed to provide efficient and transparent utilization of the grid for analysis users. We will describe how PanDA meets ATLAS requirements, the evolution of the system in light of operational experience, how the system has performed during the first LHC data-taking phase and plans for the future.

  13. Top quark pair production in ATLAS

    CERN Document Server

    Moreno Llacer, M; The ATLAS collaboration

    2010-01-01

    Top-quark pairs are expected to be produced at the LHC, even at the lower beam energy and luminosity in the first years of running. Establishing the top-pair signal and measuring the production cross-section are important benchmarks for ATLAS, and will help understand the detector performance for events with high-pT leptons, high jet multiplicity, missing transverse energy. The prospects for early top physics measurements will be shown, with a particular emphasis on the progress achieved with data so far.

  14. ATLAS TDAQ System Administration: evolution and re-design

    CERN Document Server

    Ballestrero, Sergio; The ATLAS collaboration; Brasolin, Franco; Contescu, Alexandru Cristian; Dubrov, Sergei; Fazio, Daniel; Korol, Aleksandr; Lee, Christopher Jon; Scannicchio, Diana; Twomey, Matthew Shaun

    2015-01-01

    The ATLAS Trigger and Data Acquisition (TDAQ) system is responsible for the online processing of live data, streaming from the ATLAS experiment at the Large Hadron Collider (LHC) at CERN. The online farm is composed of $\\sim 3000$ servers, processing the data readout from $\\sim 100$ million detector channels through multiple trigger levels. During the two years of the first Long Shutdown (LS1) there has been a tremendous amount of work done by the ATLAS TDAQ System Administrators, implementing numerous new software applications, upgrading the OS and the hardware, changing some design philosophies and exploiting the High Level Trigger farm with different purposes. The OS version has been upgraded to SLC6; for the largest part of the farm, which is composed by net booted nodes, this required a completely new design of the net booting system. In parallel, the migration to Puppet of the Configuration Management systems has been completed for both net booted and local booted hosts; the Post-Boot Scripts system and...

  15. System Architecture Modeling for Technology Portfolio Management using ATLAS

    Science.gov (United States)

    Thompson, Robert W.; O'Neil, Daniel A.

    2006-01-01

    Strategic planners and technology portfolio managers have traditionally relied on consensus-based tools, such as Analytical Hierarchy Process (AHP) and Quality Function Deployment (QFD) in planning the funding of technology development. While useful to a certain extent, these tools are limited in the ability to fully quantify the impact of a technology choice on system mass, system reliability, project schedule, and lifecycle cost. The Advanced Technology Lifecycle Analysis System (ATLAS) aims to provide strategic planners a decision support tool for analyzing technology selections within a Space Exploration Architecture (SEA). Using ATLAS, strategic planners can select physics-based system models from a library, configure the systems with technologies and performance parameters, and plan the deployment of a SEA. Key parameters for current and future technologies have been collected from subject-matter experts and other documented sources in the Technology Tool Box (TTB). ATLAS can be used to compare the technical feasibility and economic viability of a set of technology choices for one SEA, and compare it against another set of technology choices or another SEA. System architecture modeling in ATLAS is a multi-step process. First, the modeler defines the system level requirements. Second, the modeler identifies technologies of interest whose impact on an SEA. Third, the system modeling team creates models of architecture elements (e.g. launch vehicles, in-space transfer vehicles, crew vehicles) if they are not already in the model library. Finally, the architecture modeler develops a script for the ATLAS tool to run, and the results for comparison are generated.

  16. ATLAS Distributed Analysis Tools

    CERN Document Server

    Gonzalez de la Hoz, Santiago; Liko, Dietrich

    2008-01-01

    The ATLAS production system has been successfully used to run production of simulation data at an unprecedented scale. Up to 10000 jobs were processed in one day. The experiences obtained operating the system on several grid flavours was essential to perform a user analysis using grid resources. First tests of the distributed analysis system were then performed. In the preparation phase data was registered in the LHC File Catalog (LFC) and replicated in external sites. For the main test, few resources were used. All these tests are only a first step towards the validation of the computing model. The ATLAS management computing board decided to integrate the collaboration efforts in distributed analysis in only one project, GANGA. The goal is to test the reconstruction and analysis software in a large scale Data production using Grid flavors in several sites. GANGA allows trivial switching between running test jobs on a local batch system and running large-scale analyses on the Grid; it provides job splitting a...

  17. Scaling up ATLAS Event Service to production levels on opportunistic computing platforms

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00066086; The ATLAS collaboration; Caballero, Jose; Ernst, Michael; Guan, Wen; Hover, John; Lesny, David; Maeno, Tadashi; Nilsson, Paul; Tsulaia, Vakhtang; van Gemmeren, Peter; Vaniachine, Alexandre; Wang, Fuquan; Wenaus, Torre

    2016-01-01

    Continued growth in public cloud and HPC resources is on track to exceed the dedicated resources available for ATLAS on the WLCG. Examples of such platforms are Amazon AWS EC2 Spot Instances, Edison Cray XC30 supercomputer, backfill at Tier 2 and Tier 3 sites, opportunistic resources at the Open Science Grid (OSG), and ATLAS High Level Trigger farm between the data taking periods. Because of specific aspects of opportunistic resources such as preemptive job scheduling and data I/O, their efficient usage requires workflow innovations provided by the ATLAS Event Service. Thanks to the finer granularity of the Event Service data processing workflow, the opportunistic resources are used more efficiently. We report on our progress in scaling opportunistic resource usage to double-digit levels in ATLAS production.

  18. The next generation of the ATLAS PanDA Monitoring System

    CERN Document Server

    Schovancova, J; The ATLAS collaboration; Klimentov, A; Love, P; Potekhin, M; Wenaus, T

    2014-01-01

    For many years the PanDA Workload Management System has been the basis for distributed production and analysis for the ATLAS experiment at the LHC. Since the start of data taking PanDA usage has ramped up steadily, with up to 1M completed jobs/day in 2013. The associated monitoring data volume has been rising as well, to levels that present a new set of challenges in the areas of database scalability and monitoring system performance and efficiency. Outside of ATLAS, the PanDA system is also being used in projects like AMS, LSST and a few others. It currently undergoes a significant redesign, both of the core server components responsible for workload management, brokerage and data access, and of the monitoring part, which is critically important for efficient execution of the workflow in a way that’s transparent to the user and also provides an effective set of tools for operational support. The new generation of the PanDA Monitoring Service is designed based on a proven, scalable, industry-standard Web Fr...

  19. ATLAS : magnet industrial production Conference MT17

    CERN Multimedia

    2001-01-01

    With overall dimensions of 26 meters in length and 20 meters in diameter, the ATLAS magnet system is the largest integrated superconducting magnet ever built. The system is made up of four super-conducting magnets, a power supply, cryogenics, vacuum, control, and safety systems. The coils are built with Aluminum stabilized NbTi/Cu superconductor indirectly cooled at 4.5 K by liquid Helium forced flow.

  20. ATLAS BigPanDA Monitoring and Its Evolution

    CERN Document Server

    Wenaus, Torre; The ATLAS collaboration; Korchuganova, Tatiana

    2016-01-01

    BigPanDA is the latest generation of the monitoring system for the Production and Distributed Analysis (PanDA) system. The BigPanDA monitor is a core component of PanDA and also serves the monitoring needs of the new ATLAS Production System Prodsys-2. BigPanDA has been developed to serve the growing computation needs of the ATLAS Experiment and the wider applications of PanDA beyond ATLAS. Through a system-wide job database, the BigPanDA monitor provides a comprehensive and coherent view of the tasks and jobs executed by the system, from high level summaries to detailed drill-down job diagnostics. The system has been in production and has remained in continuous development since mid 2014, today effectively managing more than 2 million jobs per day distributed over 150 computing centers worldwide. BigPanDA also delivers web-based analytics and system state views to groups of users including distributed computing systems operators, shifters, physicist end-users, computing managers and accounting services. Provi...

  1. Single Top quark production cross section using ATLAS detector at the LHC

    CERN Document Server

    Estrada Pastor, Oscar; The ATLAS collaboration

    2018-01-01

    Measurements of single top-quark production in proton-proton collisions are presented based on the 8 TeV and 13 TeV ATLAS datasets. In the leading order process, a W boson is exchanged in the t-channel. The cross-section for the production of single top-quarks and single anti-top-quarks, their ratio, as well as differential cross-section measurements are also reported. These analyses include limits on anomalous contributions to the Wtb vertex and measurement of the top quark polarization. Measurements of the inclusive and differential cross-sections for the production of a single top quark in association with a W boson, the second largest single-top production mode, are also presented. Finally, evidence for s-channel single-top production in the 8 TeV ATLAS dataset is presented. All measurements are compared to state-of-the-art theoretical calculations.

  2. ATLAS Tile Calorimeter calibration and monitoring systems

    Science.gov (United States)

    Cortés-González, Arely

    2018-01-01

    The ATLAS Tile Calorimeter is the central section of the hadronic calorimeter of the ATLAS experiment and provides important information for reconstruction of hadrons, jets, hadronic decays of tau leptons and missing transverse energy. This sampling calorimeter uses steel plates as absorber and scintillating tiles as active medium. The light produced by the passage of charged particles is transmitted by wavelength shifting fibres to photomultiplier tubes, located in the outer part of the calorimeter. Neutral particles may also produce a signal after interacting with the material and producing charged particles. The readout is segmented into about 5000 cells, each of them being read out by two photomultipliers in parallel. To calibrate and monitor the stability and performance of each part of the readout chain during the data taking, a set of calibration systems is used. This comprises Cesium radioactive sources, Laser, charge injection elements and an integrator based readout system. Information from all systems allows to monitor and equalise the calorimeter response at each stage of the signal production, from scintillation light to digitisation. Calibration runs are monitored from a data quality perspective and used as a cross-check for physics runs. The data quality efficiency achieved during 2016 was 98.9%. These calibration and stability of the calorimeter reported here show that the TileCal performance is within the design requirements and has given essential contribution to reconstructed objects and physics results.

  3. The Production and Qualification of Scintillator Tiles for the ATLAS Hadronic Calorimeter

    CERN Document Server

    Abdallah, J; Alexa, C; Alves, R; Amaral, P; Ananiev, A; Anderson, K; Andresen, X; Antonaki, A; Batusov, V; Bednar, P; Bergeaas, E; Biscarat, C; Blanch, O; Blanchot, G; Bohm, C; Boldea, V; Bosi, F; Bosman, M; Bromberg, C; Budagov, Yu; Calvet, D; Cardeira, C; Carli, T; Carvalho, J; Cascella, M; Castillo, M V; Costello, J; Cavalli-Sforza, M; Cavasinni, V; Cerqueira, A S; Clément, C; Cobal, M; Cogswell, F; Constantinescu, S; Costanzo, D; Da Silva, P; David, M; Davidek, T; Dawson, J; De, K; Del Prete, T; Diakov, E; Di Girolamo, B; Dita, S; Dolejsi, J; Dolezal, Z; Dotti, A; Downing, R; Drake, G; Efthymiopoulos, I; Errede, D; Errede, S; Farbin, A; Fassouliotis, D; Feng, E; Fenyuk, A; Ferdi, C; Ferreira, B C; Ferrer, A; Flaminio, V; Flix, J; Francavilla, P; Fullana, E; Garde, V; Gellerstedt, K; Giakoumopoulou, V; Giangiobbe, V; Gildemeister, O; Gilewsky, V; Giokaris, N; Gollub, N; Gomes, A; González, V; Gouveia, J; Grenier, P; Gris, P; Guarino, V; Guicheney, C; Sen-Gupta, A; Hakobyan, H; Haney, M; Hellman, S; Henriques, A; Higón, E; Hill, N; Holmgren, S; Hruska, I; Hurwitz, M; Huston, J; Jen-La Plante, I; Jon-And, K; Junk, T; Karyukhin, A; Khubua, J; Klereborn, J; Konsnantinov, V; Kopikov, S; Korolkov, I; Krivkova, P; Kulchitskii, Yu A; Kurochkin, Yu; Kuzhir, P; Lapin, V; LeCompte, T; Lefèvre, R; Leitner, R; Li, J; Liablin, M; Lokajícek, M; Lomakin, Y; Lourtie, P; Lovas, L; Lupi, A; Maidantchik, C; Maio, A; Maliukov, S; Manousakis, A; Marques, C; Marroquim, F; Martin, F; Mazzoni, E; Merritt, F S; Myagkov, A; Miller, R; Minashvili, I; Miralles, L; Montarou, G; Némécek, S; Nessi, M; Nikitine, I; Nodulman, L; Norniella, O; Onofre, A; Oreglia, M; Palan, B; Pallin, D; Pantea, D; Pereira, A; Pilcher, J E; Pina, J; Pinhão, J; Pod, E; Podlyski, F; Portell, X; Poveda, J; Pribyl, a L; Price, L E; Proudfoot, J; Ramalho, M; Ramstedt, M; Raposeiro, L; Reis, J; Richards, R; Roda, C; Romanov, V; Rosnet, P; Roy, P; Ruiz, A; Rumiantsau, V; Rusakovich, N; Sada Costa, J; Salto, O; Salvachúa, B; Sanchis, E; Sanders, H; Santoni, C; Santos, J; Saraiva, J G; Sarri, F; Says, L P; Schlager, G; Schlereth, J L; Seixas, J M; Selldén, B; Shalanda, N; Shevtsov, P; Shochet, M; Silva, J; Simaitis, V; Simonyan, M; Sisakian, A; Sjölin, J; Solans, C; Solodkov, A; Solovyanov, O; Sosebee, M; Spanó, F; Speckmeyer, P; Stanek, R; Starchenko, E; Starovoitov, P; Suk, M; Sykora, I; Tang, F; Tas, P; Teuscher, R; Tischenko, M; Tokar, S; Topilin, N; Torres, J; Underwood, D; Usai, G; Valero, A; Valkár, S; Valls, J A; Vartapetian, A; Vazeille, F; Vellidis, C; Ventura, F; Vichou, I; Vivarelli, I; Volpi, M; White, A; Zaitsev, A; Zaytsev, Yu; Zenin, A; Zenis, T; Zenonos, Z; Zenz, S; Zilka, B

    2007-01-01

    The production of the scintillator tiles for the ATLAS Tile Calorimeter is presented. In addition to the manufacture and production, the properties of the tiles will be presented including light yield, uniformity and stability.

  4. The ATLAS Detector Safety System

    CERN Multimedia

    Helfried Burckhart; Kathy Pommes; Heidi Sandaker

    The ATLAS Detector Safety System (DSS) has the mandate to put the detector in a safe state in case an abnormal situation arises which could be potentially dangerous for the detector. It covers the CERN alarm severity levels 1 and 2, which address serious risks for the equipment. The highest level 3, which also includes danger for persons, is the responsibility of the CERN-wide system CSAM, which always triggers an intervention by the CERN fire brigade. DSS works independently from and hence complements the Detector Control System, which is the tool to operate the experiment. The DSS is organized in a Front- End (FE), which fulfills autonomously the safety functions and a Back-End (BE) for interaction and configuration. The overall layout is shown in the picture below. ATLAS DSS configuration The FE implementation is based on a redundant Programmable Logical Crate (PLC) system which is used also in industry for such safety applications. Each of the two PLCs alone, one located underground and one at the s...

  5. Studies of photon production in association with jets at the ATLAS detector

    CERN Document Server

    Stockton, Mark; The ATLAS collaboration

    2018-01-01

    The production of prompt photons in association with jets in proton-proton collisions at the LHC provides a testing ground for perturbative QCD (pQCD) with a hard colourless probe less affected by hadronisation effects than jet production. The measurements of the angular correlations between the photon and the jets can be used to probe the dynamics of the hard-scattering process. We present here the latest ATLAS cross-section measurement of isolated photon plus jet at $\\sqrt{s}=13$ TeV, differential in a wide range of kinematic variables describing the photon plus jet production dynamic. We will also present for the first time ATLAS measurements on the differential cross sections of isolated-photon plus heavy-flavour jet production at $\\sqrt{s}=8$ TeV. The results are compared to recent theoretical predictions.

  6. Recent Improvements in the ATLAS PanDA Pilot

    International Nuclear Information System (INIS)

    Nilsson, P; De, K; Bejar, J Caballero; Maeno, T; Potekhin, M; Wenaus, T; Compostella, G; Contreras, C; Dos Santos, T

    2012-01-01

    The Production and Distributed Analysis system (PanDA) in the ATLAS experiment uses pilots to execute submitted jobs on the worker nodes. The pilots are designed to deal with different runtime conditions and failure scenarios, and support many storage systems. This talk will give a brief overview of the PanDA pilot system and will present major features and recent improvements including CernVM File System integration, the job retry mechanism, advanced job monitoring including JEM technology, and validation of new pilot code using the HammerCloud stress-testing system. PanDA is used for all ATLAS distributed production and is the primary system for distributed analysis. It is currently used at over 130 sites worldwide. We analyze the performance of the pilot system in processing LHC data on the OSG, EGI and Nordugrid infrastructures used by ATLAS, and describe plans for its further evolution.

  7. ATLAS rewards industry

    CERN Multimedia

    2006-01-01

    Showing excellence in mechanics, electronics and cryogenics, three industries are honoured for their contributions to the ATLAS experiment. Representatives of the three award-wining companies after the ceremony. For contributing vital pieces to the ATLAS puzzle, three industries were recognized on Friday 5 May during a supplier awards ceremony. After a welcome and overview of the ATLAS experiment by spokesperson Peter Jenni, CERN Secretary-General Maximilian Metzger stressed the importance of industry to CERN's scientific goals. Close interaction with CERN was a key factor in the selection of each rewarded company, in addition to the high-quality products they delivered to the experiment. Alu Menziken Industrie AG, of Switzerland, was honoured for the production of 380,000 aluminium tubes for the Monitored Drift Tube Chambers (MDT). As Giora Mikenberg, the Muon System Project Leader stressed, the aluminium tubes were delivered on time with an extraordinary quality and precision. Between October 2000 and Jan...

  8. Detector Control System for the AFP detector in ATLAS experiment at CERN

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00211068; The ATLAS collaboration; Caforio, Davide; Czekierda, Sabina; Hajduk, Zbigniew; Olszowska, Jolanta; Oleiro Seabra, Luis Filipe; Sicho, Petr

    2017-01-01

    The ATLAS Forward Proton (AFP) detector consists of two forward detectors located at 205 m and 217 m on either side of the ATLAS experiment. The aim is to measure the momenta and angles of diffractively scattered protons. In 2016, two detector stations on one side of the ATLAS interaction point were installed and commissioned. The detector infrastructure and necessary services were installed and are supervised by the Detector Control System (DCS), which is responsible for the coherent and safe operation of the detector. A large variety of used equipment represents a considerable challenge for the AFP DCS design. Industrial Supervisory Control and Data Acquisition (SCADA) product Siemens WinCCOA, together with the CERN Joint Control Project (JCOP) framework and standard industrial and custom developed server applications and protocols are used for reading, processing, monitoring and archiving of the detector parameters. Graphical user interfaces allow for overall detector operation and visualization of the det...

  9. Multilevel Workflow System in the ATLAS Experiment

    CERN Document Server

    Borodin, M; The ATLAS collaboration; Golubkov, D; Klimentov, A; Maeno, T; Vaniachine, A

    2015-01-01

    The ATLAS experiment is scaling up Big Data processing for the next LHC run using a multilevel workflow system comprised of many layers. In Big Data processing ATLAS deals with datasets, not individual files. Similarly a task (comprised of many jobs) has become a unit of the ATLAS workflow in distributed computing, with about 0.8M tasks processed per year. In order to manage the diversity of LHC physics (exceeding 35K physics samples per year), the individual data processing tasks are organized into workflows. For example, the Monte Carlo workflow is composed of many steps: generate or configure hard-processes, hadronize signal and minimum-bias (pileup) events, simulate energy deposition in the ATLAS detector, digitize electronics response, simulate triggers, reconstruct data, convert the reconstructed data into ROOT ntuples for physics analysis, etc. Outputs are merged and/or filtered as necessary to optimize the chain. The bi-level workflow manager - ProdSys2 - generates actual workflow tasks and their jobs...

  10. The ATLAS Pixel Detector

    CERN Document Server

    Huegging, Fabian

    2006-06-26

    The contruction of the ATLAS Pixel Detector which is the innermost layer of the ATLAS tracking system is prgressing well. Because the pixel detector will contribute significantly to the ATLAS track and vertex reconstruction. The detector consists of identical sensor-chip-hybrid modules, arranged in three barrels in the centre and three disks on either side for the forward region. The position of the detector near the interaction point requires excellent radiation hardness, mechanical and thermal robustness, good long-term stability for all parts, combined with a low material budget. The final detector layout, new results from production modules and the status of assembly are presented.

  11. ATLAS Distributed Computing: Experience and Evolution

    CERN Document Server

    Nairz, A; The ATLAS collaboration

    2013-01-01

    The ATLAS experiment has just concluded its first running period which commenced in 2010. After two years of remarkable performance from the LHC and ATLAS, the experiment has accumulated more than 25 fb-1 of data. The total volume of beam and simulated data products exceeds 100 PB distributed across more than 150 computing centers around the world, managed by the experiment's distributed data management system. These sites have provided up to 150,000 computing cores to ATLAS's global production and analysis processing system, enabling a rich physics program including the discovery of the Higgs-like boson in 2012. The wealth of accumulated experience in global data-intensive computing at this massive scale, and the considerably more challenging requirements of LHC computing from 2014 when the LHC resumes operation, are driving a comprehensive design and development cycle to prepare a revised computing model together with data processing and management systems able to meet the demands of higher trigger rates, e...

  12. ATLAS distributed computing: experience and evolution

    CERN Document Server

    Nairz, A; The ATLAS collaboration

    2014-01-01

    The ATLAS experiment has just concluded its first running period which commenced in 2010. After two years of remarkable performance from the LHC and ATLAS, the experiment has accumulated more than 25/fb of data. The total volume of beam and simulated data products exceeds 100~PB distributed across more than 150 computing centres around the world, managed by the experiment's distributed data management system. These sites have provided up to 150,000 computing cores to ATLAS's global production and analysis processing system, enabling a rich physics programme including the discovery of the Higgs-like boson in 2012. The wealth of accumulated experience in global data-intensive computing at this massive scale, and the considerably more challenging requirements of LHC computing from 2015 when the LHC resumes operation, are driving a comprehensive design and development cycle to prepare a revised computing model together with data processing and management systems able to meet the demands of higher trigger rates, e...

  13. ATLAS TDAQ application gateway upgrade during LS1

    CERN Document Server

    KOROL, A; The ATLAS collaboration; BOGDANCHIKOV, A; BRASOLIN, F; CONTESCU, A C; DUBROV, S; HAFEEZ, M; LEE, C J; SCANNICCHIO, D A; TWOMEY, M; VORONKOV, A; ZAYTSEV, A

    2014-01-01

    The ATLAS Gateway service is implemented with a set of dedicated computer nodes to provide a fine-grained access control between CERN General Public Network (GPN) and ATLAS Technical Control Network (ATCN). ATCN connects the ATLAS online farm used for ATLAS Operations and data taking, including the ATLAS TDAQ (Trigger and Data Aquisition) and DCS (Detector Control System) nodes. In particular, it provides restricted access to the web services (proxy), general login sessions (via SSH and RDP protocols), NAT and mail relay from ATCN. At the Operating System level the implementation is based on virtualization technologies. Here we report on the Gateway upgrade during Long Shutdown 1 (LS1) period: it includes the transition to the last production release of the CERN Linux distribution (SLC6), the migration to the centralized configuration management system (based on Puppet) and the redesign of the internal system architecture.

  14. Calculation Sheet for the Basic Design of the ATLAS Fluid System

    Energy Technology Data Exchange (ETDEWEB)

    Park, Hyun Sik; Moon, S. K.; Yun, B. J.; Kwon, T. S.; Choi, K. Y.; Cho, S.; Park, C. K.; Lee, S. J.; Kim, Y. S.; Song, C. H.; Baek, W. P.; Hong, S. D

    2007-03-15

    The basic design of an integral effect test loop for pressurized water reactors (PWRs), the ATLAS (Advanced Thermal-hydraulic Test Loop for Accident Simulation), has been carried out by Thermal-Hydraulics Safety Research Team in Korea Atomic Energy Research Institute (KAERI). The ATLAS facility has been designed to have the length scale of 1/2 and area scale of 1/144 compared with the reference plant, APR1400, and is scaled for full pressure and temperature conditions. This report includes calculation sheets for the basic design of ATLAS fluid systems, which are consisted of a reactor pressure vessel with core simulator, the primary loop piping, a pressurizer, reactor coolant pumps, steam generators, the secondary system, the safety system, the auxiliary system, and the heat loss compensation system. The present calculation sheets will be used to help understanding the basic design of the ATLAS fluid system and its based scaling methodology.

  15. Calculation Sheet for the Basic Design of the ATLAS Fluid System

    International Nuclear Information System (INIS)

    Park, Hyun Sik; Moon, S. K.; Yun, B. J.; Kwon, T. S.; Choi, K. Y.; Cho, S.; Park, C. K.; Lee, S. J.; Kim, Y. S.; Song, C. H.; Baek, W. P.; Hong, S. D.

    2007-03-01

    The basic design of an integral effect test loop for pressurized water reactors (PWRs), the ATLAS (Advanced Thermal-hydraulic Test Loop for Accident Simulation), has been carried out by Thermal-Hydraulics Safety Research Team in Korea Atomic Energy Research Institute (KAERI). The ATLAS facility has been designed to have the length scale of 1/2 and area scale of 1/144 compared with the reference plant, APR1400, and is scaled for full pressure and temperature conditions. This report includes calculation sheets for the basic design of ATLAS fluid systems, which are consisted of a reactor pressure vessel with core simulator, the primary loop piping, a pressurizer, reactor coolant pumps, steam generators, the secondary system, the safety system, the auxiliary system, and the heat loss compensation system. The present calculation sheets will be used to help understanding the basic design of the ATLAS fluid system and its based scaling methodology

  16. Searches for direct pair production of third generation squarks with the ATLAS detector

    CERN Document Server

    Mitrevski, Jovan; The ATLAS collaboration

    2018-01-01

    Naturalness arguments for weak-scale supersymmetry favour supersymmetric partners of the third generation quarks with masses not too far from those of their Standard Model counterparts. Top or bottom squarks with masses less than or around one TeV can also give rise to direct pair production rates at the LHC that can be observed in the data sample recorded by the ATLAS detector. The talk presents recent ATLAS results from searches for direct stop and sbottom pair production, using the data collected during the LHC Run 2.

  17. Searches for direct pair production of third generation squarks with the ATLAS detector

    CERN Document Server

    Mitrevski, Jovan; The ATLAS collaboration

    2018-01-01

    Naturalness arguments for weak-scale supersymmetry favour supersymmetric partners of the third generation quarks with masses not too far from those of their Standard Model counterparts. Top or bottom squarks with masses less than or around one TeV can also give rise to direct pair production rates at the LHC that can be observed in the data sample recorded by the ATLAS detector. The paper presents recent ATLAS results from searches for direct stop and sbottom pair production, using the data collected during the LHC Run 2.

  18. WW Production and Triple Gauge Boson Couplings at ATLAS

    International Nuclear Information System (INIS)

    Simic, Lj.; Vranjes, N.; Reljic, D.; Vudragovic, D.; Popovic, D.S.

    2007-01-01

    We present a strategy for the direct measurement of the WW production rate at the LHC with ATLAS. Sensitivity limits on anomalous WWZ and WW couplings are assessed with account of the effects of higher order QCD corrections and contributions from other theoretical and detector related systematics. (author)

  19. The Detector Safety System of the ATLAS experiment

    International Nuclear Information System (INIS)

    Beltramello, O; Burckhart, H J; Franz, S; Jaekel, M; Jeckel, M; Lueders, S; Morpurgo, G; Santos Pedrosa, F dos; Pommes, K; Sandaker, H

    2009-01-01

    The ATLAS detector at the Large Hadron Collider at CERN is one of the most advanced detectors for High Energy Physics experiments ever built. It consists of the order of ten functionally independent sub-detectors, which all have dedicated services like power, cooling, gas supply. A Detector Safety System has been built to detect possible operational problems and abnormal and potentially dangerous situations at an early stage and, if needed, to bring the relevant part of ATLAS automatically into a safe state. The procedures and the configuration specific to ATLAS are described in detail and first operational experience is given.

  20. Status and Evolution of ATLAS Workload Management System PanDA

    CERN Document Server

    AUTHOR|(CDS)2067365; The ATLAS collaboration

    2012-01-01

    The ATLAS experiment at the LHC uses a sophisticated workload management system, PanDA, to provide access for thousands of physicists to distributed computing resources of unprecedented scale. This system has proved to be robust and scalable during three years of LHC operations. We describe the design and performance of PanDA in ATLAS. The features which make PanDA successful in ATLAS could be applicable to other exabyte scale scientific projects. We describe plans to evolve PanDA towards a general workload management system for the new BigData initiative announced by the US government. Other planned future improvements to PanDA will also be described

  1. HTR fuel modelling with the ATLAS code. Thermal mechanical behaviour and fission product release assessment

    International Nuclear Information System (INIS)

    Guillermier, Pierre; Daniel, Lucile; Gauthier, Laurent

    2009-01-01

    To support AREVA NP in its design on HTR reactor and its HTR fuel R and D program, the Commissariat a l'Energie Atomique developed the ATLAS code (Advanced Thermal mechanicaL Analysis Software) with the objectives: - to quantify, with a statistical approach, the failed particle fraction and fission product release of a HTR fuel core under normal and accidental conditions (compact or pebble design). - to simulate irradiation tests or benchmark in order to compare measurements or others code results with ATLAS evaluation. These two objectives aim at qualifying the code in order to predict fuel behaviour and to design fuel according to core performance and safety requirements. A statistical calculation uses numerous deterministic calculations. The finite element method is used for these deterministic calculations, in order to be able to choose among three types of meshes, depending on what must be simulated: - One-dimensional calculation of one single particle, for intact particles or particles with fully debonded layers. - Two-dimensional calculations of one single particle, in the case of particles which are cracked, partially debonded or shaped in various ways. - Three-dimensional calculations of a whole compact slice, in order to simulate the interactions between the particles, the thermal gradient and the transport of fission products up to the coolant. - Some calculations of a whole pebble, using homogenization methods are being studied. The temperatures, displacements, stresses, strains and fission product concentrations are calculated on each mesh of the model. Statistical calculations are done using these results, taking into account ceramic failure mode, but also fabrication tolerances and material property uncertainties, variations of the loads (fluence, temperature, burn-up) and core data parameters. The statistical method used in ATLAS is the importance sampling. The model of migration of long-lived fission products in the coated particle and more

  2. Mixing and CP-violation in the Bd and Bs systems at ATLAS

    CERN Document Server

    Wharton, Andrew Mark; The ATLAS collaboration

    2016-01-01

    Search for deviations from the standard model is performed in the systems of the neutral B mesons. The Bs system is studied in the decay into J/psi phi. The mixing phase phi_s and the width difference DeltaGamma_s are determined through the simultaneous study of angular distributions in the final state and of the decay time, performed together with flavour tagging at production. The measurement performed by ATLAS with the full LHC Run-1 sample is discussed and compared to the previous world average. The width difference DeltaGamma_d in the Bd system is obtained from the comparison of the decay time distributions in the flavour specific state J/psi K* and in the CP eigenstate J/psi KS. The result obtained from the full sample of data collected by ATLAS at 7 and 8 TeV is the most accurate single measurement of the width difference currently available.

  3. Electroweak probes of small and large systems with the ATLAS detector

    CERN Document Server

    Citron, Zvi Hirsh; The ATLAS collaboration

    2018-01-01

    Measurements of isolated prompt photon and massive electroweak (W and Z) boson production in different collision systems are of great interest to understanding the partonic structure of heavy nuclei, and serve as a constraint on the initial state in larger collision systems. These channels are sensitive to a variety of effects such as the modification of the parton densities in nuclei in certain kinematic regions, and the energy loss of partons as they undergo multiple interactions in the nucleus before the hard parton-parton scattering. High-statistics samples of lead--lead and proton--lead collision data at sNN‾‾‾√=5.02 TeV and 8.16 TeV, respectively, taken by the ATLAS experiment, as well as proton--proton comparison data at analogous collision energies, allow for a detailed study of these phenomena in data and comprehensive comparisons to the predictions of a variety of theoretical approaches. This talk presents the latest ATLAS results in these topics, including updated results on inclusive promp...

  4. Status of the ATLAS control system upgrade

    International Nuclear Information System (INIS)

    Munson, F.H.; Ferraretto, M.; Rutherford, B.

    1992-01-01

    Certain components of the ATLAS control system are two generations behind today's technology. It has been decided to upgrade the control system. in part, by replacing Digital Equipment Corporation (DEC) PDP-11 computers with present-day VAX technology. Two primary goals have been defined for the upgraded control system. The first of these goals is to keep additional ''in-house'' written software to a minimum, while providing the portability necessary to ensure the continued use of existing software. In an attempt to achieve this goal, commercially-available software has been utilized to provide a foundation for the final control-system configuration. The second goal is to develop the new control system, while not interfering with accelerator operations. This paper describes some of the motivation for upgrading the ATLAS control system, the basic features of the new control system, and the present status of the system's development

  5. Measurement of quarkonia production in heavy-ion collisions with the ATLAS detector

    CERN Document Server

    Tapia Araya, Sebastian; The ATLAS collaboration

    2017-01-01

    The suppression of heavy quarkonia states in heavy-ion collisions is a phenomenon understood as a consequence of QGP formation in the hot, dense system formed in heavy-ion collisions at the LHC. In addition to hot matter effects in heavy-ion collisions , cold nuclear effects may also affect quarkonia production . Therefore, a full assessment requires detailed studies on the effects present in both A-A and p+A collisions. Based on p+Pb data collected in 2013 and pp and Pb+Pb data collected in 2015 at the LHC, the ATLAS experiment has studied prompt and non-prompt J/psi and psi(2S) productions as well as Upsilon production via the di-muon decay final states. The results are of the various measurements are discussed.

  6. Measurement of quarkonia production in heavy-ion collisions with the ATLAS detector

    CERN Document Server

    Kremer, Jakub Andrzej; The ATLAS collaboration

    2017-01-01

    The suppression of heavy quarkonia states in heavy-ion collisions is a phenomenon understood as a consequence of QGP formation in the hot, dense system formed in heavy-ion collisions at the LHC. In addition to hot matter effects in heavy-ion collisions, cold nuclear effects may also affect quarkonia production. Therefore, a full assessment requires detailed studies on the effects present in both A-A and p+A collisions. Based on p+Pb data collected in 2013 and pp and Pb+Pb data collected in 2015 at the LHC, the ATLAS experiment has studied prompt and non-prompt J/psi and psi(2S) productions as well as Upsilon production via the di-muon decay final states. The results of the various measurements are discussed

  7. Performance of the ATLAS trigger system in 2015

    Energy Technology Data Exchange (ETDEWEB)

    Aaboud, M. [Univ. Mohamed Premier et LPTPM, Oujda (Morocco). Faculte des Sciences; Aad, G. [CPPM, Aix-Marseille Univ. et CNRS/IN2P3, Marseille (France); Abbott, B. [Oklahoma Univ., Norman, OK (United States). Homer L. Dodge Dept. of Physics and Astronomy; Collaboration: Atlas Collaboration; and others

    2017-05-15

    During 2015 the ATLAS experiment recorded 3.8 fb{sup -1} of proton-proton collision data at a centre-of-mass energy of 13 TeV. The ATLAS trigger system is a crucial component of the experiment, responsible for selecting events of interest at a recording rate of approximately 1 kHz from up to 40 MHz of collisions. This paper presents a short overview of the changes to the trigger and data acquisition systems during the first long shutdown of the LHC and shows the performance of the trigger system and its components based on the 2015 proton-proton collision data. (orig.)

  8. Performance of the ATLAS Trigger System in 2015

    CERN Document Server

    Aaboud, Morad; Abbott, Brad; Abdallah, Jalal; Abdinov, Ovsat; Abeloos, Baptiste; Aben, Rosemarie; AbouZeid, Ossama; Abraham, Nicola; Abramowicz, Halina; Abreu, Henso; Abreu, Ricardo; Abulaiti, Yiming; Acharya, Bobby Samir; Adachi, Shunsuke; Adamczyk, Leszek; Adams, David; Adelman, Jahred; Adomeit, Stefanie; Adye, Tim; Affolder, Tony; Agatonovic-Jovin, Tatjana; Aguilar-Saavedra, Juan Antonio; Ahlen, Steven; Ahmadov, Faig; Aielli, Giulio; Akerstedt, Henrik; Åkesson, Torsten Paul Ake; Akimov, Andrei; Alberghi, Gian Luigi; Albert, Justin; Albrand, Solveig; Alconada Verzini, Maria Josefina; Aleksa, Martin; Aleksandrov, Igor; Alexa, Calin; Alexander, Gideon; Alexopoulos, Theodoros; Alhroob, Muhammad; Ali, Babar; Aliev, Malik; Alimonti, Gianluca; Alison, John; Alkire, Steven Patrick; Allbrooke, Benedict; Allen, Benjamin William; Allport, Phillip; Aloisio, Alberto; Alonso, Alejandro; Alonso, Francisco; Alpigiani, Cristiano; Alshehri, Azzah Aziz; Alstaty, Mahmoud; Alvarez Gonzalez, Barbara; Άlvarez Piqueras, Damián; Alviggi, Mariagrazia; Amadio, Brian Thomas; Amaral Coutinho, Yara; Amelung, Christoph; Amidei, Dante; Amor Dos Santos, Susana Patricia; Amorim, Antonio; Amoroso, Simone; Amundsen, Glenn; Anastopoulos, Christos; Ancu, Lucian Stefan; Andari, Nansi; Andeen, Timothy; Anders, Christoph Falk; Anders, Gabriel; Anders, John Kenneth; Anderson, Kelby; Andreazza, Attilio; Andrei, George Victor; Angelidakis, Stylianos; Angelozzi, Ivan; Angerami, Aaron; Anghinolfi, Francis; Anisenkov, Alexey; Anjos, Nuno; Annovi, Alberto; Antel, Claire; Antonelli, Mario; Antonov, Alexey; Antrim, Daniel Joseph; Anulli, Fabio; Aoki, Masato; Aperio Bella, Ludovica; Arabidze, Giorgi; Arai, Yasuo; Araque, Juan Pedro; Arce, Ayana; Arduh, Francisco Anuar; Arguin, Jean-Francois; Argyropoulos, Spyridon; Arik, Metin; Armbruster, Aaron James; Armitage, Lewis James; Arnaez, Olivier; Arnold, Hannah; Arratia, Miguel; Arslan, Ozan; Artamonov, Andrei; Artoni, Giacomo; Artz, Sebastian; Asai, Shoji; Asbah, Nedaa; Ashkenazi, Adi; Åsman, Barbro; Asquith, Lily; Assamagan, Ketevi; Astalos, Robert; Atkinson, Markus; Atlay, Naim Bora; Augsten, Kamil; Avolio, Giuseppe; Axen, Bradley; Ayoub, Mohamad Kassem; Azuelos, Georges; Baak, Max; Baas, Alessandra; Baca, Matthew John; Bachacou, Henri; Bachas, Konstantinos; Backes, Moritz; Backhaus, Malte; Bagiacchi, Paolo; Bagnaia, Paolo; Bai, Yu; Baines, John; Bajic, Milena; Baker, Oliver Keith; Baldin, Evgenii; Balek, Petr; Balestri, Thomas; Balli, Fabrice; Balunas, William Keaton; Banas, Elzbieta; Banerjee, Swagato; Bannoura, Arwa A E; Barak, Liron; Barberio, Elisabetta Luigia; Barberis, Dario; Barbero, Marlon; Barillari, Teresa; Barisits, Martin-Stefan; Barklow, Timothy; Barlow, Nick; Barnes, Sarah Louise; Barnett, Bruce; Barnett, Michael; Barnovska-Blenessy, Zuzana; Baroncelli, Antonio; Barone, Gaetano; Barr, Alan; Barranco Navarro, Laura; Barreiro, Fernando; Barreiro Guimarães da Costa, João; Bartoldus, Rainer; Barton, Adam Edward; Bartos, Pavol; Basalaev, Artem; Bassalat, Ahmed; Bates, Richard; Batista, Santiago Juan; Batley, Richard; Battaglia, Marco; Bauce, Matteo; Bauer, Florian; Bawa, Harinder Singh; Beacham, James; Beattie, Michael David; Beau, Tristan; Beauchemin, Pierre-Hugues; Bechtle, Philip; Beck, Hans~Peter; Becker, Kathrin; Becker, Maurice; Beckingham, Matthew; Becot, Cyril; Beddall, Andrew; Beddall, Ayda; Bednyakov, Vadim; Bedognetti, Matteo; Bee, Christopher; Beemster, Lars; Beermann, Thomas; Begel, Michael; Behr, Janna Katharina; Bell, Andrew Stuart; Bella, Gideon; Bellagamba, Lorenzo; Bellerive, Alain; Bellomo, Massimiliano; Belotskiy, Konstantin; Beltramello, Olga; Belyaev, Nikita; Benary, Odette; Benchekroun, Driss; Bender, Michael; Bendtz, Katarina; Benekos, Nektarios; Benhammou, Yan; Benhar Noccioli, Eleonora; Benitez, Jose; Benjamin, Douglas; Bensinger, James; Bentvelsen, Stan; Beresford, Lydia; Beretta, Matteo; Berge, David; Bergeaas Kuutmann, Elin; Berger, Nicolas; Beringer, Jürg; Berlendis, Simon; Bernard, Nathan Rogers; Bernius, Catrin; Bernlochner, Florian Urs; Berry, Tracey; Berta, Peter; Bertella, Claudia; Bertoli, Gabriele; Bertolucci, Federico; Bertram, Iain Alexander; Bertsche, Carolyn; Bertsche, David; Besjes, Geert-Jan; Bessidskaia Bylund, Olga; Bessner, Martin Florian; Besson, Nathalie; Betancourt, Christopher; Bethani, Agni; Bethke, Siegfried; Bevan, Adrian John; Bianchi, Riccardo-Maria; Bianco, Michele; Biebel, Otmar; Biedermann, Dustin; Bielski, Rafal; Biesuz, Nicolo Vladi; Biglietti, Michela; Bilbao De Mendizabal, Javier; Billoud, Thomas Remy Victor; Bilokon, Halina; Bindi, Marcello; Bingul, Ahmet; Bini, Cesare; Biondi, Silvia; Bisanz, Tobias; Bjergaard, David Martin; Black, Curtis; Black, James; Black, Kevin; Blackburn, Daniel; Blair, Robert; Blazek, Tomas; Bloch, Ingo; Blocker, Craig; Blue, Andrew; Blum, Walter; Blumenschein, Ulrike; Blunier, Sylvain; Bobbink, Gerjan; Bobrovnikov, Victor; Bocchetta, Simona Serena; Bocci, Andrea; Bock, Christopher; Boehler, Michael; Boerner, Daniela; Bogaerts, Joannes Andreas; Bogavac, Danijela; Bogdanchikov, Alexander; Bohm, Christian; Boisvert, Veronique; Bokan, Petar; Bold, Tomasz; Boldyrev, Alexey; Bomben, Marco; Bona, Marcella; Boonekamp, Maarten; Borisov, Anatoly; Borissov, Guennadi; Bortfeldt, Jonathan; Bortoletto, Daniela; Bortolotto, Valerio; Bos, Kors; Boscherini, Davide; Bosman, Martine; Bossio Sola, Jonathan David; Boudreau, Joseph; Bouffard, Julian; Bouhova-Thacker, Evelina Vassileva; Boumediene, Djamel Eddine; Bourdarios, Claire; Boutle, Sarah Kate; Boveia, Antonio; Boyd, James; Boyko, Igor; Bracinik, Juraj; Brandt, Andrew; Brandt, Gerhard; Brandt, Oleg; Bratzler, Uwe; Brau, Benjamin; Brau, James; Breaden Madden, William Dmitri; Brendlinger, Kurt; Brennan, Amelia Jean; Brenner, Lydia; Brenner, Richard; Bressler, Shikma; Bristow, Timothy Michael; Britton, Dave; Britzger, Daniel; Brochu, Frederic; Brock, Ian; Brock, Raymond; Brooijmans, Gustaaf; Brooks, Timothy; Brooks, William; Brosamer, Jacquelyn; Brost, Elizabeth; Broughton, James; Bruckman de Renstrom, Pawel; Bruncko, Dusan; Bruneliere, Renaud; Bruni, Alessia; Bruni, Graziano; Bruni, Lucrezia Stella; Brunt, Benjamin; Bruschi, Marco; Bruscino, Nello; Bryant, Patrick; Bryngemark, Lene; Buanes, Trygve; Buat, Quentin; Buchholz, Peter; Buckley, Andrew; Budagov, Ioulian; Buehrer, Felix; Bugge, Magnar Kopangen; Bulekov, Oleg; Bullock, Daniel; Burckhart, Helfried; Burdin, Sergey; Burgard, Carsten Daniel; Burger, Angela Maria; Burghgrave, Blake; Burka, Klaudia; Burke, Stephen; Burmeister, Ingo; Burr, Jonathan Thomas Peter; Busato, Emmanuel; Büscher, Daniel; Büscher, Volker; Bussey, Peter; Butler, John; Buttar, Craig; Butterworth, Jonathan; Butti, Pierfrancesco; Buttinger, William; Buzatu, Adrian; Buzykaev, Aleksey; Cabrera Urbán, Susana; Caforio, Davide; Cairo, Valentina; Cakir, Orhan; Calace, Noemi; Calafiura, Paolo; Calandri, Alessandro; Calderini, Giovanni; Calfayan, Philippe; Callea, Giuseppe; Caloba, Luiz; Calvente Lopez, Sergio; Calvet, David; Calvet, Samuel; Calvet, Thomas Philippe; Camacho Toro, Reina; Camarda, Stefano; Camarri, Paolo; Cameron, David; Caminal Armadans, Roger; Camincher, Clement; Campana, Simone; Campanelli, Mario; Camplani, Alessandra; Campoverde, Angel; Canale, Vincenzo; Canepa, Anadi; Cano Bret, Marc; Cantero, Josu; Cao, Tingting; Capeans Garrido, Maria Del Mar; Caprini, Irinel; Caprini, Mihai; Capua, Marcella; Carbone, Ryne Michael; Cardarelli, Roberto; Cardillo, Fabio; Carli, Ina; Carli, Tancredi; Carlino, Gianpaolo; Carlson, Benjamin Taylor; Carminati, Leonardo; Carney, Rebecca; Caron, Sascha; Carquin, Edson; Carrillo-Montoya, German D; Carter, Janet; Carvalho, João; Casadei, Diego; Casado, Maria Pilar; Casolino, Mirkoantonio; Casper, David William; Castaneda-Miranda, Elizabeth; Castelijn, Remco; Castelli, Angelantonio; Castillo Gimenez, Victoria; Castro, Nuno Filipe; Catinaccio, Andrea; Catmore, James; Cattai, Ariella; Caudron, Julien; Cavaliere, Viviana; Cavallaro, Emanuele; Cavalli, Donatella; Cavalli-Sforza, Matteo; Cavasinni, Vincenzo; Ceradini, Filippo; Cerda Alberich, Leonor; Santiago Cerqueira, Augusto; Cerri, Alessandro; Cerrito, Lucio; Cerutti, Fabio; Cervelli, Alberto; Cetin, Serkant Ali; Chafaq, Aziz; Chakraborty, Dhiman; Chan, Stephen Kam-wah; Chan, Yat Long; Chang, Philip; Chapman, John Derek; Charlton, Dave; Chatterjee, Avishek; Chau, Chav Chhiv; Chavez Barajas, Carlos Alberto; Che, Siinn; Cheatham, Susan; Chegwidden, Andrew; Chekanov, Sergei; Chekulaev, Sergey; Chelkov, Gueorgui; Chelstowska, Magda Anna; Chen, Chunhui; Chen, Hucheng; Chen, Karen; Chen, Shenjian; Chen, Shion; Chen, Xin; Chen, Ye; Cheng, Hok Chuen; Cheng, Huajie; Cheng, Yangyang; Cheplakov, Alexander; Cheremushkina, Evgenia; Cherkaoui El Moursli, Rajaa; Chernyatin, Valeriy; Cheu, Elliott; Chevalier, Laurent; Chiarella, Vitaliano; Chiarelli, Giorgio; Chiodini, Gabriele; Chisholm, Andrew; Chitan, Adrian; Chizhov, Mihail; Choi, Kyungeon; Chomont, Arthur Rene; Chouridou, Sofia; Chow, Bonnie Kar Bo; Christodoulou, Valentinos; Chromek-Burckhart, Doris; Chudoba, Jiri; Chuinard, Annabelle Julia; Chwastowski, Janusz; Chytka, Ladislav; Ciapetti, Guido; Ciftci, Abbas Kenan; Cinca, Diane; Cindro, Vladimir; Cioara, Irina Antonela; Ciocca, Claudia; Ciocio, Alessandra; Cirotto, Francesco; Citron, Zvi Hirsh; Citterio, Mauro; Ciubancan, Mihai; Clark, Allan G; Clark, Brian Lee; Clark, Michael; Clark, Philip James; Clarke, Robert; Clement, Christophe; Coadou, Yann; Cobal, Marina; Coccaro, Andrea; Cochran, James H; Colasurdo, Luca; Cole, Brian; Colijn, Auke-Pieter; Collot, Johann; Colombo, Tommaso; Compostella, Gabriele; Conde Muiño, Patricia; Coniavitis, Elias; Connell, Simon Henry; Connelly, Ian; Consorti, Valerio; Constantinescu, Serban; Conti, Geraldine; Conventi, Francesco; Cooke, Mark; Cooper, Ben; Cooper-Sarkar, Amanda; Cormier, Felix; Cormier, Kyle James Read; Cornelissen, Thijs; Corradi, Massimo; Corriveau, Francois; Cortes-Gonzalez, Arely; Cortiana, Giorgio; Costa, Giuseppe; Costa, María José; Costanzo, Davide; Cottin, Giovanna; Cowan, Glen; Cox, Brian; Cranmer, Kyle; Crawley, Samuel Joseph; Cree, Graham; Crépé-Renaudin, Sabine; Crescioli, Francesco; Cribbs, Wayne Allen; Crispin Ortuzar, Mireia; Cristinziani, Markus; Croft, Vince; Crosetti, Giovanni; Cueto, Ana; Cuhadar Donszelmann, Tulay; Cummings, Jane; Curatolo, Maria; Cúth, Jakub; Czirr, Hendrik; Czodrowski, Patrick; D'amen, Gabriele; D'Auria, Saverio; D'Onofrio, Monica; Da Cunha Sargedas De Sousa, Mario Jose; Da Via, Cinzia; Dabrowski, Wladyslaw; Dado, Tomas; Dai, Tiesheng; Dale, Orjan; Dallaire, Frederick; Dallapiccola, Carlo; Dam, Mogens; Dandoy, Jeffrey; Dang, Nguyen Phuong; Daniells, Andrew Christopher; Dann, Nicholas Stuart; Danninger, Matthias; Dano Hoffmann, Maria; Dao, Valerio; Darbo, Giovanni; Darmora, Smita; Dassoulas, James; Dattagupta, Aparajita; Davey, Will; David, Claire; Davidek, Tomas; Davies, Merlin; Davison, Peter; Dawe, Edmund; Dawson, Ian; De, Kaushik; de Asmundis, Riccardo; De Benedetti, Abraham; De Castro, Stefano; De Cecco, Sandro; De Groot, Nicolo; de Jong, Paul; De la Torre, Hector; De Lorenzi, Francesco; De Maria, Antonio; De Pedis, Daniele; De Salvo, Alessandro; De Sanctis, Umberto; De Santo, Antonella; De Vivie De Regie, Jean-Baptiste; Dearnaley, William James; Debbe, Ramiro; Debenedetti, Chiara; Dedovich, Dmitri; Dehghanian, Nooshin; Deigaard, Ingrid; Del Gaudio, Michela; Del Peso, Jose; Del Prete, Tarcisio; Delgove, David; Deliot, Frederic; Delitzsch, Chris Malena; Dell'Acqua, Andrea; Dell'Asta, Lidia; Dell'Orso, Mauro; Della Pietra, Massimo; della Volpe, Domenico; Delmastro, Marco; Delsart, Pierre-Antoine; DeMarco, David; Demers, Sarah; Demichev, Mikhail; Demilly, Aurelien; Denisov, Sergey; Denysiuk, Denys; Derendarz, Dominik; Derkaoui, Jamal Eddine; Derue, Frederic; Dervan, Paul; Desch, Klaus Kurt; Deterre, Cecile; Dette, Karola; Deviveiros, Pier-Olivier; Dewhurst, Alastair; Dhaliwal, Saminder; Di Ciaccio, Anna; Di Ciaccio, Lucia; Di Clemente, William Kennedy; Di Donato, Camilla; Di Girolamo, Alessandro; Di Girolamo, Beniamino; Di Micco, Biagio; Di Nardo, Roberto; Di Simone, Andrea; Di Sipio, Riccardo; Di Valentino, David; Diaconu, Cristinel; Diamond, Miriam; Dias, Flavia; Diaz, Marco Aurelio; Diehl, Edward; Dietrich, Janet; Díez Cornell, Sergio; Dimitrievska, Aleksandra; Dingfelder, Jochen; Dita, Petre; Dita, Sanda; Dittus, Fridolin; Djama, Fares; Djobava, Tamar; Djuvsland, Julia Isabell; Barros do Vale, Maria Aline; Dobos, Daniel; Dobre, Monica; Doglioni, Caterina; Dolejsi, Jiri; Dolezal, Zdenek; Donadelli, Marisilvia; Donati, Simone; Dondero, Paolo; Donini, Julien; Dopke, Jens; Doria, Alessandra; Dova, Maria-Teresa; Doyle, Tony; Drechsler, Eric; Dris, Manolis; Du, Yanyan; Duarte-Campderros, Jorge; Duchovni, Ehud; Duckeck, Guenter; Ducu, Otilia Anamaria; Duda, Dominik; Dudarev, Alexey; Dudder, Andreas Christian; Duffield, Emily Marie; Duflot, Laurent; Dührssen, Michael; Dumancic, Mirta; Duncan, Anna Kathryn; Dunford, Monica; Duran Yildiz, Hatice; Düren, Michael; Durglishvili, Archil; Duschinger, Dirk; Dutta, Baishali; Dyndal, Mateusz; Eckardt, Christoph; Ecker, Katharina Maria; Edgar, Ryan Christopher; Edwards, Nicholas Charles; Eifert, Till; Eigen, Gerald; Einsweiler, Kevin; Ekelof, Tord; El Kacimi, Mohamed; Ellajosyula, Venugopal; Ellert, Mattias; Elles, Sabine; Ellinghaus, Frank; Elliot, Alison; Ellis, Nicolas; Elmsheuser, Johannes; Elsing, Markus; Emeliyanov, Dmitry; Enari, Yuji; Endner, Oliver Chris; Ennis, Joseph Stanford; Erdmann, Johannes; Ereditato, Antonio; Ernis, Gunar; Ernst, Jesse; Ernst, Michael; Errede, Steven; Ertel, Eugen; Escalier, Marc; Esch, Hendrik; Escobar, Carlos; Esposito, Bellisario; Etienvre, Anne-Isabelle; Etzion, Erez; Evans, Hal; Ezhilov, Alexey; Ezzi, Mohammed; Fabbri, Federica; Fabbri, Laura; Facini, Gabriel; Fakhrutdinov, Rinat; Falciano, Speranza; Falla, Rebecca Jane; Faltova, Jana; Fang, Yaquan; Fanti, Marcello; Farbin, Amir; Farilla, Addolorata; Farina, Christian; Farina, Edoardo Maria; Farooque, Trisha; Farrell, Steven; Farrington, Sinead; Farthouat, Philippe; Fassi, Farida; Fassnacht, Patrick; Fassouliotis, Dimitrios; Faucci Giannelli, Michele; Favareto, Andrea; Fawcett, William James; Fayard, Louis; Fedin, Oleg; Fedorko, Wojciech; Feigl, Simon; Feligioni, Lorenzo; Feng, Cunfeng; Feng, Eric; Feng, Haolu; Fenyuk, Alexander; Feremenga, Last; Fernandez Martinez, Patricia; Fernandez Perez, Sonia; Ferrando, James; Ferrari, Arnaud; Ferrari, Pamela; Ferrari, Roberto; Ferreira de Lima, Danilo Enoque; Ferrer, Antonio; Ferrere, Didier; Ferretti, Claudio; Fiedler, Frank; Filipčič, Andrej; Filipuzzi, Marco; Filthaut, Frank; Fincke-Keeler, Margret; Finelli, Kevin Daniel; Fiolhais, Miguel; Fiorini, Luca; Fischer, Adam; Fischer, Cora; Fischer, Julia; Fisher, Wade Cameron; Flaschel, Nils; Fleck, Ivor; Fleischmann, Philipp; Fletcher, Gareth Thomas; Fletcher, Rob Roy MacGregor; Flick, Tobias; Flierl, Bernhard Matthias; Flores Castillo, Luis; Flowerdew, Michael; Forcolin, Giulio Tiziano; Formica, Andrea; Forti, Alessandra; Foster, Andrew Geoffrey; Fournier, Daniel; Fox, Harald; Fracchia, Silvia; Francavilla, Paolo; Franchini, Matteo; Francis, David; Franconi, Laura; Franklin, Melissa; Frate, Meghan; Fraternali, Marco; Freeborn, David; Fressard-Batraneanu, Silvia; Friedrich, Felix; Froidevaux, Daniel; Frost, James; Fukunaga, Chikara; Fullana Torregrosa, Esteban; Fusayasu, Takahiro; Fuster, Juan; Gabaldon, Carolina; Gabizon, Ofir; Gabrielli, Alessandro; Gabrielli, Andrea; Gach, Grzegorz; Gadatsch, Stefan; Gagliardi, Guido; Gagnon, Louis Guillaume; Gagnon, Pauline; Galea, Cristina; Galhardo, Bruno; Gallas, Elizabeth; Gallop, Bruce; Gallus, Petr; Galster, Gorm Aske Gram Krohn; Gan, KK; Ganguly, Sanmay; Gao, Jun; Gao, Yanyan; Gao, Yongsheng; Garay Walls, Francisca; García, Carmen; García Navarro, José Enrique; Garcia-Sciveres, Maurice; Gardner, Robert; Garelli, Nicoletta; Garonne, Vincent; Gascon Bravo, Alberto; Gasnikova, Ksenia; Gatti, Claudio; Gaudiello, Andrea; Gaudio, Gabriella; Gauthier, Lea; Gavrilenko, Igor; Gay, Colin; Gaycken, Goetz; Gazis, Evangelos; Gecse, Zoltan; Gee, Norman; Geich-Gimbel, Christoph; Geisen, Marc; Geisler, Manuel Patrice; Gellerstedt, Karl; Gemme, Claudia; Genest, Marie-Hélène; Geng, Cong; Gentile, Simonetta; Gentsos, Christos; George, Simon; Gerbaudo, Davide; Gershon, Avi; Ghasemi, Sara; Ghneimat, Mazuza; Giacobbe, Benedetto; Giagu, Stefano; Giannetti, Paola; Gibson, Stephen; Gignac, Matthew; Gilchriese, Murdock; Gillam, Thomas; Gillberg, Dag; Gilles, Geoffrey; Gingrich, Douglas; Giokaris, Nikos; Giordani, MarioPaolo; Giorgi, Filippo Maria; Giraud, Pierre-Francois; Giromini, Paolo; Giugni, Danilo; Giuli, Francesco; Giuliani, Claudia; Giulini, Maddalena; Gjelsten, Børge Kile; Gkaitatzis, Stamatios; Gkialas, Ioannis; Gkougkousis, Evangelos Leonidas; Gladilin, Leonid; Glasman, Claudia; Glatzer, Julian; Glaysher, Paul; Glazov, Alexandre; Goblirsch-Kolb, Maximilian; Godlewski, Jan; Goldfarb, Steven; Golling, Tobias; Golubkov, Dmitry; Gomes, Agostinho; Gonçalo, Ricardo; Goncalves Pinto Firmino Da Costa, Joao; Gonella, Giulia; Gonella, Laura; Gongadze, Alexi; González de la Hoz, Santiago; Gonzalez-Sevilla, Sergio; Goossens, Luc; Gorbounov, Petr Andreevich; Gordon, Howard; Gorelov, Igor; Gorini, Benedetto; Gorini, Edoardo; Gorišek, Andrej; Gornicki, Edward; Goshaw, Alfred; Gössling, Claus; Gostkin, Mikhail Ivanovitch; Goudet, Christophe Raymond; Goujdami, Driss; Goussiou, Anna; Govender, Nicolin; Gozani, Eitan; Graber, Lars; Grabowska-Bold, Iwona; Gradin, Per Olov Joakim; Grafström, Per; Gramling, Johanna; Gramstad, Eirik; Grancagnolo, Sergio; Gratchev, Vadim; Gravila, Paul Mircea; Gray, Heather; Graziani, Enrico; Greenwood, Zeno Dixon; Grefe, Christian; Gregersen, Kristian; Gregor, Ingrid-Maria; Grenier, Philippe; Grevtsov, Kirill; Griffiths, Justin; Grillo, Alexander; Grimm, Kathryn; Grinstein, Sebastian; Gris, Philippe Luc Yves; Grivaz, Jean-Francois; Groh, Sabrina; Gross, Eilam; Grosse-Knetter, Joern; Grossi, Giulio Cornelio; Grout, Zara Jane; Guan, Liang; Guan, Wen; Guenther, Jaroslav; Guescini, Francesco; Guest, Daniel; Gueta, Orel; Gui, Bin; Guido, Elisa; Guillemin, Thibault; Guindon, Stefan; Gul, Umar; Gumpert, Christian; Guo, Jun; Guo, Yicheng; Gupta, Ruchi; Gupta, Shaun; Gustavino, Giuliano; Gutierrez, Phillip; Gutierrez Ortiz, Nicolas Gilberto; Gutschow, Christian; Guyot, Claude; Gwenlan, Claire; Gwilliam, Carl; Haas, Andy; Haber, Carl; Hadavand, Haleh Khani; Haddad, Nacim; Hadef, Asma; Hageböck, Stephan; Hagihara, Mutsuto; Hajduk, Zbigniew; Hakobyan, Hrachya; Haleem, Mahsana; Haley, Joseph; Halladjian, Garabed; Hallewell, Gregory David; Hamacher, Klaus; Hamal, Petr; Hamano, Kenji; Hamilton, Andrew; Hamity, Guillermo Nicolas; Hamnett, Phillip George; Han, Liang; Hanagaki, Kazunori; Hanawa, Keita; Hance, Michael; Haney, Bijan; Hanke, Paul; Hanna, Remie; Hansen, Jørgen Beck; Hansen, Jorn Dines; Hansen, Maike Christina; Hansen, Peter Henrik; Hara, Kazuhiko; Hard, Andrew; Harenberg, Torsten; Hariri, Faten; Harkusha, Siarhei; Harrington, Robert; Harrison, Paul Fraser; Hartjes, Fred; Hartmann, Nikolai Marcel; Hasegawa, Makoto; Hasegawa, Yoji; Hasib, Ahmed; Hassani, Samira; Haug, Sigve; Hauser, Reiner; Hauswald, Lorenz; Havranek, Miroslav; Hawkes, Christopher; Hawkings, Richard John; Hayakawa, Daiki; Hayden, Daniel; Hays, Chris; Hays, Jonathan Michael; Hayward, Helen; Haywood, Stephen; Head, Simon; Heck, Tobias; Hedberg, Vincent; Heelan, Louise; Heim, Sarah; Heim, Timon; Heinemann, Beate; Heinrich, Jochen Jens; Heinrich, Lukas; Heinz, Christian; Hejbal, Jiri; Helary, Louis; Hellman, Sten; Helsens, Clement; Henderson, James; Henderson, Robert; Heng, Yang; Henkelmann, Steffen; Henriques Correia, Ana Maria; Henrot-Versille, Sophie; Herbert, Geoffrey Henry; Herde, Hannah; Herget, Verena; Hernández Jiménez, Yesenia; Herten, Gregor; Hertenberger, Ralf; Hervas, Luis; Hesketh, Gavin Grant; Hessey, Nigel; Hetherly, Jeffrey Wayne; Higón-Rodriguez, Emilio; Hill, Ewan; Hill, John; Hiller, Karl Heinz; Hillier, Stephen; Hinchliffe, Ian; Hines, Elizabeth; Hirose, Minoru; Hirschbuehl, Dominic; Hoad, Xanthe; Hobbs, John; Hod, Noam; Hodgkinson, Mark; Hodgson, Paul; Hoecker, Andreas; Hoeferkamp, Martin; Hoenig, Friedrich; Hohn, David; Holmes, Tova Ray; Homann, Michael; Honda, Takuya; Hong, Tae Min; Hooberman, Benjamin Henry; Hopkins, Walter; Horii, Yasuyuki; Horton, Arthur James; Hostachy, Jean-Yves; Hou, Suen; Hoummada, Abdeslam; Howarth, James; Hoya, Joaquin; Hrabovsky, Miroslav; Hristova, Ivana; Hrivnac, Julius; Hryn'ova, Tetiana; Hrynevich, Aliaksei; Hsu, Pai-hsien Jennifer; Hsu, Shih-Chieh; Hu, Qipeng; Hu, Shuyang; Huang, Yanping; Hubacek, Zdenek; Hubaut, Fabrice; Huegging, Fabian; Huffman, Todd Brian; Hughes, Emlyn; Hughes, Gareth; Huhtinen, Mika; Huo, Peng; Huseynov, Nazim; Huston, Joey; Huth, John; Iacobucci, Giuseppe; Iakovidis, Georgios; Ibragimov, Iskander; Iconomidou-Fayard, Lydia; Ideal, Emma; Idrissi, Zineb; Iengo, Paolo; Igonkina, Olga; Iizawa, Tomoya; Ikai, Takashi; Ikegami, Yoichi; Ikeno, Masahiro; Ilchenko, Yuriy; Iliadis, Dimitrios; Ilic, Nikolina; Introzzi, Gianluca; Ioannou, Pavlos; Iodice, Mauro; Iordanidou, Kalliopi; Ippolito, Valerio; Ishijima, Naoki; Ishino, Masaya; Ishitsuka, Masaki; Ishmukhametov, Renat; Issever, Cigdem; Istin, Serhat; Ito, Fumiaki; Iturbe Ponce, Julia Mariana; Iuppa, Roberto; Iwanski, Wieslaw; Iwasaki, Hiroyuki; Izen, Joseph; Izzo, Vincenzo; Jabbar, Samina; Jackson, Brett; Jackson, Paul; Jain, Vivek; Jakobi, Katharina Bianca; Jakobs, Karl; Jakobsen, Sune; Jakoubek, Tomas; Jamin, David Olivier; Jana, Dilip; Jansky, Roland; Janssen, Jens; Janus, Michel; Janus, Piotr Andrzej; Jarlskog, Göran; Javadov, Namig; Javůrek, Tomáš; Jeanneau, Fabien; Jeanty, Laura; Jejelava, Juansher; Jeng, Geng-yuan; Jennens, David; Jenni, Peter; Jeske, Carl; Jézéquel, Stéphane; Ji, Haoshuang; Jia, Jiangyong; Jiang, Hai; Jiang, Yi; Jiang, Zihao; Jiggins, Stephen; Jimenez Pena, Javier; Jin, Shan; Jinaru, Adam; Jinnouchi, Osamu; Jivan, Harshna; Johansson, Per; Johns, Kenneth; Johnson, William Joseph; Jon-And, Kerstin; Jones, Graham; Jones, Roger; Jones, Sarah; Jones, Tim; Jongmanns, Jan; Jorge, Pedro; Jovicevic, Jelena; Ju, Xiangyang; Juste Rozas, Aurelio; Köhler, Markus Konrad; Kaczmarska, Anna; Kado, Marumi; Kagan, Harris; Kagan, Michael; Kahn, Sebastien Jonathan; Kaji, Toshiaki; Kajomovitz, Enrique; Kalderon, Charles William; Kaluza, Adam; Kama, Sami; Kamenshchikov, Andrey; Kanaya, Naoko; Kaneti, Steven; Kanjir, Luka; Kantserov, Vadim; Kanzaki, Junichi; Kaplan, Benjamin; Kaplan, Laser Seymour; Kapliy, Anton; Kar, Deepak; Karakostas, Konstantinos; Karamaoun, Andrew; Karastathis, Nikolaos; Kareem, Mohammad Jawad; Karentzos, Efstathios; Karnevskiy, Mikhail; Karpov, Sergey; Karpova, Zoya; Karthik, Krishnaiyengar; Kartvelishvili, Vakhtang; Karyukhin, Andrey; Kasahara, Kota; Kashif, Lashkar; Kass, Richard; Kastanas, Alex; Kataoka, Yousuke; Kato, Chikuma; Katre, Akshay; Katzy, Judith; Kawade, Kentaro; Kawagoe, Kiyotomo; Kawamoto, Tatsuo; Kawamura, Gen; Kazanin, Vassili; Keeler, Richard; Kehoe, Robert; Keller, John; Kempster, Jacob Julian; Keoshkerian, Houry; Kepka, Oldrich; Kerševan, Borut Paul; Kersten, Susanne; Keyes, Robert; Khader, Mazin; Khalil-zada, Farkhad; Khanov, Alexander; Kharlamov, Alexey; Kharlamova, Tatyana; Khoo, Teng Jian; Khovanskiy, Valery; Khramov, Evgeniy; Khubua, Jemal; Kido, Shogo; Kilby, Callum; Kim, Hee Yeun; Kim, Shinhong; Kim, Young-Kee; Kimura, Naoki; Kind, Oliver Maria; King, Barry; King, Matthew; Kirk, Julie; Kiryunin, Andrey; Kishimoto, Tomoe; Kisielewska, Danuta; Kiss, Florian; Kiuchi, Kenji; Kivernyk, Oleh; Kladiva, Eduard; Klein, Matthew Henry; Klein, Max; Klein, Uta; Kleinknecht, Konrad; Klimek, Pawel; Klimentov, Alexei; Klingenberg, Reiner; Klioutchnikova, Tatiana; Kluge, Eike-Erik; Kluit, Peter; Kluth, Stefan; Knapik, Joanna; Kneringer, Emmerich; Knoops, Edith; Knue, Andrea; Kobayashi, Aine; Kobayashi, Dai; Kobayashi, Tomio; Kobel, Michael; Kocian, Martin; Kodys, Peter; Koffas, Thomas; Koffeman, Els; Köhler, Nicolas Maximilian; Koi, Tatsumi; Kolanoski, Hermann; Kolb, Mathis; Koletsou, Iro; Komar, Aston; Komori, Yuto; Kondo, Takahiko; Kondrashova, Nataliia; Köneke, Karsten; König, Adriaan; Kono, Takanori; Konoplich, Rostislav; Konstantinidis, Nikolaos; Kopeliansky, Revital; Koperny, Stefan; Köpke, Lutz; Kopp, Anna Katharina; Korcyl, Krzysztof; Kordas, Kostantinos; Korn, Andreas; Korol, Aleksandr; Korolkov, Ilya; Korolkova, Elena; Kortner, Oliver; Kortner, Sandra; Kosek, Tomas; Kostyukhin, Vadim; Kotwal, Ashutosh; Koulouris, Aimilianos; Kourkoumeli-Charalampidi, Athina; Kourkoumelis, Christine; Kouskoura, Vasiliki; Kowalewska, Anna Bozena; Kowalewski, Robert Victor; Kowalski, Tadeusz; Kozakai, Chihiro; Kozanecki, Witold; Kozhin, Anatoly; Kramarenko, Viktor; Kramberger, Gregor; Krasnopevtsev, Dimitriy; Krasny, Mieczyslaw Witold; Krasznahorkay, Attila; Kravchenko, Anton; Kretz, Moritz; Kretzschmar, Jan; Kreutzfeldt, Kristof; Krieger, Peter; Krizka, Karol; Kroeninger, Kevin; Kroha, Hubert; Kroll, Joe; Kroseberg, Juergen; Krstic, Jelena; Kruchonak, Uladzimir; Krüger, Hans; Krumnack, Nils; Kruse, Mark; Kruskal, Michael; Kubota, Takashi; Kucuk, Hilal; Kuday, Sinan; Kuechler, Jan Thomas; Kuehn, Susanne; Kugel, Andreas; Kuger, Fabian; Kuhl, Thorsten; Kukhtin, Victor; Kukla, Romain; Kulchitsky, Yuri; Kuleshov, Sergey; Kuna, Marine; Kunigo, Takuto; Kupco, Alexander; Kurashige, Hisaya; Kurchaninov, Leonid; Kurochkin, Yurii; Kurth, Matthew Glenn; Kus, Vlastimil; Kuwertz, Emma Sian; Kuze, Masahiro; Kvita, Jiri; Kwan, Tony; Kyriazopoulos, Dimitrios; La Rosa, Alessandro; La Rosa Navarro, Jose Luis; La Rotonda, Laura; Lacasta, Carlos; Lacava, Francesco; Lacey, James; Lacker, Heiko; Lacour, Didier; Lacuesta, Vicente Ramón; Ladygin, Evgueni; Lafaye, Remi; Laforge, Bertrand; Lagouri, Theodota; Lai, Stanley; Lammers, Sabine; Lampl, Walter; Lançon, Eric; Landgraf, Ulrich; Landon, Murrough; Lanfermann, Marie Christine; Lang, Valerie Susanne; Lange, J örn Christian; Lankford, Andrew; Lanni, Francesco; Lantzsch, Kerstin; Lanza, Agostino; Laplace, Sandrine; Lapoire, Cecile; Laporte, Jean-Francois; Lari, Tommaso; Lasagni Manghi, Federico; Lassnig, Mario; Laurelli, Paolo; Lavrijsen, Wim; Law, Alexander; Laycock, Paul; Lazovich, Tomo; Lazzaroni, Massimo; Le, Brian; Le Dortz, Olivier; Le Guirriec, Emmanuel; Le Quilleuc, Eloi; LeBlanc, Matthew Edgar; LeCompte, Thomas; Ledroit-Guillon, Fabienne Agnes Marie; Lee, Claire Alexandra; Lee, Shih-Chang; Lee, Lawrence; Lefebvre, Benoit; Lefebvre, Guillaume; Lefebvre, Michel; Legger, Federica; Leggett, Charles; Lehan, Allan; Lehmann Miotto, Giovanna; Lei, Xiaowen; Leight, William Axel; Leister, Andrew Gerard; Leite, Marco Aurelio Lisboa; Leitner, Rupert; Lellouch, Daniel; Lemmer, Boris; Leney, Katharine; Lenz, Tatjana; Lenzi, Bruno; Leone, Robert; Leone, Sandra; Leonidopoulos, Christos; Leontsinis, Stefanos; Lerner, Giuseppe; Leroy, Claude; Lesage, Arthur; Lester, Christopher; Levchenko, Mikhail; Levêque, Jessica; Levin, Daniel; Levinson, Lorne; Levy, Mark; Lewis, Dave; Leyton, Michael; Li, Bing; Li, Changqiao; Li, Haifeng; Li, Lei; Li, Liang; Li, Qi; Li, Shu; Li, Xingguo; Li, Yichen; Liang, Zhijun; Liberti, Barbara; Liblong, Aaron; Lichard, Peter; Lie, Ki; Liebal, Jessica; Liebig, Wolfgang; Limosani, Antonio; Lin, Simon; Lin, Tai-Hua; Lindquist, Brian Edward; Lionti, Anthony Eric; Lipeles, Elliot; Lipniacka, Anna; Lisovyi, Mykhailo; Liss, Tony; Lister, Alison; Litke, Alan; Liu, Bo; Liu, Dong; Liu, Hao; Liu, Hongbin; Liu, Jian; Liu, Jianbei; Liu, Kun; Liu, Lulu; Liu, Minghui; Liu, Yanlin; Liu, Yanwen; Livan, Michele; Lleres, Annick; Llorente Merino, Javier; Lloyd, Stephen; Lo Sterzo, Francesco; Lobodzinska, Ewelina Maria; Loch, Peter; Loebinger, Fred; Loew, Kevin Michael; Loginov, Andrey; Lohse, Thomas; Lohwasser, Kristin; Lokajicek, Milos; Long, Brian Alexander; Long, Jonathan David; Long, Robin Eamonn; Longo, Luigi; Looper, Kristina Anne; Lopez Lopez, Jorge Andres; Lopez Mateos, David; Lopez Paredes, Brais; Lopez Paz, Ivan; Lopez Solis, Alvaro; Lorenz, Jeanette; Lorenzo Martinez, Narei; Losada, Marta; Lösel, Philipp Jonathan; Lou, XinChou; Lounis, Abdenour; Love, Jeremy; Love, Peter; Lu, Haonan; Lu, Nan; Lubatti, Henry; Luci, Claudio; Lucotte, Arnaud; Luedtke, Christian; Luehring, Frederick; Lukas, Wolfgang; Luminari, Lamberto; Lundberg, Olof; Lund-Jensen, Bengt; Luzi, Pierre Marc; Lynn, David; Lysak, Roman; Lytken, Else; Lyubushkin, Vladimir; Ma, Hong; Ma, Lian Liang; Ma, Yanhui; Maccarrone, Giovanni; Macchiolo, Anna; Macdonald, Calum Michael; Maček, Boštjan; Machado Miguens, Joana; Madaffari, Daniele; Madar, Romain; Maddocks, Harvey Jonathan; Mader, Wolfgang; Madsen, Alexander; Maeda, Junpei; Maeland, Steffen; Maeno, Tadashi; Maevskiy, Artem; Magradze, Erekle; Mahlstedt, Joern; Maiani, Camilla; Maidantchik, Carmen; Maier, Andreas Alexander; Maier, Thomas; Maio, Amélia; Majewski, Stephanie; Makida, Yasuhiro; Makovec, Nikola; Malaescu, Bogdan; Malecki, Pawel; Maleev, Victor; Malek, Fairouz; Mallik, Usha; Malon, David; Malone, Caitlin; Malone, Claire; Maltezos, Stavros; Malyukov, Sergei; Mamuzic, Judita; Mancini, Giada; Mandelli, Luciano; Mandić, Igor; Maneira, José; Manhaes de Andrade Filho, Luciano; Manjarres Ramos, Joany; Mann, Alexander; Manousos, Athanasios; Mansoulie, Bruno; Mansour, Jason Dhia; Mantifel, Rodger; Mantoani, Matteo; Manzoni, Stefano; Mapelli, Livio; Marceca, Gino; March, Luis; Marchiori, Giovanni; Marcisovsky, Michal; Marjanovic, Marija; Marley, Daniel; Marroquim, Fernando; Marsden, Stephen Philip; Marshall, Zach; Marti-Garcia, Salvador; Martin, Brian Thomas; Martin, Tim; Martin, Victoria Jane; Martin dit Latour, Bertrand; Martinez, Mario; Martinez Outschoorn, Verena; Martin-Haugh, Stewart; Martoiu, Victor Sorin; Martyniuk, Alex; Marzin, Antoine; Masetti, Lucia; Mashimo, Tetsuro; Mashinistov, Ruslan; Masik, Jiri; Maslennikov, Alexey; Massa, Ignazio; Massa, Lorenzo; Mastrandrea, Paolo; Mastroberardino, Anna; Masubuchi, Tatsuya; Mättig, Peter; Mattmann, Johannes; Maurer, Julien; Maxfield, Stephen; Maximov, Dmitriy; Mazini, Rachid; Maznas, Ioannis; Mazza, Simone Michele; Mc Fadden, Neil Christopher; Mc Goldrick, Garrin; Mc Kee, Shawn Patrick; McCarn, Allison; McCarthy, Robert; McCarthy, Tom; McClymont, Laurie; McDonald, Emily; Mcfayden, Josh; Mchedlidze, Gvantsa; McMahon, Steve; McNamara, Peter Charles; McPherson, Robert; Medinnis, Michael; Meehan, Samuel; Mehlhase, Sascha; Mehta, Andrew; Meier, Karlheinz; Meineck, Christian; Meirose, Bernhard; Melini, Davide; Mellado Garcia, Bruce Rafael; Melo, Matej; Meloni, Federico; Menary, Stephen Burns; Meng, Lingxin; Meng, Xiangting; Mengarelli, Alberto; Menke, Sven; Meoni, Evelin; Mergelmeyer, Sebastian; Mermod, Philippe; Merola, Leonardo; Meroni, Chiara; Merritt, Frank; Messina, Andrea; Metcalfe, Jessica; Mete, Alaettin Serhan; Meyer, Carsten; Meyer, Christopher; Meyer, Jean-Pierre; Meyer, Jochen; Meyer Zu Theenhausen, Hanno; Miano, Fabrizio; Middleton, Robin; Miglioranzi, Silvia; Mijović, Liza; Mikenberg, Giora; Mikestikova, Marcela; Mikuž, Marko; Milesi, Marco; Milic, Adriana; Miller, David; Mills, Corrinne; Milov, Alexander; Milstead, David; Minaenko, Andrey; Minami, Yuto; Minashvili, Irakli; Mincer, Allen; Mindur, Bartosz; Mineev, Mikhail; Minegishi, Yuji; Ming, Yao; Mir, Lluisa-Maria; Mistry, Khilesh; Mitani, Takashi; Mitrevski, Jovan; Mitsou, Vasiliki A; Miucci, Antonio; Miyagawa, Paul; Mizukami, Atsushi; Mjörnmark, Jan-Ulf; Mlynarikova, Michaela; Moa, Torbjoern; Mochizuki, Kazuya; Mogg, Philipp; Mohapatra, Soumya; Molander, Simon; Moles-Valls, Regina; Monden, Ryutaro; Mondragon, Matthew Craig; Mönig, Klaus; Monk, James; Monnier, Emmanuel; Montalbano, Alyssa; Montejo Berlingen, Javier; Monticelli, Fernando; Monzani, Simone; Moore, Roger; Morange, Nicolas; Moreno, Deywis; Moreno Llácer, María; Morettini, Paolo; Morgenstern, Stefanie; Mori, Daniel; Mori, Tatsuya; Morii, Masahiro; Morinaga, Masahiro; Morisbak, Vanja; Moritz, Sebastian; Morley, Anthony Keith; Mornacchi, Giuseppe; Morris, John; Mortensen, Simon Stark; Morvaj, Ljiljana; Moschovakos, Paris; Mosidze, Maia; Moss, Harry James; Moss, Josh; Motohashi, Kazuki; Mount, Richard; Mountricha, Eleni; Moyse, Edward; Muanza, Steve; Mudd, Richard; Mueller, Felix; Mueller, James; Mueller, Ralph Soeren Peter; Mueller, Thibaut; Muenstermann, Daniel; Mullen, Paul; Mullier, Geoffrey; Munoz Sanchez, Francisca Javiela; Murillo Quijada, Javier Alberto; Murray, Bill; Musheghyan, Haykuhi; Muškinja, Miha; Myagkov, Alexey; Myska, Miroslav; Nachman, Benjamin Philip; Nackenhorst, Olaf; Nagai, Koichi; Nagai, Ryo; Nagano, Kunihiro; Nagasaka, Yasushi; Nagata, Kazuki; Nagel, Martin; Nagy, Elemer; Nairz, Armin Michael; Nakahama, Yu; Nakamura, Koji; Nakamura, Tomoaki; Nakano, Itsuo; Naranjo Garcia, Roger Felipe; Narayan, Rohin; Narrias Villar, Daniel Isaac; Naryshkin, Iouri; Naumann, Thomas; Navarro, Gabriela; Nayyar, Ruchika; Neal, Homer; Nechaeva, Polina; Neep, Thomas James; Negri, Andrea; Negrini, Matteo; Nektarijevic, Snezana; Nellist, Clara; Nelson, Andrew; Nemecek, Stanislav; Nemethy, Peter; Nepomuceno, Andre Asevedo; Nessi, Marzio; Neubauer, Mark; Neumann, Manuel; Neves, Ricardo; Nevski, Pavel; Newman, Paul; Nguyen, Duong Hai; Nguyen Manh, Tuan; Nickerson, Richard; Nicolaidou, Rosy; Nielsen, Jason; Nikiforov, Andriy; Nikolaenko, Vladimir; Nikolic-Audit, Irena; Nikolopoulos, Konstantinos; Nilsen, Jon Kerr; Nilsson, Paul; Ninomiya, Yoichi; Nisati, Aleandro; Nisius, Richard; Nobe, Takuya; Nomachi, Masaharu; Nomidis, Ioannis; Nooney, Tamsin; Norberg, Scarlet; Nordberg, Markus; Norjoharuddeen, Nurfikri; Novgorodova, Olga; Nowak, Sebastian; Nozaki, Mitsuaki; Nozka, Libor; Ntekas, Konstantinos; Nurse, Emily; Nuti, Francesco; O'grady, Fionnbarr; O'Neil, Dugan; O'Rourke, Abigail Alexandra; O'Shea, Val; Oakham, Gerald; Oberlack, Horst; Obermann, Theresa; Ocariz, Jose; Ochi, Atsuhiko; Ochoa, Ines; Ochoa-Ricoux, Juan Pedro; Oda, Susumu; Odaka, Shigeru; Ogren, Harold; Oh, Alexander; Oh, Seog; Ohm, Christian; Ohman, Henrik; Oide, Hideyuki; Okawa, Hideki; Okumura, Yasuyuki; Okuyama, Toyonobu; Olariu, Albert; Oleiro Seabra, Luis Filipe; Olivares Pino, Sebastian Andres; Oliveira Damazio, Denis; Olszewski, Andrzej; Olszowska, Jolanta; Onofre, António; Onogi, Kouta; Onyisi, Peter; Oreglia, Mark; Oren, Yona; Orestano, Domizia; Orlando, Nicola; Orr, Robert; Osculati, Bianca; Ospanov, Rustem; Otero y Garzon, Gustavo; Otono, Hidetoshi; Ouchrif, Mohamed; Ould-Saada, Farid; Ouraou, Ahmimed; Oussoren, Koen Pieter; Ouyang, Qun; Owen, Mark; Owen, Rhys Edward; Ozcan, Veysi Erkcan; Ozturk, Nurcan; Pachal, Katherine; Pacheco Pages, Andres; Pacheco Rodriguez, Laura; Padilla Aranda, Cristobal; Pagáčová, Martina; Pagan Griso, Simone; Paganini, Michela; Paige, Frank; Pais, Preema; Pajchel, Katarina; Palacino, Gabriel; Palazzo, Serena; Palestini, Sandro; Palka, Marek; Pallin, Dominique; Panagiotopoulou, Evgenia; Panagoulias, Ilias; Pandini, Carlo Enrico; Panduro Vazquez, William; Pani, Priscilla; Panitkin, Sergey; Pantea, Dan; Paolozzi, Lorenzo; Papadopoulou, Theodora; Papageorgiou, Konstantinos; Paramonov, Alexander; Paredes Hernandez, Daniela; Parker, Adam Jackson; Parker, Michael Andrew; Parker, Kerry Ann; Parodi, Fabrizio; Parsons, John; Parzefall, Ulrich; Pascuzzi, Vincent; Pasqualucci, Enrico; Passaggio, Stefano; Pastore, Francesca; Pásztor, Gabriella; Pataraia, Sophio; Pater, Joleen; Pauly, Thilo; Pearce, James; Pearson, Benjamin; Pedersen, Lars Egholm; Pedersen, Maiken; Pedraza Lopez, Sebastian; Pedro, Rute; Peleganchuk, Sergey; Penc, Ondrej; Peng, Cong; Peng, Haiping; Penwell, John; Peralva, Bernardo; Perego, Marta Maria; Perepelitsa, Dennis; Perez Codina, Estel; Perini, Laura; Pernegger, Heinz; Perrella, Sabrina; Peschke, Richard; Peshekhonov, Vladimir; Peters, Krisztian; Peters, Yvonne; Petersen, Brian; Petersen, Troels; Petit, Elisabeth; Petridis, Andreas; Petridou, Chariclia; Petroff, Pierre; Petrolo, Emilio; Petrov, Mariyan; Petrucci, Fabrizio; Pettersson, Nora Emilia; Peyaud, Alan; Pezoa, Raquel; Phillips, Peter William; Piacquadio, Giacinto; Pianori, Elisabetta; Picazio, Attilio; Piccaro, Elisa; Piccinini, Maurizio; Pickering, Mark Andrew; Piegaia, Ricardo; Pilcher, James; Pilkington, Andrew; Pin, Arnaud Willy J; Pinamonti, Michele; Pinfold, James; Pingel, Almut; Pires, Sylvestre; Pirumov, Hayk; Pitt, Michael; Plazak, Lukas; Pleier, Marc-Andre; Pleskot, Vojtech; Plotnikova, Elena; Pluth, Daniel; Poettgen, Ruth; Poggioli, Luc; Pohl, David-leon; Polesello, Giacomo; Poley, Anne-luise; Policicchio, Antonio; Polifka, Richard; Polini, Alessandro; Pollard, Christopher Samuel; Polychronakos, Venetios; Pommès, Kathy; Pontecorvo, Ludovico; Pope, Bernard; Popeneciu, Gabriel Alexandru; Poppleton, Alan; Pospisil, Stanislav; Potamianos, Karolos; Potrap, Igor; Potter, Christina; Potter, Christopher; Poulard, Gilbert; Poveda, Joaquin; Pozdnyakov, Valery; Pozo Astigarraga, Mikel Eukeni; Pralavorio, Pascal; Pranko, Aliaksandr; Prell, Soeren; Price, Darren; Price, Lawrence; Primavera, Margherita; Prince, Sebastien; Prokofiev, Kirill; Prokoshin, Fedor; Protopopescu, Serban; Proudfoot, James; Przybycien, Mariusz; Puddu, Daniele; Purohit, Milind; Puzo, Patrick; Qian, Jianming; Qin, Gang; Qin, Yang; Quadt, Arnulf; Quayle, William; Queitsch-Maitland, Michaela; Quilty, Donnchadha; Raddum, Silje; Radeka, Veljko; Radescu, Voica; Radhakrishnan, Sooraj Krishnan; Radloff, Peter; Rados, Pere; Ragusa, Francesco; Rahal, Ghita; Raine, John Andrew; Rajagopalan, Srinivasan; Rammensee, Michael; Rangel-Smith, Camila; Ratti, Maria Giulia; Rauch, Daniel; Rauscher, Felix; Rave, Stefan; Ravenscroft, Thomas; Ravinovich, Ilia; Raymond, Michel; Read, Alexander Lincoln; Readioff, Nathan Peter; Reale, Marilea; Rebuzzi, Daniela; Redelbach, Andreas; Redlinger, George; Reece, Ryan; Reed, Robert; Reeves, Kendall; Rehnisch, Laura; Reichert, Joseph; Reiss, Andreas; Rembser, Christoph; Ren, Huan; Rescigno, Marco; Resconi, Silvia; Rezanova, Olga; Reznicek, Pavel; Rezvani, Reyhaneh; Richter, Robert; Richter, Stefan; Richter-Was, Elzbieta; Ricken, Oliver; Ridel, Melissa; Rieck, Patrick; Riegel, Christian Johann; Rieger, Julia; Rifki, Othmane; Rijssenbeek, Michael; Rimoldi, Adele; Rimoldi, Marco; Rinaldi, Lorenzo; Ristić, Branislav; Ritsch, Elmar; Riu, Imma; Rizatdinova, Flera; Rizvi, Eram; Rizzi, Chiara; Robertson, Steven; Robichaud-Veronneau, Andree; Robinson, Dave; Robinson, James; Robson, Aidan; Roda, Chiara; Rodina, Yulia; Rodriguez Perez, Andrea; Rodriguez Rodriguez, Daniel; Roe, Shaun; Rogan, Christopher Sean; Røhne, Ole; Roloff, Jennifer; Romaniouk, Anatoli; Romano, Marino; Romano Saez, Silvestre Marino; Romero Adam, Elena; Rompotis, Nikolaos; Ronzani, Manfredi; Roos, Lydia; Ros, Eduardo; Rosati, Stefano; Rosbach, Kilian; Rose, Peyton; Rosien, Nils-Arne; Rossetti, Valerio; Rossi, Elvira; Rossi, Leonardo Paolo; Rosten, Jonatan; Rosten, Rachel; Rotaru, Marina; Roth, Itamar; Rothberg, Joseph; Rousseau, David; Rozanov, Alexandre; Rozen, Yoram; Ruan, Xifeng; Rubbo, Francesco; Rudolph, Matthew Scott; Rühr, Frederik; Ruiz-Martinez, Aranzazu; Rurikova, Zuzana; Rusakovich, Nikolai; Ruschke, Alexander; Russell, Heather; Rutherfoord, John; Ruthmann, Nils; Ryabov, Yury; Rybar, Martin; Rybkin, Grigori; Ryu, Soo; Ryzhov, Andrey; Rzehorz, Gerhard Ferdinand; Saavedra, Aldo; Sabato, Gabriele; Sacerdoti, Sabrina; Sadrozinski, Hartmut; Sadykov, Renat; Safai Tehrani, Francesco; Saha, Puja; Sahinsoy, Merve; Saimpert, Matthias; Saito, Tomoyuki; Sakamoto, Hiroshi; Sakurai, Yuki; Salamanna, Giuseppe; Salamon, Andrea; Salazar Loyola, Javier Esteban; Salek, David; Sales De Bruin, Pedro Henrique; Salihagic, Denis; Salnikov, Andrei; Salt, José; Salvatore, Daniela; Salvatore, Pasquale Fabrizio; Salvucci, Antonio; Salzburger, Andreas; Sammel, Dirk; Sampsonidis, Dimitrios; Sánchez, Javier; Sanchez Martinez, Victoria; Sanchez Pineda, Arturo; Sandaker, Heidi; Sandbach, Ruth Laura; Sandhoff, Marisa; Sandoval, Carlos; Sankey, Dave; Sannino, Mario; Sansoni, Andrea; Santoni, Claudio; Santonico, Rinaldo; Santos, Helena; Santoyo Castillo, Itzebelt; Sapp, Kevin; Sapronov, Andrey; Saraiva, João; Sarrazin, Bjorn; Sasaki, Osamu; Sato, Koji; Sauvan, Emmanuel; Savage, Graham; Savard, Pierre; Savic, Natascha; Sawyer, Craig; Sawyer, Lee; Saxon, James; Sbarra, Carla; Sbrizzi, Antonio; Scanlon, Tim; Scannicchio, Diana; Scarcella, Mark; Scarfone, Valerio; Schaarschmidt, Jana; Schacht, Peter; Schachtner, Balthasar Maria; Schaefer, Douglas; Schaefer, Leigh; Schaefer, Ralph; Schaeffer, Jan; Schaepe, Steffen; Schaetzel, Sebastian; Schäfer, Uli; Schaffer, Arthur; Schaile, Dorothee; Schamberger, R Dean; Scharf, Veit; Schegelsky, Valery; Scheirich, Daniel; Schernau, Michael; Schiavi, Carlo; Schier, Sheena; Schillo, Christian; Schioppa, Marco; Schlenker, Stefan; Schmidt-Sommerfeld, Korbinian Ralf; Schmieden, Kristof; Schmitt, Christian; Schmitt, Stefan; Schmitz, Simon; Schneider, Basil; Schnoor, Ulrike; Schoeffel, Laurent; Schoening, Andre; Schoenrock, Bradley Daniel; Schopf, Elisabeth; Schott, Matthias; Schouwenberg, Jeroen; Schovancova, Jaroslava; Schramm, Steven; Schreyer, Manuel; Schuh, Natascha; Schulte, Alexandra; Schultens, Martin Johannes; Schultz-Coulon, Hans-Christian; Schulz, Holger; Schumacher, Markus; Schumm, Bruce; Schune, Philippe; Schwartzman, Ariel; Schwarz, Thomas Andrew; Schweiger, Hansdieter; Schwemling, Philippe; Schwienhorst, Reinhard; Schwindling, Jerome; Schwindt, Thomas; Sciolla, Gabriella; Scuri, Fabrizio; Scutti, Federico; Searcy, Jacob; Seema, Pienpen; Seidel, Sally; Seiden, Abraham; Seifert, Frank; Seixas, José; Sekhniaidze, Givi; Sekhon, Karishma; Sekula, Stephen; Seliverstov, Dmitry; Semprini-Cesari, Nicola; Serfon, Cedric; Serin, Laurent; Serkin, Leonid; Sessa, Marco; Seuster, Rolf; Severini, Horst; Sfiligoj, Tina; Sforza, Federico; Sfyrla, Anna; Shabalina, Elizaveta; Shaikh, Nabila Wahab; Shan, Lianyou; Shang, Ruo-yu; Shank, James; Shapiro, Marjorie; Shatalov, Pavel; Shaw, Kate; Shaw, Savanna Marie; Shcherbakova, Anna; Shehu, Ciwake Yusufu; Sherwood, Peter; Shi, Liaoshan; Shimizu, Shima; Shimmin, Chase Owen; Shimojima, Makoto; Shirabe, Shohei; Shiyakova, Mariya; Shmeleva, Alevtina; Shoaleh Saadi, Diane; Shochet, Mel; Shojaii, Seyed Ruhollah; Shope, David Richard; Shrestha, Suyog; Shulga, Evgeny; Shupe, Michael; Sicho, Petr; Sickles, Anne Marie; Sidebo, Per Edvin; Sideras Haddad, Elias; Sidiropoulou, Ourania; Sidorov, Dmitri; Sidoti, Antonio; Siegert, Frank; Sijacki, Djordje; Silva, José; Silverstein, Samuel; Simak, Vladislav; Simic, Ljiljana; Simion, Stefan; Simioni, Eduard; Simmons, Brinick; Simon, Dorian; Simon, Manuel; Sinervo, Pekka; Sinev, Nikolai; Sioli, Maximiliano; Siragusa, Giovanni; Sivoklokov, Serguei; Sjölin, Jörgen; Skinner, Malcolm Bruce; Skottowe, Hugh Philip; Skubic, Patrick; Slater, Mark; Slavicek, Tomas; Slawinska, Magdalena; Sliwa, Krzysztof; Slovak, Radim; Smakhtin, Vladimir; Smart, Ben; Smestad, Lillian; Smiesko, Juraj; Smirnov, Sergei; Smirnov, Yury; Smirnova, Lidia; Smirnova, Oxana; Smith, Joshua Wyatt; Smith, Matthew; Smith, Russell; Smizanska, Maria; Smolek, Karel; Snesarev, Andrei; Snyder, Ian Michael; Snyder, Scott; Sobie, Randall; Socher, Felix; Soffer, Abner; Soh, Dart-yin; Sokhrannyi, Grygorii; Solans Sanchez, Carlos; Solar, Michael; Soldatov, Evgeny; Soldevila, Urmila; Solodkov, Alexander; Soloshenko, Alexei; Solovyanov, Oleg; Solovyev, Victor; Sommer, Philip; Son, Hyungsuk; Song, Hong Ye; Sood, Alexander; Sopczak, Andre; Sopko, Vit; Sorin, Veronica; Sosa, David; Sotiropoulou, Calliope Louisa; Soualah, Rachik; Soukharev, Andrey; South, David; Sowden, Benjamin; Spagnolo, Stefania; Spalla, Margherita; Spangenberg, Martin; Spanò, Francesco; Sperlich, Dennis; Spettel, Fabian; Spieker, Thomas Malte; Spighi, Roberto; Spigo, Giancarlo; Spiller, Laurence Anthony; Spousta, Martin; St Denis, Richard Dante; Stabile, Alberto; Stamen, Rainer; Stamm, Soren; Stanecka, Ewa; Stanek, Robert; Stanescu, Cristian; Stanescu-Bellu, Madalina; Stanitzki, Marcel Michael; Stapnes, Steinar; Starchenko, Evgeny; Stark, Giordon; Stark, Jan; Staroba, Pavel; Starovoitov, Pavel; Stärz, Steffen; Staszewski, Rafal; Steinberg, Peter; Stelzer, Bernd; Stelzer, Harald Joerg; Stelzer-Chilton, Oliver; Stenzel, Hasko; Stewart, Graeme; Stillings, Jan Andre; Stockton, Mark; Stoebe, Michael; Stoicea, Gabriel; Stolte, Philipp; Stonjek, Stefan; Stradling, Alden; Straessner, Arno; Stramaglia, Maria Elena; Strandberg, Jonas; Strandberg, Sara; Strandlie, Are; Strauss, Michael; Strizenec, Pavol; Ströhmer, Raimund; Strom, David; Stroynowski, Ryszard; Strubig, Antonia; Stucci, Stefania Antonia; Stugu, Bjarne; Styles, Nicholas Adam; Su, Dong; Su, Jun; Suchek, Stanislav; Sugaya, Yorihito; Suk, Michal; Sulin, Vladimir; Sultansoy, Saleh; Sumida, Toshi; Sun, Siyuan; Sun, Xiaohu; Sundermann, Jan Erik; Suruliz, Kerim; Suster, Carl; Sutton, Mark; Suzuki, Shota; Svatos, Michal; Swiatlowski, Maximilian; Swift, Stewart Patrick; Sykora, Ivan; Sykora, Tomas; Ta, Duc; Taccini, Cecilia; Tackmann, Kerstin; Taenzer, Joe; Taffard, Anyes; Tafirout, Reda; Taiblum, Nimrod; Takai, Helio; Takashima, Ryuichi; Takeshita, Tohru; Takubo, Yosuke; Talby, Mossadek; Talyshev, Alexey; Tan, Kong Guan; Tanaka, Junichi; Tanaka, Masahiro; Tanaka, Reisaburo; Tanaka, Shuji; Tanioka, Ryo; Tannenwald, Benjamin Bordy; Tapia Araya, Sebastian; Tapprogge, Stefan; Tarem, Shlomit; Tartarelli, Giuseppe Francesco; Tas, Petr; Tasevsky, Marek; Tashiro, Takuya; Tassi, Enrico; Tavares Delgado, Ademar; Tayalati, Yahya; Taylor, Aaron; Taylor, Geoffrey; Taylor, Pierre Thor Elliot; Taylor, Wendy; Teischinger, Florian Alfred; Teixeira-Dias, Pedro; Temming, Kim Katrin; Temple, Darren; Ten Kate, Herman; Teng, Ping-Kun; Teoh, Jia Jian; Tepel, Fabian-Phillipp; Terada, Susumu; Terashi, Koji; Terron, Juan; Terzo, Stefano; Testa, Marianna; Teuscher, Richard; Theveneaux-Pelzer, Timothée; Thomas, Juergen; Thomas-Wilsker, Joshuha; Thompson, Paul; Thompson, Stan; Thomsen, Lotte Ansgaard; Thomson, Evelyn; Tibbetts, Mark James; Ticse Torres, Royer Edson; Tikhomirov, Vladimir; Tikhonov, Yury; Timoshenko, Sergey; Tipton, Paul; Tisserant, Sylvain; Todome, Kazuki; Todorov, Theodore; Todorova-Nova, Sharka; Tojo, Junji; Tokár, Stanislav; Tokushuku, Katsuo; Tolley, Emma; Tomlinson, Lee; Tomoto, Makoto; Tompkins, Lauren; Toms, Konstantin; Tong, Baojia(Tony); Tornambe, Peter; Torrence, Eric; Torres, Heberth; Torró Pastor, Emma; Toth, Jozsef; Touchard, Francois; Tovey, Daniel; Trefzger, Thomas; Tricoli, Alessandro; Trigger, Isabel Marian; Trincaz-Duvoid, Sophie; Tripiana, Martin; Trischuk, William; Trocmé, Benjamin; Trofymov, Artur; Troncon, Clara; Trottier-McDonald, Michel; Trovatelli, Monica; Truong, Loan; Trzebinski, Maciej; Trzupek, Adam; Tseng, Jeffrey; Tsiareshka, Pavel; Tsipolitis, Georgios; Tsirintanis, Nikolaos; Tsiskaridze, Shota; Tsiskaridze, Vakhtang; Tskhadadze, Edisher; Tsui, Ka Ming; Tsukerman, Ilya; Tsulaia, Vakhtang; Tsuno, Soshi; Tsybychev, Dmitri; Tu, Yanjun; Tudorache, Alexandra; Tudorache, Valentina; Tulbure, Traian Tiberiu; Tuna, Alexander Naip; Tupputi, Salvatore; Turchikhin, Semen; Turgeman, Daniel; Turk Cakir, Ilkay; Turra, Ruggero; Tuts, Michael; Ucchielli, Giulia; Ueda, Ikuo; Ughetto, Michael; Ukegawa, Fumihiko; Unal, Guillaume; Undrus, Alexander; Unel, Gokhan; Ungaro, Francesca; Unno, Yoshinobu; Unverdorben, Christopher; Urban, Jozef; Urquijo, Phillip; Urrejola, Pedro; Usai, Giulio; Usui, Junya; Vacavant, Laurent; Vacek, Vaclav; Vachon, Brigitte; Valderanis, Chrysostomos; Valdes Santurio, Eduardo; Valencic, Nika; Valentinetti, Sara; Valero, Alberto; Valery, Loic; Valkar, Stefan; Valls Ferrer, Juan Antonio; Van Den Wollenberg, Wouter; Van Der Deijl, Pieter; van der Graaf, Harry; van Eldik, Niels; van Gemmeren, Peter; Van Nieuwkoop, Jacobus; van Vulpen, Ivo; van Woerden, Marius Cornelis; Vanadia, Marco; Vandelli, Wainer; Vanguri, Rami; Vaniachine, Alexandre; Vankov, Peter; Vardanyan, Gagik; Vari, Riccardo; Varnes, Erich; Varol, Tulin; Varouchas, Dimitris; Vartapetian, Armen; Varvell, Kevin; Vasquez, Jared Gregory; Vasquez, Gerardo; Vazeille, Francois; Vazquez Schroeder, Tamara; Veatch, Jason; Veeraraghavan, Venkatesh; Veloce, Laurelle Maria; Veloso, Filipe; Veneziano, Stefano; Ventura, Andrea; Venturi, Manuela; Venturi, Nicola; Venturini, Alessio; Vercesi, Valerio; Verducci, Monica; Verkerke, Wouter; Vermeulen, Jos; Vest, Anja; Vetterli, Michel; Viazlo, Oleksandr; Vichou, Irene; Vickey, Trevor; Vickey Boeriu, Oana Elena; Viehhauser, Georg; Viel, Simon; Vigani, Luigi; Villa, Mauro; Villaplana Perez, Miguel; Vilucchi, Elisabetta; Vincter, Manuella; Vinogradov, Vladimir; Vittori, Camilla; Vivarelli, Iacopo; Vlachos, Sotirios; Vlasak, Michal; Vogel, Marcelo; Vokac, Petr; Volpi, Guido; Volpi, Matteo; von der Schmitt, Hans; von Toerne, Eckhard; Vorobel, Vit; Vorobev, Konstantin; Vos, Marcel; Voss, Rudiger; Vossebeld, Joost; Vranjes, Nenad; Vranjes Milosavljevic, Marija; Vrba, Vaclav; Vreeswijk, Marcel; Vuillermet, Raphael; Vukotic, Ilija; Wagner, Peter; Wagner, Wolfgang; Wahlberg, Hernan; Wahrmund, Sebastian; Wakabayashi, Jun; Walder, James; Walker, Rodney; Walkowiak, Wolfgang; Wallangen, Veronica; Wang, Chao; Wang, Chao; Wang, Fuquan; Wang, Haichen; Wang, Hulin; Wang, Jike; Wang, Jin; Wang, Kuhan; Wang, Rui; Wang, Song-Ming; Wang, Tingting; Wang, Wenxiao; Wanotayaroj, Chaowaroj; Warburton, Andreas; Ward, Patricia; Wardrope, David Robert; Washbrook, Andrew; Watkins, Peter; Watson, Alan; Watson, Miriam; Watts, Gordon; Watts, Stephen; Waugh, Ben; Webb, Samuel; Weber, Michele; Weber, Stefan Wolf; Weber, Stephen; Webster, Jordan S; Weidberg, Anthony; Weinert, Benjamin; Weingarten, Jens; Weiser, Christian; Weits, Hartger; Wells, Phillippa; Wenaus, Torre; Wengler, Thorsten; Wenig, Siegfried; Wermes, Norbert; Werner, Michael David; Werner, Per; Wessels, Martin; Wetter, Jeffrey; Whalen, Kathleen; Whallon, Nikola Lazar; Wharton, Andrew Mark; White, Andrew; White, Martin; White, Ryan; Whiteson, Daniel; Wickens, Fred; Wiedenmann, Werner; Wielers, Monika; Wiglesworth, Craig; Wiik-Fuchs, Liv Antje Mari; Wildauer, Andreas; Wilk, Fabian; Wilkens, Henric George; Williams, Hugh; Williams, Sarah; Willis, Christopher; Willocq, Stephane; Wilson, John; Wingerter-Seez, Isabelle; Winklmeier, Frank; Winston, Oliver James; Winter, Benedict Tobias; Wittgen, Matthias; Wolf, Tim Michael Heinz; Wolff, Robert; Wolter, Marcin Wladyslaw; Wolters, Helmut; Worm, Steven D; Wosiek, Barbara; Wotschack, Jorg; Woudstra, Martin; Wozniak, Krzysztof; Wu, Mengqing; Wu, Miles; Wu, Sau Lan; Wu, Xin; Wu, Yusheng; Wyatt, Terry Richard; Wynne, Benjamin; Xella, Stefania; Xi, Zhaoxu; Xu, Da; Xu, Lailin; Yabsley, Bruce; Yacoob, Sahal; Yamaguchi, Daiki; Yamaguchi, Yohei; Yamamoto, Akira; Yamamoto, Shimpei; Yamanaka, Takashi; Yamauchi, Katsuya; Yamazaki, Yuji; Yan, Zhen; Yang, Haijun; Yang, Hongtao; Yang, Yi; Yang, Zongchang; Yao, Weiming; Yap, Yee Chinn; Yasu, Yoshiji; Yatsenko, Elena; Yau Wong, Kaven Henry; Ye, Jingbo; Ye, Shuwei; Yeletskikh, Ivan; Yildirim, Eda; Yorita, Kohei; Yoshida, Rikutaro; Yoshihara, Keisuke; Young, Charles; Young, Christopher John; Youssef, Saul; Yu, David Ren-Hwa; Yu, Jaehoon; Yu, Jiaming; Yu, Jie; Yuan, Li; Yuen, Stephanie P; Yusuff, Imran; Zabinski, Bartlomiej; Zacharis, George; Zaidan, Remi; Zaitsev, Alexander; Zakharchuk, Nataliia; Zalieckas, Justas; Zaman, Aungshuman; Zambito, Stefano; Zanello, Lucia; Zanzi, Daniele; Zeitnitz, Christian; Zeman, Martin; Zemla, Andrzej; Zeng, Jian Cong; Zeng, Qi; Zenin, Oleg; Ženiš, Tibor; Zerwas, Dirk; Zhang, Dongliang; Zhang, Fangzhou; Zhang, Guangyi; Zhang, Huijun; Zhang, Jinlong; Zhang, Lei; Zhang, Liqing; Zhang, Matt; Zhang, Rui; Zhang, Ruiqi; Zhang, Xueyao; Zhang, Zhiqing; Zhao, Xiandong; Zhao, Yongke; Zhao, Zhengguo; Zhemchugov, Alexey; Zhong, Jiahang; Zhou, Bing; Zhou, Chen; Zhou, Lei; Zhou, Li; Zhou, Mingliang; Zhou, Ning; Zhu, Cheng Guang; Zhu, Hongbo; Zhu, Junjie; Zhu, Yingchun; Zhuang, Xuai; Zhukov, Konstantin; Zibell, Andre; Zieminska, Daria; Zimine, Nikolai; Zimmermann, Christoph; Zimmermann, Stephanie; Zinonos, Zinonas; Zinser, Markus; Ziolkowski, Michael; Živković, Lidija; Zobernig, Georg; Zoccoli, Antonio; zur Nedden, Martin; Zwalinski, Lukasz

    2017-05-18

    During 2015 the ATLAS experiment recorded $3.8 \\mathrm{fb}^{-1}$ of proton--proton collision data at a centre-of-mass energy of $13 \\mathrm{TeV}$. The ATLAS trigger system is a crucial component of the experiment, responsible for selecting events of interest at a recording rate of approximately 1 kHz from up to 40 MHz of collisions. This paper presents a short overview of the changes to the trigger and data acquisition systems during the first long shutdown of the LHC and shows the performance of the trigger system and its components based on the 2015 proton--proton collision data.

  9. Di-boson production measurements with the ATLAS detector

    CERN Document Server

    Chevalier, Laurent; The ATLAS collaboration

    2015-01-01

    Measurements diboson production cross sections in proton-proton interactions at 7 and 8 TeV are reported from the ATLAS experiment. The cross section results are measured in phase space regions defined by the decay kinematics and then extrapolated to the full phase spaces. Cross sections for WV (V=W or Z) production in the leptonic or semileptonic channels are compared to (N)NLO predictions of the Standard Model and are used to place constraints on anomalous triple-gauge-boson couplings. First LHC Run-2 results will be included if available.

  10. PD2P : PanDA Dynamic Data Placement for ATLAS

    OpenAIRE

    Maeno, T; De, K; Panitkin, S

    2012-01-01

    The PanDA (Production and Distributed Analysis) system plays a key role in the ATLAS distributed computing infrastructure. PanDA is the ATLAS workload management system for processing all Monte-Carlo (MC) simulation and data reprocessing jobs in addition to user and group analysis jobs. The PanDA Dynamic Data Placement (PD2P) system has been developed to cope with difficulties of data placement for ATLAS. We will describe the design of the new system, its performance during the past year of d...

  11. Supervision of the ATLAS High Level Trigger System

    CERN Document Server

    Wheeler, S.; Meessen, C.; Qian, Z.; Touchard, F.; Negri, France A.; Zobernig, H.; CHEP 2003 Computing in High Energy Physics; Negri, France A.

    2003-01-01

    The ATLAS High Level Trigger (HLT) system provides software-based event selection after the initial LVL1 hardware trigger. It is composed of two stages, the LVL2 trigger and the Event Filter. The HLT is implemented as software tasks running on large processor farms. An essential part of the HLT is the supervision system, which is responsible for configuring, coordinating, controlling and monitoring the many hundreds of processes running in the HLT. A prototype implementation of the supervision system, using tools from the ATLAS Online Software system is presented. Results from scalability tests are also presented where the supervision system was shown to be capable of controlling over 1000 HLT processes running on 230 nodes.

  12. The ATLAS Distributed Data Management project: Past and Future

    CERN Document Server

    Garonne, V; The ATLAS collaboration; Lassnig, M; Molfetas, A; Barisits, M; Beermann, T; Nairz, A; Goossens, L; Barreiro Megino, F; Serfon, C; Oleynik, D; Petrosyan, A

    2012-01-01

    ATLAS has recorded almost 8PB of RAW data since the LHC started running at the end of 2009. Many more derived data products and complimentary simulation data have also been produced by the collaboration and, in total, 90PB is currently stored in the Worldwide LHC Computing Grid by ATLAS. All this data is managed by the ATLAS Distributed Data Management system, called Don Quijote 2 (DQ2). DQ2 has evolved rapidly to help ATLAS Computing operations manage these large quantities of data across the many grid sites at which ATLAS runs, and to help ATLAS physicists get access to this data. In this paper, we describe new and improved DQ2 services, and the experience of data management operation in ATLAS computing, showing how these services enable the management of petabyte scale computing operations. We also present the concepts of the new version of the ATLAS Distributed Data Management (DDM) system, Rucio.

  13. The ATLAS Distributed Data Management project: Past and Future

    International Nuclear Information System (INIS)

    Garonne, Vincent; Stewart, Graeme A; Lassnig, Mario; Molfetas, Angelos; Barisits, Martin; Beermann, Thomas; Nairz, Armin; Goossens, Luc; Barreiro Megino, Fernando; Serfon, Cedric; Oleynik, Danila; Petrosyan, Artem

    2012-01-01

    ATLAS has recorded more than 8 petabyte(PB) of RAW data since the LHC started running at the end of 2009. Many more derived data products and complimentary simulation data have also been produced by the collaboration and, in total, 90PB are currently stored in the Worldwide LHC Computing Grid by ATLAS. All these data are managed by the ATLAS Distributed Data Management system, called Don Quijote 2 (DQ2). DQ2 has evolved rapidly to help ATLAS Computing operations manage these large quantities of data across the many grid sites at which ATLAS runs, and to help ATLAS physicists get access to these data. In this paper, we describe new and improved DQ2 services, and the experience of data management operation in ATLAS computing, showing how these services enable the management of PB scale computing operations. We also present the concepts of the new version of the ATLAS Distributed Data Management (DDM) system, Rucio.

  14. The ATLAS Distributed Data Management project: Past and Future

    Science.gov (United States)

    Garonne, Vincent; Stewart, Graeme A.; Lassnig, Mario; Molfetas, Angelos; Barisits, Martin; Beermann, Thomas; Nairz, Armin; Goossens, Luc; Barreiro Megino, Fernando; Serfon, Cedric; Oleynik, Danila; Petrosyan, Artem

    2012-12-01

    ATLAS has recorded more than 8 petabyte(PB) of RAW data since the LHC started running at the end of 2009. Many more derived data products and complimentary simulation data have also been produced by the collaboration and, in total, 90PB are currently stored in the Worldwide LHC Computing Grid by ATLAS. All these data are managed by the ATLAS Distributed Data Management system, called Don Quijote 2 (DQ2). DQ2 has evolved rapidly to help ATLAS Computing operations manage these large quantities of data across the many grid sites at which ATLAS runs, and to help ATLAS physicists get access to these data. In this paper, we describe new and improved DQ2 services, and the experience of data management operation in ATLAS computing, showing how these services enable the management of PB scale computing operations. We also present the concepts of the new version of the ATLAS Distributed Data Management (DDM) system, Rucio.

  15. Operation of the ATLAS distributed computing

    CERN Document Server

    Barreiro Megino, Fernando Harald; The ATLAS collaboration

    2018-01-01

    We describe the central operation of the ATLAS distributed computing system. The majority of compute intensive activities within ATLAS are carried out on some 350,000 CPU cores on the Grid, augmented by opportunistic usage of significant HPC and volunteer resources. The increasing scale, and challenging new payloads, demand fine-tuning of operational procedures together with timely developments of the production system. We describe several such developments, motivated directly from operational experience. Optimization of inefficient task requests, from both official production and users, is made possible by automatic detection of payload properties. User education, job shaping or preventative throttling help to increase the overall throughput of the available resources.

  16. Measurements of the Vector boson production with the ATLAS Detector

    CERN Document Server

    Lapertosa, Alessandro; The ATLAS collaboration

    2017-01-01

    Measurements of the Drell-Yan production of W and Z/γ bosons at the LHC provide a benchmark of our understanding of perturbative QCD and probe the proton structure in a unique way. The ATLAS collaboration has performed new high precision measurements at center-of-mass energies of 7 TeV. The measurements are performed for W+, W− and Z/γ bosons integrated and as a function of the boson or lepton rapidity and the Z/γ∗ mass. Unprecedented precision is reached and strong constraints on Parton Distribution functions, in particular the strange density are found. Z cross sections are also measured at a center-of-mass energies of 8 TeV and 13 TeV, and cross-section ratios to the top-quark pair production have been derived. This ratio measurement leads to a cancellation of systematic effects and allows for a high precision comparison to the theory predictions. The production of jets in association with vector bosons is a further important process to study perturbative QCD in a multi-scale environment. The ATLAS ...

  17. Recent ATLAS results on inclusive and differential top quark production

    CERN Multimedia

    CERN. Geneva

    2016-01-01

    Recent results from measurements of top quark pair and single top production by the ATLAS experiment are presented. Final inclusive and differential results from Run 1 are shown along with a first look at the 13 TeV data. Measurements of single top quark production in all three possible modes have been performed at ATLAS at 8 TeV. The first 13 TeV data was used to measure the t-channel production cross section, showing good agreement with the prediction. The Run 1 data was used to measure particle and parton level differential cross sections in top-pair events over a wide range of phase-space up to high masses, high momentum or high jet multiplicity. These results are compared to state-of-the-art NLO Monte Carlo predictions. A number of parton level results are compared to recent NNLO differential distributions, which show good agreement with the data. Results in both the lepton-plus-jets and dilepton channels are shown. The 2015 data at 13 TeV has allowed us to measure the additional jet multiplicity in dile...

  18. ATLAS distributed computing: experience and evolution

    International Nuclear Information System (INIS)

    Nairz, A

    2014-01-01

    The ATLAS experiment has just concluded its first running period which commenced in 2010. After two years of remarkable performance from the LHC and ATLAS, the experiment has accumulated more than 25 fb −1 of data. The total volume of beam and simulated data products exceeds 100 PB distributed across more than 150 computing centres around the world, managed by the experiment's distributed data management system. These sites have provided up to 150,000 computing cores to ATLAS's global production and analysis processing system, enabling a rich physics programme including the discovery of the Higgs-like boson in 2012. The wealth of accumulated experience in global data-intensive computing at this massive scale, and the considerably more challenging requirements of LHC computing from 2015 when the LHC resumes operation, are driving a comprehensive design and development cycle to prepare a revised computing model together with data processing and management systems able to meet the demands of higher trigger rates, energies and event complexities. An essential requirement will be the efficient utilisation of current and future processor technologies as well as a broad range of computing platforms, including supercomputing and cloud resources. We will report on experience gained thus far and our progress in preparing ATLAS computing for the future

  19. Evaluation and proposal of improvement for the measurement system in ATLAS

    International Nuclear Information System (INIS)

    Cho, Dong Woo; Kim, Jong Rok; Park, Jun Kwon

    2007-03-01

    The project independently evaluated the validities and reliability of measurement system in ATLAS, then proposed plans to improve the measurement system from evaluated results. For this objectives, we evaluated the design, technical backgrounds, verifying data of measurement system in ATLAS. From this evaluation, we proposed plans for improvement on parts which need improvement

  20. ATLAS results on diffraction and exclusive production

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00224260; The ATLAS collaboration

    2017-01-01

    Various aspects of forward physics have been studied by the ATLAS collaboration using data from Run I at the LHC. In this text, main results of three published analyses are summarized, based on data from proton-proton collisions at $\\sqrt{s} = 7$ or 8 TeV collected between 2010 and 2012. One analysis deals with diffractive signature with at least two jets in the final state, the other two study exclusive production of a pair of leptons or W bosons.

  1. The Message Reporting System of the ATLAS DAQ System

    CERN Document Server

    Caprini, M; Kolos, S; 10th ICATPP Conference on Astroparticle, Particle, Space Physics, Detectors and Medical Physics Applications

    2008-01-01

    The Message Reporting System (MRS) in the ATLAS data acquisition system (DAQ) is one package of the Online Software which acts as a glue of various elements of DAQ, High Level Trigger (HLT) and Detector Control System (DCS). The aim of the MRS is to provide a facility which allows all software components in ATLAS to report messages to other components of the distributed DAQ system. The processes requiring a MRS are on one hand applications that report error conditions or information and on the other hand message processors that receive reported messages. A message reporting application can inject one or more messages into the MRS at any time. An application wishing to receive messages can subscribe to a message group according to defined criteria. The application receives messages that fulfill the subscription criteria when they are reported to MRS. The receiver message processing can consist of anything from simply logging the messages in a file/terminal to performing message analysis. The inter-process comm...

  2. ATLAS Job Transforms

    CERN Document Server

    Stewart, G A; The ATLAS collaboration; Maddocks, H J; Harenberg, T; Sandhoff, M; Sarrazin, B

    2013-01-01

    The need to run complex workflows for a high energy physics experiment such as ATLAS has always been present. However, as computing resources have become even more constrained, compared to the wealth of data generated by the LHC, the need to use resources efficiently and manage complex workflows within a single grid job have increased. In ATLAS, a new Job Transform framework has been developed that we describe in this paper. This framework manages the multiple execution steps needed to `transform' one data type into another (e.g., RAW data to ESD to AOD to final ntuple) and also provides a consistent interface for the ATLAS production system. The new framework uses a data driven workflow definition which is both easy to manage and powerful. After a transform is defined, jobs are expressed simply by specifying the input data and the desired output data. The transform infrastructure then executes only the necessary substeps to produce the final data products. The global execution cost of running the job is mini...

  3. ATLAS Job Transforms

    CERN Document Server

    Stewart, G A; The ATLAS collaboration; Maddocks, H J; Harenberg, T; Sandhoff, M; Sarrazin, B

    2013-01-01

    The need to run complex workflows for a high energy physics experiment such as ATLAS has always been present. However, as computing resources have become even more constrained, compared to the wealth of data generated by the LHC, the need to use resources efficiently and manage complex workflows within a single grid job have increased. In ATLAS, a new Job Transform framework has been developed that we describe in this paper. This framework manages the multiple execution steps needed to 'transform' one data type into another (e.g., RAW data to ESD to AOD to final ntuple) and also provides a consistent interface for the ATLAS production system. The new framework uses a data driven workflow definition which is both easy to manage and powerful. After a transform is defined, jobs are expressed simply by specifying the input data and the desired output data. The transform infrastructure then executes only the necessary substeps to produce the final data products. The global execution cost of running the job is mini...

  4. Associated top production (ttV, tV) in ATLAS and CMS

    CERN Document Server

    Sánchez Cruz, Sergio

    2018-01-01

    A set of measurements of top quark pair and single top quark production in association with standard model (SM) bosons performed by the CMS and ATLAS Collaborations is presented. The consistency of these results among themselves and with the SM prediction is discussed. The latest results on four top quark production are also shown.

  5. Single top quark production cross section using the ATLAS detector at the LHC

    CERN Document Server

    Cioara, Irina Antonela; The ATLAS collaboration

    2017-01-01

    Measurements of single top-quark production in proton-proton collisions are presented based on the 8 TeV and 13 TeV ATLAS datasets. In the leading order process, a W boson is exchanged in the t-channel. The cross-section for the production of single top-quarks and single anti-top-quarks total production cross sections, their ratio, as well as a measurement of the inclusive production cross section is presented. At 8 TeV, differential cross-section measurements of the t-channel process are also reported, these analyses include limits on anomalous contributions to the Wtb vertex and measurement of the top quark polarization. A measurement of the production cross section of a single top quark in association with a W boson, the second largest single-top production mode, is also presented. Finally, evidence for s-channel single-top production in the 8 TeV ATLAS dataset is presented. All measurements are compared to state-of-the-art theoretical calculations.

  6. Studies of quarkonium production and polarisation with early data at ATLAS

    CERN Document Server

    Price, Darren David

    2008-01-01

    This thesis presents a range of studies of the production cross-section and polarisation of quarkonium at the ATLAS experiment at the LHC. As data-taking has not yet begun, these studies are conducted on fully simulated and reconstructed Monte Carlo samples. The main studies of this thesis are an analysis of the production cross-sections of J/psi and Upsilon mesons at ATLAS with 10 pb^-1 of data, along with an analysis of the polarisation of the J/psi and Upsilon that can provide information towards understanding the underlying QCD production mechanism. In addition, we are able to predict the observability of the three chi_c states through its radiative decays to J/psi. I present details of a new method of data-driven muon reconstruction efficiency determination using Inner Detector tracking information in quarkonium decays, and studies into the uses of quarkonia for detector performance, commissioning and data quality monitoring. Using a new method of measuring the polarisation distribution that reduces syst...

  7. The ATLAS Forward Physics Program

    OpenAIRE

    Royon, C

    2010-01-01

    After a brief review of the approved ATLAS forward detector system we describe the main ATLAS forward physics program. This program currently includes such topics as soft and hard diffraction, double pomeron exchange, central exclusive production, rapidity gap survival, two photon physics, the determination of the total cross-section and the determination of the absolute luminosity A possible high luminosity upgrade program involving new forward proton detectors is also briefly reviewed. This...

  8. Detector Control System for the AFP detector in ATLAS experiment at CERN

    Science.gov (United States)

    Banaś, E.; Caforio, D.; Czekierda, S.; Hajduk, Z.; Olszowska, J.; Seabra, L.; Šícho, P.

    2017-10-01

    The ATLAS Forward Proton (AFP) detector consists of two forward detectors located at 205 m and 217 m on either side of the ATLAS experiment. The aim is to measure the momenta and angles of diffractively scattered protons. In 2016, two detector stations on one side of the ATLAS interaction point were installed and commissioned. The detector infrastructure and necessary services were installed and are supervised by the Detector Control System (DCS), which is responsible for the coherent and safe operation of the detector. A large variety of used equipment represents a considerable challenge for the AFP DCS design. Industrial Supervisory Control and Data Acquisition (SCADA) product Siemens WinCCOA, together with the CERN Joint Control Project (JCOP) framework and standard industrial and custom developed server applications and protocols are used for reading, processing, monitoring and archiving of the detector parameters. Graphical user interfaces allow for overall detector operation and visualization of the detector status. Parameters, important for the detector safety, are used for alert generation and interlock mechanisms.

  9. ATLAS barrel hadron tile calorimeter: spacers plates mass production

    International Nuclear Information System (INIS)

    Artikov, A.M.; Budagov, Yu.A.; Khubua, J.

    1999-01-01

    In this article we expose the main problems of the mass production of the so-called 'spacer plates' for the ATLAS Barrel Hadron Tile Calorimeter. We describe all practical solutions of these problems. Particularly we present the measurement procedures and calculation schemes we used for the spacers dimensions determination. The results of the calculations are presented

  10. Results on QCD jet production at ATLAS and CMS

    International Nuclear Information System (INIS)

    Meyer, C.J.

    2014-01-01

    The production of jets at the Large Hadron Collider (LHC) at √(s)=7 TeV is summarized, including results from both ATLAS and CMS detectors. Current jet performance is described, followed by inclusive jet and multi-jet measurements in various final state configurations. Finally some results on heavy flavour and jet substructure are presented. (author)

  11. Measurements of photon and jet production properties with ATLAS

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00237016; The ATLAS collaboration

    2016-09-09

    Summary of recent ATLAS measurements of jet and photon production using proton–proton (pp) collisions from the Large Hadron Collider. The charged-particle multiplicity in jets, and jet charge measurements are presented using 8 TeV pp collisions. Differential measurements of jet and photon cross-sections are shown for 7, 8 and 13 TeV pp collisions.

  12. Measurement of W/Z production with the ATLAS detector

    CERN Document Server

    Sauvan, JB; The ATLAS collaboration

    2012-01-01

    The production of W and Z bosons at the LHC has been measured with the ATLAS detector. These measurements include inclusive and differential cross sections, $W$ polarisation, and the polarisation of tau leptons from W decays. They show sensitivity to the proton structure and are also used to test predictions from perturbative QCD and phenomenological models.

  13. Exploring the human body space: A geographical information system based anatomical atlas

    Directory of Open Access Journals (Sweden)

    Antonio Barbeito

    2016-06-01

    Full Text Available Anatomical atlases allow mapping the anatomical structures of the human body. Early versions of these systems consisted of analogical representations with informative text and labeled images of the human body. With computer systems, digital versions emerged and the third and fourth dimensions were introduced. Consequently, these systems increased their efficiency, allowing more realistic visualizations with improved interactivity and functionality. The 4D atlases allow modeling changes over time on the structures represented. The anatomical atlases based on geographic information system (GIS environments allow the creation of platforms with a high degree of interactivity and new tools to explore and analyze the human body. In this study we expand the functions of a human body representation system by creating new vector data, topology, functions, and an improved user interface. The new prototype emulates a 3D GIS with a topological model of the human body, replicates the information provided by anatomical atlases, and provides a higher level of functionality and interactivity. At this stage, the developed system is intended to be used as an educational tool and integrates into the same interface the typical representations of surface and sectional atlases.

  14. IT Infrastructure Design and Implementation Considerations for the ATLAS TDAQ System

    CERN Document Server

    Dobson, M; The ATLAS collaboration; Caramarcu, C; Dumitru, I; Valsan, L; Darlea, G L; Bujor, F; Bogdanchikov, A G; Korol, A A; Zaytsev, A S; Ballestrero, S

    2013-01-01

    This paper gives a thorough overview of the ATLAS TDAQ SysAdmin group activities which deals with administration of the TDAQ computing environment supporting Front End detector hardware, Data Flow, Event Filter and other subsystems of the ATLAS detector operating on the LHC accelerator at CERN. The current installation consists of approximately 1500 netbooted nodes managed by more than 60 dedicated servers, a high performance centralized storage system, about 50 multi-screen user interface systems installed in the control rooms and various hardware and critical service monitoring machines. In the final configuration, the online computer farm will be capable of hosting tens of thousands applications running simultaneously. The ATLAS TDAQ computing environment is now serving more than 3000 users subdivided into approximately 300 categories in correspondence with their roles in the system. The access and role management system is custom built on top of an LDAP schema. The engineering infrastructure of the ATLAS ...

  15. NODC Standard Product: World Ocean Atlas 2001 (6 disc set) (NODC Accession 0095600)

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — The World Ocean Atlas 2001 (WOA01) Series consists of two sets of products. The first set of products consists of three DATA CD-ROMs containing global data...

  16. Experience with ATLAS MySQL PanDA database service

    International Nuclear Information System (INIS)

    Smirnov, Y; Wlodek, T; Hover, J; Smith, J; Wenaus, T; Yu, D; De, K; Ozturk, N

    2010-01-01

    The PanDA distributed production and analysis system has been in production use for ATLAS data processing and analysis since late 2005 in the US, and globally throughout ATLAS since early 2008. Its core architecture is based on a set of stateless web services served by Apache and backed by a suite of MySQL databases that are the repository for all PanDA information: active and archival job queues, dataset and file catalogs, site configuration information, monitoring information, system control parameters, and so on. This database system is one of the most critical components of PanDA, and has successfully delivered the functional and scaling performance required by PanDA, currently operating at a scale of half a million jobs per week, with much growth still to come. In this paper we describe the design and implementation of the PanDA database system, its architecture of MySQL servers deployed at BNL and CERN, backup strategy and monitoring tools. The system has been developed, thoroughly tested, and brought to production to provide highly reliable, scalable, flexible and available database services for ATLAS Monte Carlo production, reconstruction and physics analysis.

  17. Experience with ATLAS MySQL PanDA database service

    Energy Technology Data Exchange (ETDEWEB)

    Smirnov, Y; Wlodek, T; Hover, J; Smith, J; Wenaus, T; Yu, D [Physics Department, Brookhaven National Laboratory, Upton, NY 11973-5000 (United States); De, K; Ozturk, N [Department of Physics, University of Texas at Arlington, Arlington, TX, 76019 (United States)

    2010-04-01

    The PanDA distributed production and analysis system has been in production use for ATLAS data processing and analysis since late 2005 in the US, and globally throughout ATLAS since early 2008. Its core architecture is based on a set of stateless web services served by Apache and backed by a suite of MySQL databases that are the repository for all PanDA information: active and archival job queues, dataset and file catalogs, site configuration information, monitoring information, system control parameters, and so on. This database system is one of the most critical components of PanDA, and has successfully delivered the functional and scaling performance required by PanDA, currently operating at a scale of half a million jobs per week, with much growth still to come. In this paper we describe the design and implementation of the PanDA database system, its architecture of MySQL servers deployed at BNL and CERN, backup strategy and monitoring tools. The system has been developed, thoroughly tested, and brought to production to provide highly reliable, scalable, flexible and available database services for ATLAS Monte Carlo production, reconstruction and physics analysis.

  18. arXiv Production and Integration of the ATLAS Insertable B-Layer

    CERN Document Server

    Abbott, B.; Alberti, F.; Alex, M.; Alimonti, G.; Alkire, S.; Allport, P.; Altenheiner, S.; Ancu, L.S.; Anderssen, E.; Andreani, A.; Andreazza, A.; Axen, B.; Arguin, J.; Backhaus, M.; Balbi, G.; Ballansat, J.; Barbero, M.; Barbier, G.; Bassalat, A.; Bates, R.; Baudin, P.; Battaglia, M.; Beau, T.; Beccherle, R.; Bell, A.; Benoit, M.; Bermgan, A.; Bertsche, C.; Bertsche, D.; Bilbao de Mendizabal, J.; Bindi, F.; Bomben, M.; Borri, M.; Bortolin, C.; Bousson, N.; Boyd, R.G.; Breugnon, P.; Bruni, G.; Brossamer, J.; Bruschi, M.; Buchholz, P.; Budun, E.; Buttar, C.; Cadoux, F.; Calderini, G.; Caminada, L.; Capeans, M.; Carney, R.; Casse, G.; Catinaccio, A.; Cavalli-Sforza, M.; Červ, M.; Cervelli, A.; Chau, C.C.; Chauveau, J.; Chen, S.P.; Chu, M.; Ciapetti, M.; Cindro, V.; Citterio, M.; Clark, A.; Cobal, M.; Coelli, S.; Collot, J.; Crespo-Lopez, O.; Dalla Betta, G.F.; Daly, C.; D'Amen, G.; Dann, N.; Dao, V.; Darbo, G.; DaVia, C.; David, P.; Debieux, S.; Delebecque, P.; De Lorenzi, F.; de Oliveira, R.; Dette, K.; Dietsche, W.; Di Girolamo, B.; Dinu, N.; Dittus, F.; Diyakov, D.; Djama, F.; Dobos, D.; Dondero, P.; Doonan, K.; Dopke, J.; Dorholt, O.; Dube, S.; Dzahini, D.; Egorov, K.; Ehrmann, O.; Einsweiler, K.; Elles, S.; Elsing, M.; Eraud, L.; Ereditato, A.; Eyring, A.; Falchieri, D.; Falou, A.; Fausten, C.; Favareto, A.; Favre, Y.; Feigl, S.; Fernandez Perez, S.; Ferrere, D.; Fleury, J.; Flick, T.; Forshaw, D.; Fougeron, D.; Franconi, L.; Gabrielli, A.; Gaglione, R.; Gallrapp, C.; Gan, K.K.; Garcia-Sciveres, M.; Gariano, G.; Gastaldi, T.; Gavrilenko, I.; Gaudiello, A.; Geffroy, N.; Gemme, C.; Gensolen, F.; George, M.; Ghislain, P.; Giangiacomi, N.; Gibson, S.; Giordani, M.P.; Giugni, D.; Gjersdal, H.; Glitza, K.W.; Gnani, D.; Godlewski, J.; Gonella, L.; Gonzalez-Sevilla, S.; Gorelov, I.; Gorišek, A.; Gössling, C.; Grancagnolo, S.; Gray, H.; Gregor, I.; Grenier, P.; Grinstein, S.; Gris, A.; Gromov, V.; Grondin, D.; Grosse-Knetter, J.; Guescini, F.; Guido, E.; Gutierrez, P.; Hallewell, G.; Hartman, N.; Hauck, S.; Hasi, J.; Hasib, A.; Hegner, F.; Heidbrink, S.; Heim, T.; Heinemann, B.; Hemperek, T.; Hessey, N.P.; Hetmánek, M.; Hinman, R.R.; Hoeferkamp, M.; Holmes, T.; Hostachy, J.; Hsu, S.C.; Hügging, F.; Husi, C.; Iacobucci, G.; Ibragimov, I.; Idarraga, J.; Ikegami, Y.; Ince, T.; Ishmukhametov, R.; Izen, J.M.; Janoška, Z.; Janssen, J.; Jansen, L.; Jeanty, L.; Jensen, F.; Jentzsch, J.; Jezequel, S.; Joseph, J.; Kagan, H.; Kagan, M.; Karagounis, M.; Kass, R.; Kastanas, A.; Kenney, C.; Kersten, S.; Kind, P.; Klein, M.; Klingenberg, R.; Kluit, R.; Kocian, M.; Koffeman, E.; Korchak, O.; Korolkov, I.; Kostyukhina-Visoven, I.; Kovalenko, S.; Kretz, M.; Krieger, N.; Krüger, H.; Kruth, A.; Kugel, A.; Kuykendall, W.; La Rosa, A.; Lai, C.; Lantzsch, K.; Lapoire, C.; Laporte, D.; Lari, T.; Latorre, S.; Leyton, M.; Lindquist, B.; Looper, K.; Lopez, I.; Lounis, A.; Lu, Y.; Lubatti, H.J.; Maeland, S.; Maier, A.; Mallik, U.; Manca, F.; Mandelli, B.; Mandić, I.; Marchand, D.; Marchiori, G.; Marx, M.; Massol, N.; Mättig, P.; Mayer, J.; Mc Goldrick, G.; Mekkaoui, A.; Menouni, M.; Menu, J.; Meroni, C.; Mesa, J.; Michal, S.; Miglioranzi, S.; Mikuž, M.; Miucci, A.; Mochizuki, K.; Monti, M.; Moore, J.; Morettini, P.; Morley, A.; Moss, J.; Muenstermann, D.; Murray, P.; Nakamura, K.; Nellist, C.; Nelson, D.; Nessi, M.; Nisius, R.; Nordberg, M.; Nuiry, F.; Obermann, T.; Ockenfels, W.; Oide, H.; Oriunno, M.; Ould-Saada, F.; Padilla, C.; Pangaud, P.; Parker, S.; Pelleriti, G.; Pernegger, H.; Piacquadio, G.; Picazio, A.; Pohl, D.; Polini, A.; Pons, X.; Popule, J.; Portell Bueso, X.; Potamianos, K.; Povoli, M.; Puldon, D.; Pylypchenko, Y.; Quadt, A.; Quayle, B.; Rarbi, F.; Ragusa, F.; Rambure, T.; Richards, E.; Riegel, C.; Ristic, B.; Rivière, F.; Rizatdinova, F.; Rø hne, O.; Rossi, C.; Rossi, L.P.; Rovani, A.; Rozanov, A.; Rubinskiy, I.; Rudolph, M.S.; Rummier, A.; Ruscino, E.; Sabatini, F.; Salek, D.; Salzburger, A.; Sandaker, H.; Sannino, M.; Sanny, B.; Scanlon, T.; Schipper, J.; Schmidt, U.; Schneider, B.; Schorlemmer, A.; Schroer, N.; Schwemling, P.; Sciuccati, A.; Seidel, S.; Seiden, A.; Šícho, P.; Skubic, P.; Sloboda, M.; Smith, D.S.; Smith, M.; Sood, A.; Spencer, E.; Stramaglia, M.; Strauss, M.; Stucci, S.; Stugu, B.; Stupak, J.; Styles, N.; Su, D.; Takubo, Y.; Tassan, J.; Teng, P.; Teixeira, A.; Terzo, S.; Therry, X.; Todorov, T.; Tomášek, M.; Toms, K.; Travaglini, R.; Trischuk, W.; Troncon, C.; Troska, G.; Tsiskaridze, S.; Tsurin, I.; Tsybychev, D.; Unno, Y.; Vacavant, L.; Verlaat, B.; Vigeolas, E.; Vogt, M.; Vrba, V.; Vuillermet, R.; Wagner, W.; Walkowiak, W.; Wang, R.; Watts, S.; Weber, M.S.; Weber, M.; Weingarten, J.; Welch, S.; Wenig, S.; Wensing, M.; Wermes, N.; Wittig, T.; Wittgen, M.; Yildizkaya, T.; Yang, Y.; Yao, W.; Yi, Y.; Zaman, A.; Zaidan, R.; Zeitnitz, C.; Ziolkowski, M.; Zivkovic, V.; Zoccoli, A.; Zwalinski, L.

    2018-05-16

    During the shutdown of the CERN Large Hadron Collider in 2013-2014, an additional pixel layer was installed between the existing Pixel detector of the ATLAS experiment and a new, smaller radius beam pipe. The motivation for this new pixel layer, the Insertable B-Layer (IBL), was to maintain or improve the robustness and performance of the ATLAS tracking system, given the higher instantaneous and integrated luminosities realised following the shutdown. Because of the extreme radiation and collision rate environment, several new radiation-tolerant sensor and electronic technologies were utilised for this layer. This paper reports on the IBL construction and integration prior to its operation in the ATLAS detector.

  19. The Fiber Optic System for the Advanced Topographic Laser Altimeter System (ATLAS) Instrument

    Science.gov (United States)

    Ott, Melanie N.; Thomes, Joe; Onuma, Eleanya; Switzer, Robert; Chuska, Richard; Blair, Diana; Frese, Erich; Matyseck, Marc

    2016-01-01

    The Advanced Topographic Laser Altimeter System (ATLAS) Instrument has been in integration and testing over the past 18 months in preparation for the Ice, Cloud and Land Elevation Satellite - 2 (ICESat-2) Mission, scheduled to launch in 2017. ICESat-2 is the follow on to ICESat which launched in 2003 and operated until 2009. ATLAS will measure the elevation of ice sheets, glaciers and sea ice or the "cryosphere" (as well as terrain) to provide data for assessing the earth's global climate changes. Where ICESat's instrument, the Geo-Science Laser Altimeter (GLAS) used a single beam measured with a 70 m spot on the ground and a distance between spots of 170 m, ATLAS will measure a spot size of 10 m with a spacing of 70 cm using six beams to measure terrain height changes as small as 4 mm. The ATLAS pulsed transmission system consists of two lasers operating at 532 nm with transmitter optics for beam steering, a diffractive optical element that splits the signal into 6 separate beams, receivers for start pulse detection and a wavelength tracking system. The optical receiver telescope system consists of optics that focus all six beams into optical fibers that feed a filter system that transmits the signal via fiber assemblies to the detectors. Also included on the instrument is a system that calibrates the alignment of the transmitted pulses to the receiver optics for precise signal capture. The larger electro optical subsystems for transmission, calibration, and signal receive, stay aligned and transmitting sufficiently due to the optical fiber system that links them together. The robust design of the fiber optic system, consisting of a variety of multi fiber arrays and simplex assemblies with multiple fiber core sizes and types, will enable the system to maintain consistent critical alignments for the entire life of the mission. Some of the development approaches used to meet the challenging optical system requirements for ATLAS are discussed here.

  20. The fiber optic system for the Advanced Topographic Laser Altimeter System (ATLAS) instrument.

    Science.gov (United States)

    Ott, Melanie N; Thomes, Joe; Onuma, Eleanya; Switzer, Robert; Chuska, Richard; Blair, Diana; Frese, Erich; Matyseck, Marc

    2016-08-28

    The Advanced Topographic Laser Altimeter System (ATLAS) Instrument has been in integration and testing over the past 18 months in preparation for the Ice, Cloud and Land Elevation Satellite - 2 (ICESat-2) Mission, scheduled to launch in 2017. ICESat-2 is the follow on to ICESat which launched in 2003 and operated until 2009. ATLAS will measure the elevation of ice sheets, glaciers and sea ice or the "cryosphere" (as well as terrain) to provide data for assessing the earth's global climate changes. Where ICESat's instrument, the Geo-Science Laser Altimeter (GLAS) used a single beam measured with a 70 m spot on the ground and a distance between spots of 170 m, ATLAS will measure a spot size of 10 m with a spacing of 70 cm using six beams to measure terrain height changes as small as 4 mm.[1] The ATLAS pulsed transmission system consists of two lasers operating at 532 nm with transmitter optics for beam steering, a diffractive optical element that splits the signal into 6 separate beams, receivers for start pulse detection and a wavelength tracking system. The optical receiver telescope system consists of optics that focus all six beams into optical fibers that feed a filter system that transmits the signal via fiber assemblies to the detectors. Also included on the instrument is a system that calibrates the alignment of the transmitted pulses to the receiver optics for precise signal capture. The larger electro optical subsystems for transmission, calibration, and signal receive, stay aligned and transmitting sufficiently due to the optical fiber system that links them together. The robust design of the fiber optic system, consisting of a variety of multi fiber arrays and simplex assemblies with multiple fiber core sizes and types, will enable the system to maintain consistent critical alignments for the entire life of the mission. Some of the development approaches used to meet the challenging optical system requirements for ATLAS are discussed here.

  1. Studies of $Z\\gamma$ electroweak production in association with a high-mass di-jet system in $pp$ collisions at $\\sqrt{s}$ = 8 TeV with the ATLAS detector

    CERN Document Server

    Li, Shu; The ATLAS collaboration

    2017-01-01

    Proceeding for the poster presentation at LHCP2017, Shanghai, China on the topic of "Studies of $Z\\gamma$ electroweak production in association with a high-mass di-jet system in $pp$ collisions at $\\sqrt{s}$ = 8 TeV with the ATLAS detector" (ATL-PHYS-SLIDE-2017-266 https://cds.cern.ch/record/2265390) Deadline: 01/09/2017

  2. Control and Data Acquisition System of the ATLAS Facility

    International Nuclear Information System (INIS)

    Choi, Ki-Yong; Kwon, Tae-Soon; Cho, Seok; Park, Hyun-Sik; Baek, Won-Pil; Kim, Jung-Taek

    2007-02-01

    This report describes the control and data acquisition system of an integral effect test facility, ATLAS (Advanced Thermal-hydraulic Test Loop for Accident Simulation) facility, which recently has been constructed at KAERI (Korea Atomic Energy Research Institute). The control and data acquisition system of the ATLAS is established with the hybrid distributed control system (DCS) by RTP corp. The ARIDES system on a LINUX platform which is provided by BNF Technology Inc. is used for a control software. The IO signals consists of 1995 channels and they are processed at 10Hz. The Human-Machine-Interface (HMI) consists of 43 processing windows and they are classified according to fluid system. All control devices can be controlled by manual, auto, sequence, group, and table control methods. The monitoring system can display the real time trend or historical data of the selected IO signals on LCD monitors in a graphical form. The data logging system can be started or stopped by operator and the logging frequency can be selected among 0.5, 1, 2, 10Hz. The fluid system of the ATLAS facility consists of several systems including a primary system to auxiliary system. Each fluid system has a control similarity to the prototype plant, APR1400/OPR1000

  3. The Run-2 ATLAS Trigger System

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00222798; The ATLAS collaboration

    2016-01-01

    The ATLAS trigger successfully collected collision data during the first run of the LHC between 2009-2013 at different centre-of-mass energies between 900 GeV and 8 TeV. The trigger system consists of a hardware Level-1 and a software-based high level trigger (HLT) that reduces the event rate from the design bunch-crossing rate of 40 MHz to an average recording rate of a few hundred Hz. In Run-2, the LHC will operate at centre-of-mass energies of 13 and 14 TeV and higher luminosity, resulting in roughly five times higher trigger rates. A brief review of the ATLAS trigger system upgrades that were implemented between Run-1 and Run-2, allowing to cope with the increased trigger rates while maintaining or even improving the efficiency to select physics processes of interest, will be given. This includes changes to the Level-1 calorimeter and muon trigger systems, the introduction of a new Level-1 topological trigger module and the merging of the previously two-level HLT system into a single event filter farm. A ...

  4. Searches for direct pair production of third generation squarks with the ATLAS detector

    CERN Document Server

    Anders, John Kenneth; The ATLAS collaboration

    2015-01-01

    Naturalness arguments for weak-scale supersymmetry favour supersymmetric partners of the third generation quarks with masses not too far from those of their Standard Model counterparts. Top or bottom squarks with masses less than a few hundred GeV can also give rise to direct pair production rates at the LHC that can be observed in the data sample recorded by the ATLAS detector. The talk presents a summary of the ATLAS results from searches for direct stop and sbottom pair production, using the data collected during LHC run1, as well as sensitivity projections for the data that will be collected in 2015.

  5. Evolution of the ReadOut System of the ATLAS experiment

    CERN Document Server

    Borga, A; The ATLAS collaboration; Joos, M; Schumacher, J; Tremblet, L; Vandelli, W; Vermeulen, J; Werner, P; Wickens, F

    2014-01-01

    The ReadOut System (ROS) is a central and essential part of the ATLAS data-acquisition system. It receives and buffers event data accepted from all sub-detectors and first-level trigger subsystems. Event data are subsequently forwarded to the High-Level Trigger system and Event Builder via a GbE-based network. The ATLAS ROS will be completely renewed in view of the demanding conditions expected during LHC Run 2 and Run 3. The new ROS will consist of roughly 100 Linux-based 2U-high rack-mounted server PCs, each equipped with 2 PCIe I/O cards and four 10GbE interfaces. The FPGA-based PCIe I/O cards, developed by the ALICE collaboration, will be configured with ATLAS-specific firmware, called RobinNP. They will provide connectivity to about 2000 point-to-point optical links conveying the ATLAS event data. This dense configuration provides an excellent test bench for studying I/O efficiency and challenges in current COTS PC architectures with non-uniform memory and I/O access paths. In this paper the requirements...

  6. Evolution of the ReadOut System of the ATLAS experiment

    CERN Document Server

    Borga, A; The ATLAS collaboration; Green, B; Kugel, A; Joos, M; Panduro Vazquez, W; Schumacher, J; Teixeira-Dias, P; Tremblet, L; Vandelli, W; Vermeulen, J; Werner, P; Wickens, F

    2014-01-01

    The ReadOut System (ROS) is a central and essential part of the ATLAS DAQ system. It receives and buffers data of events accepted by the first-level trigger from all subdetectors and first-level trigger subsystems. Event data are subsequently forwarded to the High-Level Trigger system and Event Builder via a 1 GbE-based network. The ATLAS ROS is completely renewed in view of the demanding conditions expected during LHC Run 2 and Run 3, to replace obsolete technologies and space constraints require it to be compact. The new ROS will consist of roughly 100 Linux-based 2U high rack mounted server PCs, each equipped with 2 PCIe I/O cards and two four 10 GbE interfaces. The FPGA-based PCIe I/O cards, developed by the ALICE collaboration, will be configured with ATLAS-specific firmware, the so-called RobinNP firmware. They will provide the connectivity to about 2000 optical point-to-point links conveying the ATLAS event data. This dense configuration provides an excellent test bench for studying I/O efficiency and ...

  7. Multicore in production: advantages and limits of the multiprocess approach in the ATLAS experiment

    International Nuclear Information System (INIS)

    Binet, S; Calafiura, P; Lavrijsen, W; Leggett, C; Tatarkhanov, M; Tsulaia, V; Jha, M K; Lesny, D; Severini, H; Smith, D; Snyder, S; VanGemmeren, P; Washbrook, A

    2012-01-01

    The shared memory architecture of multicore CPUs provides HEP developers with the opportunity to reduce the memory footprint of their applications by sharing memory pages between the cores in a processor. ATLAS pioneered the multi-process approach to parallelize HEP applications. Using Linux fork() and the Copy On Write mechanism we implemented a simple event task farm, which allowed us to achieve sharing of almost 80% of memory pages among event worker processes for certain types of reconstruction jobs with negligible CPU overhead. By leaving the task of managing shared memory pages to the operating system, we have been able to parallelize large reconstruction and simulation applications originally written to be run in a single thread of execution with little to no change to the application code. The process of validating AthenaMP for production took ten months of concentrated effort and is expected to continue for several more months. Besides validating the software itself, an important and time-consuming aspect of running multicore applications in production was to configure the ATLAS distributed production system to handle multicore jobs. This entailed defining multicore batch queues, where the unit resource is not a core, but a whole computing node; monitoring the output of many event workers; and adapting the job definition layer to handle computing resources with different event throughputs. We will present scalability and memory usage studies, based on data gathered both on dedicated hardware and at the CERN Computer Center.

  8. gLExec Integration with the ATLAS PanDA Workload Management System

    CERN Document Server

    Karavakis, Edward; Campana, Simone; De, Kaushik; Di Girolamo, Alessandro; Litmaath, Maarten; Maeno, Tadashi; Medrano Llamas, Ramon; Nilsson, Paul; Wenaus, Torre

    2015-01-01

    ATLAS user jobs are executed on Worker Nodes (WNs) by pilots sent to sites by pilot factories. This paradigm serves to allow a high job reliability and although it has clear advantages, such as making the working environment homogeneous, the approach presents security and traceability challenges. To address these challenges, gLExec can be used to let the payloads for each user be executed under a different UNIX user id that uniquely identifies the ATLAS user. This paper describes the recent improvements and evolution of the security model within the ATLAS PanDA system, including improvements in the PanDA pilot, in the PanDA server and their integration with MyProxy, a credential caching system that entitles a person or a service to act in the name of the issuer of the credential. Finally, it presents results from ATLAS user jobs running with gLExec and describes the deployment campaign within ATLAS.

  9. Single Top quark production cross-section measurements using the ATLAS detector at the LHC

    CERN Document Server

    Jimenez Pena, Javier; The ATLAS collaboration

    2018-01-01

    Measurements of single top-quark production in proton-proton collisions are presented based on the 13 TeV and 8 TeV ATLAS datasets. In the leading order process, a W-boson is exchanged in the t-channel. The cross-section for the production of single top-quarks and single antitop-quarks, their ratio, as well as differential cross-section measurements are also reported. Measurements of the inclusive and differential cross-sections for the production of a single top quark in association with a W-boson, the second largest single top production mode are also presented. Evidence for the s-channel single top-quark production in the 8 TeV dataset is presented. Finally, the first measurement of the tZq electroweak production is presented. All measurements are compared to state-of-the art theoretical calculations. (On behalf of the ATLAS collaboration)

  10. Measurement of the particle production properties in ATLAS

    CERN Document Server

    Bruni, Alessia; The ATLAS collaboration

    2017-01-01

    Measurements of the particle production properties with the ATLAS detector A correct modelling of the underlying event in proton-proton collisions is important for the proper simulation of kinematic distributions of high-energy collisions. The ATLAS collaboration extended previous studies at 7 TeV with a leading track or jet or Z boson by a new study at 13 TeV, measuring the number and transverse-momentum sum of charged particles as a function of pseudorapidity and azimuthal angle in dependence of the reconstructed leading track. These measurements are sensitive to the underlying-event as well as the onset of hard emissions. The results are compared to predictions of several MC generators. A similar comparison between measurements and MC generator predictions will be shown for the strange meson content in topquark pair events. Studies of particle correlations in high-energy collisions can provide valuable insights into the detailed understanding of the space-time geometry of the hadronization region. The ATLA...

  11. Distributed analysis in ATLAS using GANGA

    International Nuclear Information System (INIS)

    Elmsheuser, Johannes; Brochu, Frederic; Egede, Ulrik; Reece, Will; Williams, Michael; Gaidioz, Benjamin; Maier, Andrew; Moscicki, Jakub; Vanderster, Daniel; Lee, Hurng-Chun; Pajchel, Katarina; Samset, Bjorn; Slater, Mark; Soroko, Alexander; Cowan, Greig

    2010-01-01

    Distributed data analysis using Grid resources is one of the fundamental applications in high energy physics to be addressed and realized before the start of LHC data taking. The needs to manage the resources are very high. In every experiment up to a thousand physicists will be submitting analysis jobs to the Grid. Appropriate user interfaces and helper applications have to be made available to assure that all users can use the Grid without expertise in Grid technology. These tools enlarge the number of Grid users from a few production administrators to potentially all participating physicists. The GANGA job management system (http://cern.ch/ganga), developed as a common project between the ATLAS and LHCb experiments, provides and integrates these kind of tools. GANGA provides a simple and consistent way of preparing, organizing and executing analysis tasks within the experiment analysis framework, implemented through a plug-in system. It allows trivial switching between running test jobs on a local batch system and running large-scale analyzes on the Grid, hiding Grid technicalities. We will be reporting on the plug-ins and our experiences of distributed data analysis using GANGA within the ATLAS experiment. Support for all Grids presently used by ATLAS, namely the LCG/EGEE, NDGF/NorduGrid, and OSG/PanDA is provided. The integration and interaction with the ATLAS data management system DQ2 into GANGA is a key functionality. An intelligent job brokering is set up by using the job splitting mechanism together with data-set and file location knowledge. The brokering is aided by an automated system that regularly processes test analysis jobs at all ATLAS DQ2 supported sites. Large numbers of analysis jobs can be sent to the locations of data following the ATLAS computing model. GANGA supports amongst other things tasks of user analysis with reconstructed data and small scale production of Monte Carlo data.

  12. Commissioning the ATLAS Level-1 Central Trigger System

    CERN Document Server

    Sherman, Daniel

    2010-01-01

    The ATLAS Level-1 central trigger is a critical part of ATLAS operation. It receives the 40 MHz bunch clock from the LHC and distributes it to all sub-detectors. It initiates their read-out by forming the Level-1 Accept decision, which is based on information from the calorimeter and muon trigger processors and a variety of additional trigger inputs from detectors in the forward region. It also provides trigger summary information to the data acquisition system and the Level-2 trigger system. In this paper, we present the completion of the installed central trigger system, its performance during cosmic-ray data taking and the experience gained with triggering on the first LHC beams.

  13. ATLAS Maintenance and Operation management system

    CERN Document Server

    Copy, B

    2007-01-01

    The maintenance and operation of the ATLAS detector will involve thousands of contributors from 170 physics institutes. Planning and coordinating the action of ATLAS members, ensuring their expertise is properly leveraged and that no parts of the detector are understaffed or overstaffed will be a challenging task. The ATLAS Maintenance and Operation application (referred to as Operation Task Planner inside the ATLAS experiment) offers a fluent web based interface that combines the flexibility and comfort of a desktop application, intuitive data visualization and navigation techniques, with a lightweight service oriented architecture. We will review the application, its usage within the ATLAS experiment, its underlying design and implementation.

  14. Atlas Basemaps in Web 2.0 Epoch

    Science.gov (United States)

    Chabaniuk, V.; Dyshlyk, O.

    2016-06-01

    The authors have analyzed their experience of the production of various Electronic Atlases (EA) and Atlas Information Systems (AtIS) of so-called "classical type". These EA/AtIS have been implemented in the past decade in the Web 1.0 architecture (e.g., National Atlas of Ukraine, Atlas of radioactive contamination of Ukraine, and others). One of the main distinguishing features of these atlases was their static nature - the end user could not change the content of EA/AtIS. Base maps are very important element of any EA/AtIS. In classical type EA/AtIS they were static datasets, which consisted of two parts: the topographic data of a fixed scale and data of the administrative-territorial division of Ukraine. It is important to note that the technique of topographic data production was based on the use of direct channels of topographic entity observation (such as aerial photography) for the selected scale. Changes in the information technology of the past half-decade are characterized by the advent of the "Web 2.0 epoch". Due to this, in cartography appeared such phenomena as, for example, "neo-cartography" and various mapping platforms like OpenStreetMap. These changes have forced developers of EA/AtIS to use new atlas basemaps. Our approach is described in the article. The phenomenon of neo-cartography and/or Web 2.0 cartography are analysed by authors using previously developed Conceptual framework of EA/AtIS. This framework logically explains the cartographic phenomena relations of three formations: Web 1.0, Web 1.0x1.0 and Web 2.0. Atlas basemaps of the Web 2.0 epoch are integrated information systems. We use several ways to integrate separate atlas basemaps into the information system - by building: weak integrated information system, structured system and meta-system. This integrated information system consists of several basemaps and falls under the definition of "big data". In real projects it is already used the basemaps of three strata: Conceptual

  15. ATLAS BASEMAPS IN WEB 2.0 EPOCH

    Directory of Open Access Journals (Sweden)

    V. Chabaniuk

    2016-06-01

    Full Text Available The authors have analyzed their experience of the production of various Electronic Atlases (EA and Atlas Information Systems (AtIS of so-called "classical type". These EA/AtIS have been implemented in the past decade in the Web 1.0 architecture (e.g., National Atlas of Ukraine, Atlas of radioactive contamination of Ukraine, and others. One of the main distinguishing features of these atlases was their static nature - the end user could not change the content of EA/AtIS. Base maps are very important element of any EA/AtIS. In classical type EA/AtIS they were static datasets, which consisted of two parts: the topographic data of a fixed scale and data of the administrative-territorial division of Ukraine. It is important to note that the technique of topographic data production was based on the use of direct channels of topographic entity observation (such as aerial photography for the selected scale. Changes in the information technology of the past half-decade are characterized by the advent of the “Web 2.0 epoch”. Due to this, in cartography appeared such phenomena as, for example, "neo-cartography" and various mapping platforms like OpenStreetMap. These changes have forced developers of EA/AtIS to use new atlas basemaps. Our approach is described in the article. The phenomenon of neo-cartography and/or Web 2.0 cartography are analysed by authors using previously developed Conceptual framework of EA/AtIS. This framework logically explains the cartographic phenomena relations of three formations: Web 1.0, Web 1.0x1.0 and Web 2.0. Atlas basemaps of the Web 2.0 epoch are integrated information systems. We use several ways to integrate separate atlas basemaps into the information system – by building: weak integrated information system, structured system and meta-system. This integrated information system consists of several basemaps and falls under the definition of "big data". In real projects it is already used the basemaps of three strata

  16. Evolution of User Analysis on the Grid in ATLAS

    CERN Document Server

    Legger, Federica; The ATLAS collaboration

    2016-01-01

    More than one thousand physicists analyse data collected by the ATLAS experiment at the Large Hadron Collider (LHC) at CERN through 150 computing facilities around the world. Efficient distributed analysis requires optimal resource usage and the interplay of several factors: robust grid and software infrastructures, system capability to adapt to different workloads. The continuous automatic validation of grid sites and the user support provided by a dedicated team of expert shifters have been proven to provide a solid distributed analysis system for ATLAS users. Based on the experience from the first run of the LHC, substantial improvements to the ATLAS computing system have been made to optimize both production and analysis workflows. These include the re-design of the production and data management systems, a new analysis data format and event model, and the development of common reduction and analysis frameworks. The impact of such changes on the distributed analysis system is evaluated. More than 100 mill...

  17. Searches for electroweak production of supersymmetric gauginos and sleptons with the ATLAS detector

    CERN Document Server

    Carra, Sonia; The ATLAS collaboration

    2018-01-01

    Supersymmetry is one of the most motivated Standard Model extensions. Despite the meticulous search during the LHC Run I, there is no evidence supporting this theory. Starting from 2015, LHC is performing a second data taking run with a higher center of mass energy (13 TeV), providing a great occasion for the search of beyond the Standard Model physics. An important sector is the direct production of supersymmetric electroweak particles, such as sleptons and charginos. Electroweak production cross section is lower compared to strong production, but searches performed by the ATLAS and CMS experiments during LHC Run 2 excluded squark and gluinos with masses up to 2 TeV, making electroweak production an increasingly promising probe for SUSY signals at the LHC. Results obtained with the 2015-2016 ATLAS detector data will be presented. Direct production of electroweak particles like sleptons, charginos and neutralinos, with different signatures, will be considered. A good sensitivity is obtained in the signal regi...

  18. Advances in service and operations for ATLAS data management

    International Nuclear Information System (INIS)

    Stewart, Graeme A; Garonne, Vincent; Lassnig, Mario; Molfetas, Angelos; Barisits, Martin; Calvet, Ivan; Beermann, Thomas; Megino, Fernando Barreiro; Campana, Simone; Zhang, Donal; Tykhonov, Andrii; Serfon, Cedric; Oleynik, Danila; Petrosyan, Artem

    2012-01-01

    ATLAS has recorded almost 5PB of RAW data since the LHC started running at the end of 2009. Many more derived data products and complimentary simulation data have also been produced by the collaboration and, in total, 70PB is currently stored in the Worldwide LHC Computing Grid by ATLAS. All of this data is managed by the ATLAS Distributed Data Management system, called Don Quixote 2 (DQ2). DQ2 has evolved rapidly to help ATLAS Computing operations manage these large quantities of data across the many grid sites at which ATLAS runs and to help ATLAS physicists get access to this data. In this paper we describe new and improved DQ2 services: popularity; space monitoring and accounting; exclusion service; cleaning agents; deletion agents. We describe the experience of data management operation in ATLAS computing, showing how these services enable management of petabyte scale computing operations. We illustrate the coupling of data management services to other parts of the ATLAS computing infrastructure, in particular showing how feedback from the distributed analysis system in ATLAS has enabled dynamic placement of the most popular data, helping users and groups to analyse the increasing data volumes on the grid.

  19. Advances in service and operations for ATLAS data management

    Science.gov (United States)

    Stewart, Graeme A.; Garonne, Vincent; Lassnig, Mario; Molfetas, Angelos; Barisits, Martin; Zhang, Donal; Calvet, Ivan; Beermann, Thomas; Barreiro Megino, Fernando; Tykhonov, Andrii; Campana, Simone; Serfon, Cedric; Oleynik, Danila; Petrosyan, Artem; ATLAS Collaboration

    2012-06-01

    ATLAS has recorded almost 5PB of RAW data since the LHC started running at the end of 2009. Many more derived data products and complimentary simulation data have also been produced by the collaboration and, in total, 70PB is currently stored in the Worldwide LHC Computing Grid by ATLAS. All of this data is managed by the ATLAS Distributed Data Management system, called Don Quixote 2 (DQ2). DQ2 has evolved rapidly to help ATLAS Computing operations manage these large quantities of data across the many grid sites at which ATLAS runs and to help ATLAS physicists get access to this data. In this paper we describe new and improved DQ2 services: popularity; space monitoring and accounting; exclusion service; cleaning agents; deletion agents. We describe the experience of data management operation in ATLAS computing, showing how these services enable management of petabyte scale computing operations. We illustrate the coupling of data management services to other parts of the ATLAS computing infrastructure, in particular showing how feedback from the distributed analysis system in ATLAS has enabled dynamic placement of the most popular data, helping users and groups to analyse the increasing data volumes on the grid.

  20. Measurement of quarkonia production in heavy-ion collisions with the ATLAS detector

    CERN Document Server

    Kremer, Jakub Andrzej; The ATLAS collaboration

    2017-01-01

    The suppression of heavy quarkonia states in heavy-ion collisions is a phenomenon understood as a consequence of QGP formation in the hot, dense system formed in nucleus-nucleus collisions at the LHC. In addition to hot matter effects in heavy-ion collisions, cold nuclear effects may also affect quarkonia production. Therefore, a full assessment requires detailed studies on the effects present in both A+A and $\\textit{p}$+A collisions. Based on $\\textit{p}$+Pb data collected in 2013 and $\\textit{pp}$ and Pb+Pb data collected in 2015 at the LHC, the ATLAS experiment has studied prompt and non-prompt $J/\\psi$ and $\\psi\\left(2\\mathrm{S}\\right)$ productions as well as $\\Upsilon\\left(n\\mathrm{S}\\right)$ production via the di-muon decay final states. The results of the various measurements are discussed.

  1. The ATLAS distributed analysis system

    OpenAIRE

    Legger, F.

    2014-01-01

    In the LHC operations era, analysis of the multi-petabyte ATLAS data sample by globally distributed physicists is a challenging task. To attain the required scale the ATLAS Computing Model was designed around the concept of grid computing, realized in the Worldwide LHC Computing Grid (WLCG), the largest distributed computational resource existing in the sciences. The ATLAS experiment currently stores over 140 PB of data and runs about 140,000 concurrent jobs continuously at WLCG sites. During...

  2. AGIS: Integration of new technologies used in ATLAS Distributed Computing

    Science.gov (United States)

    Anisenkov, Alexey; Di Girolamo, Alessandro; Alandes Pradillo, Maria

    2017-10-01

    The variety of the ATLAS Distributed Computing infrastructure requires a central information system to define the topology of computing resources and to store different parameters and configuration data which are needed by various ATLAS software components. The ATLAS Grid Information System (AGIS) is the system designed to integrate configuration and status information about resources, services and topology of the computing infrastructure used by ATLAS Distributed Computing applications and services. Being an intermediate middleware system between clients and external information sources (like central BDII, GOCDB, MyOSG), AGIS defines the relations between experiment specific used resources and physical distributed computing capabilities. Being in production during LHC Runl AGIS became the central information system for Distributed Computing in ATLAS and it is continuously evolving to fulfil new user requests, enable enhanced operations and follow the extension of the ATLAS Computing model. The ATLAS Computing model and data structures used by Distributed Computing applications and services are continuously evolving and trend to fit newer requirements from ADC community. In this note, we describe the evolution and the recent developments of AGIS functionalities, related to integration of new technologies recently become widely used in ATLAS Computing, like flexible computing utilization of opportunistic Cloud and HPC resources, ObjectStore services integration for Distributed Data Management (Rucio) and ATLAS workload management (PanDA) systems, unified storage protocols declaration required for PandDA Pilot site movers and others. The improvements of information model and general updates are also shown, in particular we explain how other collaborations outside ATLAS could benefit the system as a computing resources information catalogue. AGIS is evolving towards a common information system, not coupled to a specific experiment.

  3. ATLAS Potential for Beauty Physics Measurements

    International Nuclear Information System (INIS)

    Smizanska, M.

    2001-01-01

    The main focus of ATLAS b physics has traditionally been on the standard model. In the last few years also the aspects of new physics in B-decays has been addressed. Another new field of studies started recently is a beauty production. We give an overview of the older as well as more recent results. After an introduction outlining selected trigger and detector performance characteristics, we explain methods and goals of CP violation measurements in decay channels of B d 0 meson, physics of B s 0 system and of rare decays. Finally, the ATLAS program for beauty production measurements is presented. (author)

  4. Deploying the ATLAS Metadata Interface (AMI) on the cloud with Jenkins

    Science.gov (United States)

    Lambert, F.; Odier, J.; Fulachier, J.; ATLAS Collaboration

    2017-10-01

    The ATLAS Metadata Interface (AMI) is a mature application of more than 15 years of existence. Mainly used by the ATLAS experiment at CERN, it consists of a very generic tool ecosystem for metadata aggregation and cataloguing. AMI is used by the ATLAS production system, therefore the service must guarantee a high level of availability. We describe our monitoring and administration systems, and the Jenkins-based strategy used to dynamically test and deploy cloud OpenStack nodes on demand.

  5. Deploying the ATLAS Metadata Interface (AMI) on the cloud with Jenkins.

    CERN Document Server

    AUTHOR|(SzGeCERN)637120; The ATLAS collaboration; Odier, Jerome; Fulachier, Jerome

    2017-01-01

    The ATLAS Metadata Interface (AMI) is a mature application of more than 15 years of existence. Mainly used by the ATLAS experiment at CERN, it consists of a very generic tool ecosystem for metadata aggregation and cataloguing. AMI is used by the ATLAS production system, therefore the service must guarantee a high level of availability. We describe our monitoring and administration systems, and the Jenkins-based strategy used to dynamically test and deploy cloud OpenStack nodes on demand.

  6. ATLAS Tile calorimeter calibration and monitoring systems

    CERN Document Server

    Marjanovic, Marija; The ATLAS collaboration

    2018-01-01

    The ATLAS Tile Calorimeter (TileCal) is the central section of the hadronic calorimeter of the ATLAS experiment. This sampling calorimeter uses steel plates as absorber and scintillating tiles as active medium. The light produced by the passage of charged particles is transmitted by wavelength shifting fibers to photo-multiplier tubes (PMTs), located in the outer part of the calorimeter. The readout is segmented into about 5000 cells, each one being read out by two PMTs in parallel. To calibrate and monitor the stability and performance of the full readout chain during the data taking, a set of calibration sub-systems is used. The TileCal calibration system comprises Cesium radioactive sources, laser, charge injection elements, and an integrator based readout system. Combined information from all systems allows to monitor and to equalize the calorimeter response at each stage of the signal evolution, from scintillation light to digitization. Calibration runs are monitored from a data quality perspective and u...

  7. ATLAS Tile calorimeter calibration and monitoring systems

    CERN Document Server

    Cortes-Gonzalez, Arely; The ATLAS collaboration

    2017-01-01

    The ATLAS Tile Calorimeter is the central section of the hadronic calorimeter of the ATLAS experiment and provides important information for reconstruction of hadrons, jets, hadronic decays of tau leptons and missing transverse energy. This sampling calorimeter uses steel plates as absorber and scintillating tiles as active medium. The light produced by the passage of charged particles is transmitted by wavelength shifting fibres to photomultiplier tubes, located in the outer part of the calorimeter. The readout is segmented into about 5000 cells (longitudinally and transversally), each of them being read out by two photomultiplier in parallel. To calibrate and monitor the stability and performance of each part of the readout chain during the data taking, a set of calibration systems is used. The calibration system comprises Cesium radioactive sources, laser, charge injection elements and an integrator based readout system. Combined information from all systems allows to monitor and equalise the calorimeter r...

  8. Production facility for ATLAS new small wheel drift panels at JGU Mainz

    Energy Technology Data Exchange (ETDEWEB)

    Duedder, Andreas; Lin, Tai-Hua; Schott, Matthias [Johannes Gutenberg-Universitaet Mainz (Germany)

    2016-07-01

    The ATLAS Phase-I Upgrade in 2018 includes the replacement of the ATLAS Muon Small Wheel by the so-called New Small Wheel (NSW). Large-scale Micromegas detectors will serve as tracking detectors in the NSW. Parts of these detectors will be constructed at the Johannes Gutenberg University Mainz (JGU). In order to fulfill the requirements of the envisioned detector performance, a high precision detector construction is crucial. Especially the surface planarity of the produced detector panels has to better than 30 μm over an area of 2 m{sup 2}. Methods for the quality control of the raw material and the constructed parts have been developed and implemented. This talk gives an overview of the production facility at JGU Mainz which is used during the mass production of NSW components in coming years.

  9. Evolution of the ATLAS PanDA workload management system for exascale computational science

    International Nuclear Information System (INIS)

    Maeno, T; Klimentov, A; Panitkin, S; Schovancova, J; Wenaus, T; Yu, D; De, K; Nilsson, P; Oleynik, D; Petrosyan, A; Vaniachine, A

    2014-01-01

    An important foundation underlying the impressive success of data processing and analysis in the ATLAS experiment [1] at the LHC [2] is the Production and Distributed Analysis (PanDA) workload management system [3]. PanDA was designed specifically for ATLAS and proved to be highly successful in meeting all the distributed computing needs of the experiment. However, the core design of PanDA is not experiment specific. The PanDA workload management system is capable of meeting the needs of other data intensive scientific applications. Alpha-Magnetic Spectrometer [4], an astro-particle experiment on the International Space Station, and the Compact Muon Solenoid [5], an LHC experiment, have successfully evaluated PanDA and are pursuing its adoption. In this paper, a description of the new program of work to develop a generic version of PanDA will be given, as well as the progress in extending PanDA's capabilities to support supercomputers and clouds and to leverage intelligent networking. PanDA has demonstrated at a very large scale the value of automated dynamic brokering of diverse workloads across distributed computing resources. The next generation of PanDA will allow other data-intensive sciences and a wider exascale community employing a variety of computing platforms to benefit from ATLAS' experience and proven tools.

  10. Single top quark production in ATLAS at the LHC

    Directory of Open Access Journals (Sweden)

    Stillings Jan A.

    2013-11-01

    Full Text Available Measurements of the single top quark production cross-sections in proton collisions with the ATLAS detector at the Large Hadron Collider are presented. The single top-quark production in the t and Wt-channels and the determination of the CKM matrix element |Vtb| are discussed. A separate measurement of the top and antitop quark cross-sections and their ratio is shown. These measurements are sensitive to the parton distribution function in the proton. In addition, limits on exotic production in single top quark processes are outlined. This also includes the search for flavour-changing neutral currents and for additional W’ bosons in the s-channel.

  11. ATLAS measurements of vector boson production

    CERN Document Server

    Vittori, Camilla; The ATLAS collaboration

    2016-01-01

    Measurements of the Drell­Yan production of W and Z/gamma* bosons at the LHC provide a benchmark of our understanding of perturbative QCD and probe the proton structure in a unique way. The ATLAS collaboration has performed new high precision measurements at center­of­mass energies of 8 and 13 TeV. In the 8 TeV data, we present recent measurements in the di­lepton mass range up to the TeV scale, double­differentially in dilepton mass and rapidity(­separation). The measurements are compared to state­of­the­art calculations at NNLO in QCD and constrain the photon content of the proton. First precise inclusive measurements of W and Z production at 13 TeV are presented. W/Z and W charge ratios profit from a cancellation of experimental uncertainties. The angular distributions of the Drell­Yan lepton pairs around the Z­boson mass peak probe the underlying QCD dynamic of the Z­boson production mechanisms. We present a measurement of the complete set of angular coefficients describing these distributions...

  12. The Pig PeptideAtlas

    DEFF Research Database (Denmark)

    Hesselager, Marianne Overgaard; Codrea, Marius; Sun, Zhi

    2016-01-01

    Biological research of Sus scrofa, the domestic pig, is of immediate relevance for food production sciences, and for developing pig as a model organism for human biomedical research. Publicly available data repositories play a fundamental role for all biological sciences, and protein data...... repositories are in particular essential for the successful development of new proteomic methods. Cumulative proteome data repositories, including the PeptideAtlas, provide the means for targeted proteomics, system-wide observations, and cross-species observational studies, but pigs have so far been...... underrepresented in existing repositories. We here present a significantly improved build of the Pig PeptideAtlas, which includes pig proteome data from 25 tissues and three body fluid types mapped to 7139 canonical proteins. The content of the Pig PeptideAtlas reflects actively ongoing research within...

  13. The Run-2 ATLAS Trigger System

    CERN Document Server

    Ruiz-Martinez, Aranzazu; The ATLAS collaboration

    2016-01-01

    The ATLAS trigger has been successfully collecting collision data during the first run of the LHC between 2009-2013 at a centre-of-mass energy between 900 GeV and 8 TeV. The trigger system consists of a hardware Level-1 (L1) and a software based high-level trigger (HLT) that reduces the event rate from the design bunch-crossing rate of 40 MHz to an average recording rate of a few hundred Hz. In Run-2, the LHC will operate at centre-of-mass energies of 13 and 14 TeV resulting in roughly five times higher trigger rates. We will briefly review the ATLAS trigger system upgrades that were implemented during the shutdown, allowing us to cope with the increased trigger rates while maintaining or even improving our efficiency to select relevant physics processes. This includes changes to the L1 calorimeter and muon trigger systems, the introduction of a new L1 topological trigger module and the merging of the previously two-level HLT system into a single event filter farm. At hand of a few examples, we will show the ...

  14. Associated production of $J/\\psi$ pairs with the ATLAS detector

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00225428; The ATLAS collaboration

    2017-01-01

    A recent measurement of prompt $J/\\psi$ pair production is presented, using a sample of 11.4 fb$^{-1}$ of proton-proton collision data collected at $\\sqrt{s} = 8$ TeV in the ATLAS detector. The differential cross-section is measured as a function of kinematic distributions for the lower-$p_T J/\\psi$ meson and for the di-$J/\\psi$ system. A data-driven approach is used to extract the fraction of prompt $J/\\psi$ pair events due to double parton scattering and an effective cross-section of double parton scattering is measured to be $\\sigma_{eff} = 6.3 \\pm 1.6 (\\text{stat}) \\pm 1.0 (\\text{syst}) \\pm 0.1 (\\text{BF}) \\pm 0.1 (\\text{lumi})$ mb.

  15. PD2P: PanDA Dynamic Data Placement for ATLAS

    International Nuclear Information System (INIS)

    Maeno, T; Panitkin, S; De, K

    2012-01-01

    The PanDA (Production and Distributed Analysis) system plays a key role in the ATLAS distributed computing infrastructure. PanDA is the ATLAS workload management system for processing all Monte-Carlo (MC) simulation and data reprocessing jobs in addition to user and group analysis jobs. The PanDA Dynamic Data Placement (PD2P) system has been developed to cope with difficulties of data placement for ATLAS. We will describe the design of the new system, its performance during the past year of data taking, dramatic improvements it has brought about in the efficient use of storage and processing resources, and plans for the future.

  16. ATLAS Open Data project

    CERN Document Server

    The ATLAS collaboration

    2018-01-01

    The current ATLAS model of Open Access to recorded and simulated data offers the opportunity to access datasets with a focus on education, training and outreach. This mandate supports the creation of platforms, projects, software, and educational products used all over the planet. We describe the overall status of ATLAS Open Data (http://opendata.atlas.cern) activities, from core ATLAS activities and releases to individual and group efforts, as well as educational programs, and final web or software-based (and hard-copy) products that have been produced or are under development. The relatively large number and heterogeneous use cases currently documented is driving an upcoming release of more data and resources for the ATLAS Community and anyone interested to explore the world of experimental particle physics and the computer sciences through data analysis.

  17. Measurement of quarkonia production in heavy-ion collisions with the ATLAS detector

    CERN Document Server

    Tapia Araya, Sebastian; The ATLAS collaboration

    2017-01-01

    The suppression of heavy quarkonia states in heavy-ion collisions is a phenomenon understood as a consequence of QGP formation in the hot, dense system produced in interactions of heavy ions at high energy. In addition to hot matter effects, cold nuclear effects can play an important role in quarkonia production. Therefore, a full assessment of different physics scenarios requires detailed studies on the effects present in Pb+Pb and p+Pb collisions in comparison to the pp collisions. Results of the studies based on p+Pb data collected in 2013 and pp and Pb+Pb data collected in 2015 at the LHC by the ATLAS experiment at the centre of mass energy of 5.02 TeV allowed studying prompt and non-prompt J/ψ and ψ(2S) productions as well as Υ(nS) (n = 1, 2, 3) production via the di-muon decay final states. The results of the measurements presented as a function rapidity and transverse momentum as well as the ratios between different species and systems are presented and discussed in the talk.

  18. Single top-quark production cross section measurements using the ATLAS detector at the LHC

    CERN Document Server

    Rieck, Patrick; The ATLAS collaboration

    2016-01-01

    Measurements of single top­quark production in proton proton collisions are presented. The measurements include the first such measurements from the 13 TeV ATLAS dataset. In the leading order process, a W boson is exchanged in the t­channel. The single top­quark and anti­top total production cross sections, their ratio, as well as a measurement of the inclusive production cross section is presented. At 8 TeV, differential cross­section measurements of the t­channel process are also presented, these measurements include limits on anomalous contributions to the Wtb vertex. A measurement of the production cross section of a single top quark in association with a W boson, the second largest single­top production mode, is also presented. Finally, evidence for single­top production in the 8 TeV ATLAS dataset is presented. All measurements are compared to state­of­ the­art theoretical calculations.

  19. Evolution of the ATLAS PanDA Workload Management System for Exascale Computational Science

    CERN Document Server

    Maeno, T; The ATLAS collaboration; Klimentov, A; Nilsson, P; Oleynik, D; Panitkin, S; Petrosyan, A; Schovancova, J; Vaniachine, A; Wenaus, T; Yu, D

    2013-01-01

    An important foundation underlying the impressive success of data processing and analysis in the ATLAS experiment [1] at the LHC [2] is the Production and Distributed Analysis (PanDA) workload management system [3]. PanDA was designed specifically for ATLAS and proved to be highly successful in meeting all the distributed computing needs of the experiment. However, the core design of PanDA is not experiment specific. The PanDA workload management system is capable of meeting the needs of other data intensive scientific applications. Alpha-Magnetic Spectrometer [4], an astro-particle experiment on the International Space Station, and the Compact Muon Solenoid [5], an LHC experiment, have successfully evaluated PanDA and are pursuing its adoption. In this paper, a description of the new program of work to develop a generic version of PanDA will be given, as well as the progress in extending PanDA's capabilities to support supercomputers and clouds and to leverage intelligent networking. PanDA has demonstrated a...

  20. Evolution of the ATLAS PanDA Workload Management System for Exascale Computational Science

    CERN Document Server

    Maeno, T; The ATLAS collaboration; Klimentov, A; Nilsson, P; Oleynik, D; Panitkin, S; Petrosyan, A; Schovancova, J; Vaniachine, A; Wenaus, T; Yu, D

    2014-01-01

    An important foundation underlying the impressive success of data processing and analysis in the ATLAS experiment [1] at the LHC [2] is the Production and Distributed Analysis (PanDA) workload management system [3]. PanDA was designed specifically for ATLAS and proved to be highly successful in meeting all the distributed computing needs of the experiment. However, the core design of PanDA is not experiment specific. The PanDA workload management system is capable of meeting the needs of other data intensive scientific applications. Alpha-Magnetic Spectrometer [4], an astro-particle experiment on the International Space Station, and the Compact Muon Solenoid [5], an LHC experiment, have successfully evaluated PanDA and are pursuing its adoption. In this paper, a description of the new program of work to develop a generic version of PanDA will be given, as well as the progress in extending PanDA's capabilities to support supercomputers and clouds and to leverage intelligent networking. PanDA has demonstrated a...

  1. RiceAtlas, a spatial database of global rice calendars and production

    NARCIS (Netherlands)

    Laborte, Alice G.; Gutierrez, Mary Anne; Balanza, Jane Girly; Saito, Kazuki; Zwart, Sander; Boschetti, Mirco; Murty, M. V.R.; Villano, Lorena; Aunario, Jorrel Khalil; Reinke, Russell; Koo, Jawoo; Hijmans, Robert J.; Nelson, Andrew

    2017-01-01

    Knowing where, when, and how much rice is planted and harvested is crucial information for understanding the effects of policy, trade, and global and technological change on food security. We developed RiceAtlas, a spatial database on the seasonal distribution of the world's rice production. It

  2. Searching for Higgs boson production in association with a top quark pair with the ATLAS detector

    CERN Document Server

    Ospanov, Rustem; The ATLAS collaboration

    2018-01-01

    This contribution will report on most recent searches for Higgs boson production in association with a top quark pair (ttH) performed by the ATLAS experiment at the LHC. The ttH process is interesting because it can measure directly the Higgs boson coupling to the top quark, which otherwise is known only via virtual top quark loop contributions to the inclusive Higgs boson production at the LHC. The ttH production process has not been observed yet and it is a subject of extensive experimental studies by the ATLAS and CMS collaborations. This presentation will discuss most recent ATLAS searches for the ttH production, focusing on Higgs boson decays to a pair of W or Z bosons, or to a pair of tau leptons. These decay modes result in final states with multiple leptons and jets, thus allowing significant suppression of background processes. This presentation will discuss new experimental techniques for suppressing these background processes and present latest results in different multi-lepton channels. The presen...

  3. ATLAS Forward Proton Detector

    CERN Document Server

    Grieco, Chiara; The ATLAS collaboration

    2018-01-01

    The aim of the ATLAS Forward Proton (AFP) detector system is the measurement of protons scattered diffractively or electromagnetically at very small angles. The full two-arm setup was installed during the 2016/2017 EYETS. This allows measurements of processes with two forward protons: central diffraction, exclusive production, and two-photon processes. In 2017, AFP participated in the ATLAS high-luminosity data taking on the day-by-day basis. In addition, several special runs with reduced luminosity were taken. The poster will present the AFP detectors and the lessons learned from the last year operation and some performance from 2016 and 2017.

  4. ATLAS Rewards Russian Supplier for Scintillating Tile Production

    CERN Multimedia

    2001-01-01

    At a ceremony held at CERN on 30 July, the ATLAS collaboration awarded Russian firm SIA Luch from Podolsk in the Moscow region an ATLAS Suppliers Award. This follows delivery by the company of the final batch of scintillating tiles for the collaboration's Tile Calorimeter some six months ahead of schedule.   Representatives of Russian firm Luch Podolsk received the ATLAS Suppliers Award in the collaboration's Tile Calorimeter instrumentation plant at CERN on 30 July. In front of one Tile Calorimeter module instrumented by scintillating tiles are (left to right) IHEP physicists Evgueni Startchenko and Andrei Karioukhine, Luch Podolsk representatives Igor Karetnikov and Yuri Zaitsev, Tile Calorimeter Project Leader Rupert Leitner, ATLAS spokesperson Peter Jenni, and CERN Tile Calorimeter group leader Ana Henriques-Correia. Scintillating tiles form the active part of the ATLAS hadronic Tile Calorimeter, which will measure the energy and direction of particles produced in LHC collisions. They are emb...

  5. The Run-2 ATLAS Trigger System: Design, Performance and Plan

    CERN Document Server

    zur Nedden, Martin; The ATLAS collaboration

    2016-01-01

    In high-energy physics experiments, online selection is crucial to select interesting collisions from the large data volume. The ATLAS experiment at the Large Hadron Collider (LHC) utilizes the trigger system that consists of a hardware Level-1 (L1) and a software based high-level trigger (HLT), reducing the event rate from the design bunch-crossing rate of 40 MHz to an average recording rate of about 1000 Hz. The ATLAS trigger has been successfully collecting collision data during the first run of the LHC (Run-1) between 2009-2013 at a centre-of-mass energy between 900 GeV and 8 TeV. In the second run of LHC (Run-2) starting from 2015, the LHC operates at centre-of-mass energy of 13 TeV and provides a higher luminosity of collisions. Also, the number of collisions occurring in a same bunch crossing increases. The ATLAS trigger system has to cope with these challenges, while maintaining or even improving the efficiency to select relevant physics processes. In this talk, first we will review the ATLAS trigger ...

  6. An Embedded Real-Time System on ATLAS ROBIN

    OpenAIRE

    Yu, Maoyuan

    2012-01-01

    ATLAS is the largest particle detector at the Large Hadron Collider for high energy physics experiments that produces over 40 TB/s event data. The ATLAS Readout Buffer INput(ROBIN) subsystem is an essential device to buffer and reduce the data, which has a IBM PowerPC core for the control functionalities. This dissertation addresses the software design of an embedded real-time system centering on the PowerPC micro-controller, as the management core of the ROBIN. A page-based solution is pr...

  7. Simulation of Top Quark Production for the ATLAS Experiment

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00352736; The ATLAS collaboration

    2016-01-01

    The Monte Carlo setups used by ATLAS to model the $t\\bar{t}$ and single-top production in 13 TeV pp collisions are described. The performance of different event generators is assessed by comparing measurements at 7 TeV, 8 TeV and 13 TeV to predictions from simulated data. The evaluation of systematic uncertainties and the dependence of generator predictions on the tuning parameters are also discussed.

  8. The ATLAS PanDA Pilot in Operation

    International Nuclear Information System (INIS)

    Nilsson, P; De, K; Stradling, A; Caballero, J; Maeno, T; Wenaus, T

    2011-01-01

    The Production and Distributed Analysis system (PanDA) was designed to meet ATLAS requirements for a data-driven workload management system capable of operating at LHC data processing scale. Submitted jobs are executed on worker nodes by pilot jobs sent to the grid sites by pilot factories. This paper provides an overview of the PanDA pilot system and presents major features added in light of recent operational experience, including multi-job processing, advanced job recovery for jobs with output storage failures, gLExec based identity switching from the generic pilot to the actual user, and other security measures. The PanDA system serves all ATLAS distributed processing and is the primary system for distributed analysis; it is currently used at over 100 sites worldwide. We analyze the performance of the pilot system in processing real LHC data on the OSG, EGI and Nordugrid infrastructures used by ATLAS, and describe plans for its evolution.

  9. The ATLAS software installation system for LCG/EGEE

    Energy Technology Data Exchange (ETDEWEB)

    Salvo, A D [Istituto Nazionale di Fisica Nucleare, sez. Roma 1 (Italy); Barchiesi, A [Universita di Roma I ' La Sapienza' (Italy); Gnanvo, K [Queen Mary and Westfield College (United Kingdom); Gwilliam, C [University of Liverpool (United Kingdom); Kennedy, J; Krobath, G [Ludwig-Maximilians-Universitaet Muenchen (Germany); Olszewski, A [Henryk Niewodniczanski Institute of Nuclear Physics, Polish Academy of Sciences (Poland); Rybkine, G [Royal Holloway College (United Kingdom)

    2008-07-15

    The huge amount of resources available in the Grids, and the necessity to have the most up-to-date experimental software deployed in all the sites within a few hours, have driven the need for an automatic installation system for the LHC experiments. In this work we describe the ATLAS system for the experiment software installation in LCG/EGEE, based on the Light Job Submission Framework for Installation (LJSFi), an independent job submission framework for generic submission and job tracking in EGEE. LJSFi is able to automatically discover, check, install, test and tag the full set of resources made available in LCG/EGEE to the ATLAS Virtual Organization in a few hours, depending on the site availability.

  10. ATLAS BigPanDA Monitoring

    CERN Document Server

    Padolski, Siarhei; The ATLAS collaboration; Klimentov, Alexei; Korchuganova, Tatiana

    2017-01-01

    BigPanDA monitoring is a web based application which provides various processing and representation of the Production and Distributed Analysis (PanDA) system objects states. Analyzing hundreds of millions of computation entities such as an event or a job BigPanDA monitoring builds different scale and levels of abstraction reports in real time mode. Provided information allows users to drill down into the reason of a concrete event failure or observe system bigger picture such as tracking the computation nucleus and satellites performance or the progress of whole production campaign. PanDA system was originally developed for the Atlas experiment and today effectively managing more than 2 million jobs per day distributed over 170 computing centers worldwide. BigPanDA is its core component commissioned in the middle of 2014 and now is the primary source of information for ATLAS users about state of their computations and the source of decision support information for shifters, operators and managers. In this wor...

  11. ATLAS BigPanDA Monitoring

    CERN Document Server

    Padolski, Siarhei; The ATLAS collaboration

    2017-01-01

    BigPanDA monitoring is a web-based application that provides various processing and representation of the Production and Distributed Analysis (PanDA) system objects states. Analysing hundreds of millions of computation entities such as an event or a job BigPanDA monitoring builds different scale and levels of abstraction reports in real time mode. Provided information allows users to drill down into the reason of a concrete event failure or observe system bigger picture such as tracking the computation nucleus and satellites performance or the progress of whole production campaign. PanDA system was originally developed for the Atlas experiment and today effectively managing more than 2 million jobs per day distributed over 170 computing centers worldwide. BigPanDA is its core component commissioned in the middle of 2014 and now is the primary source of information for ATLAS users about state of their computations and the source of decision support information for shifters, operators and managers. In this work...

  12. The ATLAS Distributed Data Management System & Databases

    CERN Document Server

    Garonne, V; The ATLAS collaboration; Barisits, M; Beermann, T; Vigne, R; Serfon, C

    2013-01-01

    The ATLAS Distributed Data Management (DDM) System is responsible for the global management of petabytes of high energy physics data. The current system, DQ2, has a critical dependency on Relational Database Management Systems (RDBMS), like Oracle. RDBMS are well-suited to enforcing data integrity in online transaction processing applications, however, concerns have been raised about the scalability of its data warehouse-like workload. In particular, analysis of archived data or aggregation of transactional data for summary purposes is problematic. Therefore, we have evaluated new approaches to handle vast amounts of data. We have investigated a class of database technologies commonly referred to as NoSQL databases. This includes distributed filesystems, like HDFS, that support parallel execution of computational tasks on distributed data, as well as schema-less approaches via key-value stores, like HBase. In this talk we will describe our use cases in ATLAS, share our experiences with various databases used ...

  13. ATLAS measurement of Electroweak Vector Boson production

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00453010; The ATLAS collaboration

    2017-01-01

    The measurements of the Drell-Yan production of W and Z/γ⁎ bosons at the LHC provide a benchmark of our understanding of the perturbative QCD and probe the proton structure in a unique way. The ATLAS collaboration has performed new high precision measurements of the double differential cross-sections as a function of the dilepton mass and rapidity. The measurements are compared to state of calculations at NNLO in QCD and constrain the photon content of the proton. The angular distributions of the Drell-Yan lepton pairs around the Z-boson mass peak probe the underlying QCD dynamics of the Z-boson production mechanisms. The complete set of angular coefficients describing these distributions is presented and compared to theoretical predictions highlighting different approaches of the QCD and EW modelling. First precise inclusive measurements of W and Z production at 13 TeV are presented. $W/Z$ and $W^{+}/W^{−}$ ratios profit from a cancellation of experimental uncertainties.

  14. The Evolution of Cloud Computing in ATLAS

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00224309; Berghaus, Frank; Brasolin, Franco; Cordeiro, Cristovao; Desmarais, Ron; Field, Laurence; Gable, Ian; Giordano, Domenico; Di Girolamo, Alessandro; Hover, John; Leblanc, Matthew Edgar; Love, Peter; Paterson, Michael; Sobie, Randall; Zaytsev, Alexandr

    2015-01-01

    The ATLAS experiment has successfully incorporated cloud computing technology and cloud resources into its primarily grid-based model of distributed computing. Cloud R&D activities continue to mature and transition into stable production systems, while ongoing evolutionary changes are still needed to adapt and refine the approaches used, in response to changes in prevailing cloud technology. In addition, completely new developments are needed to handle emerging requirements. This paper describes the overall evolution of cloud computing in ATLAS. The current status of the virtual machine (VM) management systems used for harnessing infrastructure as a service (IaaS) resources are discussed. Monitoring and accounting systems tailored for clouds are needed to complete the integration of cloud resources within ATLAS' distributed computing framework. We are developing and deploying new solutions to address the challenge of operation in a geographically distributed multi-cloud scenario, including a system for ma...

  15. The Evolution of Cloud Computing in ATLAS

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00224309; The ATLAS collaboration; Berghaus, Frank; Love, Peter; Leblanc, Matthew Edgar; Di Girolamo, Alessandro; Paterson, Michael; Gable, Ian; Sobie, Randall; Field, Laurence

    2015-01-01

    The ATLAS experiment has successfully incorporated cloud computing technology and cloud resources into its primarily grid-based model of distributed computing. Cloud R&D activities continue to mature and transition into stable production systems, while ongoing evolutionary changes are still needed to adapt and refine the approaches used, in response to changes in prevailing cloud technology. In addition, completely new developments are needed to handle emerging requirements. This work will describe the overall evolution of cloud computing in ATLAS. The current status of the VM management systems used for harnessing IAAS resources will be discussed. Monitoring and accounting systems tailored for clouds are needed to complete the integration of cloud resources within ATLAS' distributed computing framework. We are developing and deploying new solutions to address the challenge of operation in a geographically distributed multi-cloud scenario, including a system for managing VM images across multiple clouds, ...

  16. Performance of a proximity cryogenic system for the ATLAS central solenoid magnet

    CERN Document Server

    Doi, Y; Makida, Y; Kondo, Y; Kawai, M; Aoki, K; Haruyama, T; Kondo, T; Mizumaki, S; Wachi, Y; Mine, S; Haug, F; Delruelle, N; Passardi, Giorgio; ten Kate, H H J

    2002-01-01

    The ATLAS central solenoid magnet has been designed and constructed as a collaborative work between KEK and CERN for the ATLAS experiment in the LHC project The solenoid provides an axial magnetic field of 2 Tesla at the center of the tracking volume of the ATLAS detector. The solenoid is installed in a common cryostat of a liquid-argon calorimeter in order to minimize the mass of the cryostat wall. The coil is cooled indirectly by using two-phase helium flow in a pair of serpentine cooling line. The cryogen is supplied by the ATLAS cryogenic plant, which also supplies helium to the Toroid magnet systems. The proximity cryogenic system for the solenoid has two major components: a control dewar and a valve unit In addition, a programmable logic controller, PLC, was prepared for the automatic operation and solenoid test in Japan. This paper describes the design of the proximity cryogenic system and results of the performance test. (7 refs).

  17. Evolution of the ATLAS Distributed Computing system during the LHC Long shutdown

    CERN Document Server

    Campana, S; The ATLAS collaboration

    2014-01-01

    The ATLAS Distributed Computing project (ADC) was established in 2007 to develop and operate a framework, following the ATLAS computing model, to enable data storage, processing and bookkeeping on top of the WLCG distributed infrastructure. ADC development has always been driven by operations and this contributed to its success. The system has fulfilled the demanding requirements of ATLAS, daily consolidating worldwide up to 1PB of data and running more than 1.5 million payloads distributed globally, supporting almost one thousand concurrent distributed analysis users. Comprehensive automation and monitoring minimized the operational manpower required. The flexibility of the system to adjust to operational needs has been important to the success of the ATLAS physics program. The LHC shutdown in 2013-2015 affords an opportunity to improve the system in light of operational experience and scale it to cope with the demanding requirements of 2015 and beyond, most notably a much higher trigger rate and event pileu...

  18. Measurements of the Vector boson production with the ATLAS Detector

    CERN Document Server

    Weber, Stephen Albert; The ATLAS collaboration

    2018-01-01

    Measurements of the Drell-Yan production of W and Z bosons at the LHC are an essential part of the ATLAS standard model physics program. These processes are a probe to understand perturbative QCD and the structure of the proton. Differential measurements of the Z boson cross section produced in association with jets can be compared to different LO and NLO MC generators to test the accuracy of these predictions. A selection of recent results are presented on varying studies of W and Z boson production.

  19. Operation of the Upgraded ATLAS Level-1 Central Trigger System

    CERN Document Server

    Glatzer, Julian Maximilian Volker; The ATLAS collaboration

    2015-01-01

    The ATLAS Level-1 Central Trigger (L1CT) system is a central part of ATLAS data-taking and has undergone a major upgrade for Run 2 of the LHC, in order to cope with the expected increase of instantaneous luminosity of a factor of 2 with respect to Run 1. The upgraded hardware offers more flexibility in the trigger decisions due to the double amount of trigger inputs and usable trigger channels. It also provides an interface to the new topological trigger system. Operationally - particularly useful for commissioning, calibration and test runs - it allows concurrent running of up to 3 different sub-detector combinations. In this contribution, we give an overview of the operational software framework of the L1CT system with particular emphasis of the configuration, controls and monitoring aspects. The software framework allows a consistent configuration with respect to the ATLAS experiment and the LHC machine, upstream and downstream trigger processors, and the data acquisition. Trigger and dead-time rates are m...

  20. Diagnostic Systems and Resources utilization of the ATLAS High Level Trigger

    CERN Document Server

    Sidoti, A; The ATLAS collaboration; Ospanov, R

    2010-01-01

    Since the LHC started colliding protons in December 2009, the ATLAS trigger has operated very successfully with a collision rate which has increased by several orders of magnitude. The trigger monitoring and data quality infrastructure was essential to this success. We describe the software tools used to monitor the trigger system performance and assess the overall quality of the trigger selection during collisions running. ATLAS has broad physics goals which require a large number of different active triggers due to complex event topology, requiring quite sophisticated software structures and concepts. The trigger of the ATLAS experiment is built as a three level system. The first level is realized in hardware while the high level triggers (HLT) are software based and run on large PC farms. The trigger reduces the bunch crossing rate of 40 MHz, at design, to an average event rate of about 200 Hz for storage. Since the ATLAS detector is a general purpose detector, the trigger must be sensitive to a large numb...

  1. The Future of Distributed Computing Systems in ATLAS: Boldly Venturing Beyond Grids

    CERN Document Server

    Barreiro Megino, Fernando Harald; The ATLAS collaboration

    2018-01-01

    The Production and Distributed Analysis system (PanDA) for the ATLAS experiment at the Large Hadron Collider has seen big changes over the past couple of years to accommodate new types of distributed computing resources: clouds, HPCs, volunteer computers and other external resources. While PanDA was originally designed for fairly homogeneous resources available through the Worldwide LHC Computing Grid, the new resources are heterogeneous, at diverse scales and with diverse interfaces. Up to a fifth of the resources available to ATLAS are of such new types and require special techniques for integration into PanDA. In this talk, we present the nature and scale of these resources. We provide an overview of the various challenges faced, spanning infrastructure, software distribution, workload requirements, scaling requirements, workflow management, data management, network provisioning, and associated software and computing facilities. We describe the strategies for integrating these heterogeneous resources into ...

  2. ATLAS database application enhancements using Oracle 11g

    CERN Document Server

    Dimitrov, G; The ATLAS collaboration; Blaszczyk, M; Sorokoletov, R

    2012-01-01

    The ATLAS experiment at LHC relies on databases for detector online data-taking, storage and retrieval of configurations, calibrations and alignments, post data-taking analysis, file management over the grid, job submission and management, condition data replication to remote sites. Oracle Relational Database Management System (RDBMS) has been addressing the ATLAS database requirements to a great extent for many years. Ten database clusters are currently deployed for the needs of the different applications, divided in production, integration and standby databases. The data volume, complexity and demands from the users are increasing steadily with time. Nowadays more than 20 TB of data are stored in the ATLAS production Oracle databases at CERN (not including the index overhead), but the most impressive number is the hosted 260 database schemas (for the most common case each schema is related to a dedicated client application with its own requirements). At the beginning of 2012 all ATLAS databases at CERN have...

  3. Standard Model tt̄H Production at ATLAS & CMS

    CERN Document Server

    Thomas-Wilsker, Joshuha; The ATLAS collaboration

    2016-01-01

    The Run II ttH analyses benefit from a large increase in the ttH signal cross-section making it a particularly attractive production channel. However, the particularly challenging backgrounds seen in most channels require innovative techniques in order to make the analyses as sensitive as possible. Presented are the first results from the ATLAS ttH analyses using 13TeV collisions at the LHC and updated results presented by CMS at Moriond 2016.

  4. Improving ATLAS grid site reliability with functional tests using HammerCloud

    Science.gov (United States)

    Elmsheuser, Johannes; Legger, Federica; Medrano Llamas, Ramon; Sciacca, Gianfranco; van der Ster, Dan

    2012-12-01

    With the exponential growth of LHC (Large Hadron Collider) data in 2011, and more coming in 2012, distributed computing has become the established way to analyse collider data. The ATLAS grid infrastructure includes almost 100 sites worldwide, ranging from large national computing centers to smaller university clusters. These facilities are used for data reconstruction and simulation, which are centrally managed by the ATLAS production system, and for distributed user analysis. To ensure the smooth operation of such a complex system, regular tests of all sites are necessary to validate the site capability of successfully executing user and production jobs. We report on the development, optimization and results of an automated functional testing suite using the HammerCloud framework. Functional tests are short lightweight applications covering typical user analysis and production schemes, which are periodically submitted to all ATLAS grid sites. Results from those tests are collected and used to evaluate site performances. Sites that fail or are unable to run the tests are automatically excluded from the PanDA brokerage system, therefore avoiding user or production jobs to be sent to problematic sites.

  5. The detector control system of the ATLAS experiment

    International Nuclear Information System (INIS)

    Poy, A Barriuso; Burckhart, H J; Cook, J; Franz, S; Gutzwiller, O; Hallgren, B; Schlenker, S; Varela, F; Boterenbrood, H; Filimonov, V; Khomutnikov, V

    2008-01-01

    The ATLAS experiment is one of the experiments at the Large Hadron Collider, constructed to study elementary particle interactions in collisions of high-energy proton beams. The individual detector components as well as the common experimental infrastructure are supervised by the Detector Control System (DCS). The DCS enables equipment supervision using operator commands, reads, processes and archives the operational parameters of the detector, allows for error recognition and handling, manages the communication with external control systems, and provides a synchronization mechanism with the physics data acquisition system. Given the enormous size and complexity of ATLAS, special emphasis was put on the use of standardized hardware and software components enabling efficient development and long-term maintainability of the DCS over the lifetime of the experiment. Currently, the DCS is being used successfully during the experiment commissioning phase

  6. Large Scale Software Building with CMake in ATLAS

    Science.gov (United States)

    Elmsheuser, J.; Krasznahorkay, A.; Obreshkov, E.; Undrus, A.; ATLAS Collaboration

    2017-10-01

    The offline software of the ATLAS experiment at the Large Hadron Collider (LHC) serves as the platform for detector data reconstruction, simulation and analysis. It is also used in the detector’s trigger system to select LHC collision events during data taking. The ATLAS offline software consists of several million lines of C++ and Python code organized in a modular design of more than 2000 specialized packages. Because of different workflows, many stable numbered releases are in parallel production use. To accommodate specific workflow requests, software patches with modified libraries are distributed on top of existing software releases on a daily basis. The different ATLAS software applications also require a flexible build system that strongly supports unit and integration tests. Within the last year this build system was migrated to CMake. A CMake configuration has been developed that allows one to easily set up and build the above mentioned software packages. This also makes it possible to develop and test new and modified packages on top of existing releases. The system also allows one to detect and execute partial rebuilds of the release based on single package changes. The build system makes use of CPack for building RPM packages out of the software releases, and CTest for running unit and integration tests. We report on the migration and integration of the ATLAS software to CMake and show working examples of this large scale project in production.

  7. ATLAS studies of spectroscopy and B-decays

    CERN Document Server

    Dolezal, Zdenek; The ATLAS collaboration

    2015-01-01

    ATLAS has a wide programme to study the production cross section and decay properties of particles with beauty, as well as charmonium and bottomonium states. This presentation will cover the latest ATLAS results, including searches for excited b-hadrons, new decay modes of b-hadrons, indirect New Physics searches in rare B-hadron decays, as well as sthe study of mixing and CP violation in the Bs system. We will also review the results in the domain of charmonium production, including J/psi, psi(2s) and chi_c states, associated vector boson + J/psi production and search for hidden-beauty state Xb, the bottomonium counterpart of X(3876) in the Upsilon pi pi channel.

  8. Deploying the ATLAS Metadata Interface (AMI) on the cloud with Jenkins

    CERN Document Server

    Lambert, Fabian; The ATLAS collaboration

    2016-01-01

    The ATLAS Metadata Interface (AMI) is a mature application of more than 15 years of existence. Mainly used by the ATLAS experiment at CERN, it consists of a very generic tool ecosystem for metadata aggregation and cataloguing. AMI is used by the ATLAS production system, therefore the service must guarantee a high level of availability. We describe our monitoring system and the Jenkins-based strategy used to dynamically test and deploy cloud OpenStack nodes on demand. Moreover, we describe how to switch to a distant replica in case of downtime.

  9. Development and test of the DAQ system for a Micromegas prototype installed into the ATLAS experiment

    CERN Document Server

    Zibell, Andre; The ATLAS collaboration; Bianco, Michele; Martoiu, Victor Sorin

    2015-01-01

    A Micromegas (MM) quadruplet prototype with an active area of 0.5 m$^2$ that adopts the general design foreseen for the upgrade of the innermost forward muon tracking systems (Small Wheels) of the ATLAS detector in 2018-2019, has been built at CERN and is going to be tested in the ATLAS cavern environment during the LHC RUN-II period 2015-2017. The integration of this prototype detector into the ATLAS data acquisition system using custom ATCA equipment is presented. An ATLAS compatible ReadOutDriver (ROD) based on the Scalable Readout System (SRS), the Scalable Readout Unit (SRU), will be used in order to transmit the data after generating valid event fragments to the high-level Read Out System (ROS). The SRU will be synchronized with the LHC bunch crossing clock (40.08 MHz) and will receive the Level-1 trigger signals from the Central Trigger Processor (CTP) through the TTCrx receiver ASIC. The configuration of the system will be driven directly from the ATLAS Run Control System. By using the ATLAS TDAQ Soft...

  10. Next generation PanDA pilot for ATLAS and other experiments

    International Nuclear Information System (INIS)

    Nilsson, P; De, K; Megino, F Barreiro; Llamas, R Medrano; Bejar, J Caballero; Hover, J; Maeno, T; Wenaus, T; Love, P; Walker, R

    2014-01-01

    The Production and Distributed Analysis system (PanDA) has been in use in the ATLAS Experiment since 2005. It uses a sophisticated pilot system to execute submitted jobs on the worker nodes. While originally designed for ATLAS, the PanDA Pilot has recently been refactored to facilitate use outside of ATLAS. Experiments are now handled as plug-ins such that a new PanDA Pilot user only has to implement a set of prototyped methods in the plug-in classes, and provide a script that configures and runs the experiment-specific payload. We will give an overview of the Next Generation PanDA Pilot system and will present major features and recent improvements including live user payload debugging, data access via the Federated XRootD system, stage-out to alternative storage elements, support for the new ATLAS DDM system (Rucio), and an improved integration with glExec, as well as a description of the experiment-specific plug-in classes. The performance of the pilot system in processing LHC data on the OSG, LCG and Nordugrid infrastructures used by ATLAS will also be presented. We will describe plans for future development on the time scale of the next few years.

  11. ATLAS construction schedule

    CERN Multimedia

    Kotamaki, M

    The goal during the last few months has been to freeze and baseline as much as possible the schedules of various ATLAS systems and activities. The main motivations for the re-baselining of the schedules have been the new LHC schedule aiming at first collisions in early 2006 and the encountered delays in civil engineering as well as in the production of some of the detectors. The process was started by first preparing a new installation schedule that takes into account all the new external constraints and the new ATLAS staging scenario. The installation schedule version 3 was approved in the March EB and it provides the Ready For Installation (RFI) milestones for each system, i.e. the date when the system should be available for the start of the installation. TCn is now interacting with the systems aiming at a more realistic and resource loaded version 4 before the end of the year. Using the new RFI milestones as driving dates a new summary schedule has been prepared, or is under preparation, for each system....

  12. Measurement of charmonium production in heavy-ion collisions with the ATLAS detector

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00511724; The ATLAS collaboration

    2017-01-01

    The suppression of heavy charmonia states in heavy-ion collisions is a phenomenon understood as a consequence of quark gluon plasma formation in the hot, dense system formed in heavy ion collisions at the LHC. In addition to hot matter effects in heavy-ion collisions, cold nuclear effects may also affect heavy charmonia production. Therefore, a full assessment requires detailed studies on the effects present in both A+A and p+A collisions. Based on p+Pb data collected in 2013 and pp and Pb+Pb data collected in 2015 at the LHC, the ATLAS experiment has studied prompt and non-prompt $J/\\psi$ and $\\psi$(2S) productions via the dimuon decay final states. The production and excited-to-ground state ratios of heavy charmonia measured in both p+Pb and Pb+Pb collision data with respect to that measured in pp collision data will be presented in intervals of transverse momentum, rapidity and centrality.

  13. First experience and adaptation of existing tools to ATLAS distributed analysis

    International Nuclear Information System (INIS)

    De La Hoz, S.G.; Ruiz, L.M.; Liko, D.

    2008-01-01

    The ATLAS production system has been successfully used to run production of simulation data at an unprecedented scale in ATLAS. Up to 10000 jobs were processed on about 100 sites in one day. The experiences obtained operating the system on several grid flavours was essential to perform a user analysis using grid resources. First tests of the distributed analysis system were then performed. In the preparation phase data was registered in the LHC file catalog (LFC) and replicated in external sites. For the main test, few resources were used. All these tests are only a first step towards the validation of the computing model. The ATLAS management computing board decided to integrate the collaboration efforts in distributed analysis in only one project, GANGA. The goal is to test the reconstruction and analysis software in a large scale Data production using grid flavors in several sites. GANGA allows trivial switching between running test jobs on a local batch system and running large-scale analyses on the grid; it provides job splitting and merging, and includes automated job monitoring and output retrieval. (orig.)

  14. Di-boson production measurements with the ATLAS detector

    CERN Document Server

    Chevalier, Laurent

    2015-01-01

    This talk covers the results of the measurements of diboson production cross sections in proton- proton interactions at 7 TeV in 2011 and at 8 TeV in 2012, performed by the ATLAS experiment. The cross sections are measured in phase space regions defined by the decay kinematics, the detector acceptance and then extrapolated to the full phase space. Cross sections for VV (V=W, Z, γ) production in the leptonic or semileptonic channels are compared to (N)NLO predictions of the Standard Model. Kinematic spectra are used to place constraints on anomalous triple-gauge-boson couplings. All cross sections are in agreement with the Standard Model predictions. No experimental evidence is found of anomalous triple-gauge-boson couplings. PACS: 14.70.-e Gauge bosons.

  15. Measurement of photon production cross sections also in association with jets with the ATLAS detector

    CERN Document Server

    Tassi, Enrico; The ATLAS collaboration

    2017-01-01

    The production of prompt isolated photons at hadron colliders provides a stringent test of perturbative QCD and can be used to probe the proton structure. The ATLAS collaboration has performed precise measurements of the inclusive production of isolated prompt photons at a center-of-mass energy of 13 TeV, differential in both rapidity and the photon transverse momentum. In addition, the integrated and differential cross sections for isolated photon pairs and tri-photon production 8 TeV have been measured. The results are compared with state-of-the-art theory predictions at NLO in QCD and with predictions of several MC generators. The production of prompt photons in association with jets provides an additional testing ground for perturbative QCD (pQCD) with a hard colourless probe less affected by hadronisation effects than jet production. The ATLAS collaboration has studied the dynamics of isolated-photon production in association with gluon, light and heavy quark final states in pp collisions at a centre-of-...

  16. Argonne Tandem Linac Accelerator System (ATLAS)

    Data.gov (United States)

    Federal Laboratory Consortium — ATLAS is a national user facility at Argonne National Laboratory in Argonne, Illinois. The ATLAS facility is a leading facility for nuclear structure research in the...

  17. A High-Resolution In Vivo Atlas of the Human Brain's Serotonin System.

    Science.gov (United States)

    Beliveau, Vincent; Ganz, Melanie; Feng, Ling; Ozenne, Brice; Højgaard, Liselotte; Fisher, Patrick M; Svarer, Claus; Greve, Douglas N; Knudsen, Gitte M

    2017-01-04

    The serotonin (5-hydroxytryptamine, 5-HT) system modulates many important brain functions and is critically involved in many neuropsychiatric disorders. Here, we present a high-resolution, multidimensional, in vivo atlas of four of the human brain's 5-HT receptors (5-HT 1A , 5-HT 1B , 5-HT 2A , and 5-HT 4 ) and the 5-HT transporter (5-HTT). The atlas is created from molecular and structural high-resolution neuroimaging data consisting of positron emission tomography (PET) and magnetic resonance imaging (MRI) scans acquired in a total of 210 healthy individuals. Comparison of the regional PET binding measures with postmortem human brain autoradiography outcomes showed a high correlation for the five 5-HT targets and this enabled us to transform the atlas to represent protein densities (in picomoles per milliliter). We also assessed the regional association between protein concentration and mRNA expression in the human brain by comparing the 5-HT density across the atlas with data from the Allen Human Brain atlas and identified receptor- and transporter-specific associations that show the regional relation between the two measures. Together, these data provide unparalleled insight into the serotonin system of the human brain. We present a high-resolution positron emission tomography (PET)- and magnetic resonance imaging-based human brain atlas of important serotonin receptors and the transporter. The regional PET-derived binding measures correlate strongly with the corresponding autoradiography protein levels. The strong correlation enables the transformation of the PET-derived human brain atlas into a protein density map of the serotonin (5-hydroxytryptamine, 5-HT) system. Next, we compared the regional receptor/transporter protein densities with mRNA levels and uncovered unique associations between protein expression and density at high detail. This new in vivo neuroimaging atlas of the 5-HT system not only provides insight in the human brain's regional protein

  18. The Locomotive is running full speed in the ATLAS MUONs

    CERN Multimedia

    Mikenberg, G.

    The ATLAS MUON Spectrometer is, like most of the ATLAS systems, a large collection of detectors that operate at the limit of the technology. They have to provide the MUON trigger for the ATLAS detector over very large surfaces (7000m2) and measure the passage of MUONs over distances ranging between 5 to 13m, with relative precisions between the various measurement planes of few tenths of microns, while controlling various external parameters ranging from the relative positions of the detectors (alignment systems controlled to the level of 20 microns) to the magnetic field (to be reconstructed at the level of 20 Gauss). Although many of the integration problems with the rest of the ATLAS detectors have not been fully clarified, one needs to start production, in order to be ready on time to enjoy the Physics of the LHC. This means to start the coordinated work in more than 25 production and testing sites, located all around the world, that have to produce precision detectors at industrial speed, which sho...

  19. ATLAS Global Shares Implementation in the PanDA Workload Management System

    CERN Document Server

    Barreiro Megino, Fernando Harald; The ATLAS collaboration

    2018-01-01

    PanDA (Production and Distributed Analysis) is the workload management system for ATLAS across the Worldwide LHC Computing Grid. While analysis tasks are submitted to PanDA by over a thousand users following personal schedules (e.g. PhD or conference deadlines), production campaigns are scheduled by a central Physics Coordination group based on the organization’s calendar. The Physics Coordination group needs to allocate the amount of Grid resources dedicated to each activity, in order to manage sharing of CPU resources among various parallel campaigns and to make sure that results can be achieved in time for important deadlines. While dynamic and static shares on batch systems have been around for a long time, we are trying to move away from local resource partitioning and manage shares at a global level in the PanDA system. The global solution is not straightforward, given different requirements of the activities (number of cores, memory, I/O and CPU intensity), the heterogeneity of Grid resources (site/H...

  20. The Database Driven ATLAS Trigger Configuration System

    CERN Document Server

    Martyniuk, Alex; The ATLAS collaboration

    2015-01-01

    This contribution describes the trigger selection configuration system of the ATLAS low- and high-level trigger (HLT) and the upgrades it received in preparation for LHC Run 2. The ATLAS trigger configuration system is responsible for applying the physics selection parameters for the online data taking at both trigger levels and the proper connection of the trigger lines across those levels. Here the low-level trigger consists of the already existing central trigger (CT) and the new Level-1 Topological trigger (L1Topo), which has been added for Run 2. In detail the tasks of the configuration system during the online data taking are Application of the selection criteria, e.g. energy cuts, minimum multiplicities, trigger object correlation, at the three trigger components L1Topo, CT, and HLT On-the-fly, e.g. rate-dependent, generation and application of prescale factors to the CT and HLT to adjust the trigger rates to the data taking conditions, such as falling luminosity or rate spikes in the detector readout ...

  1. Module and electronics developments for the ATLAS ITK pixel system

    CERN Document Server

    Munoz Sanchez, Francisca Javiela; The ATLAS collaboration

    2017-01-01

    The ATLAS experiment is preparing for an extensive modification of its detectors in the course of the planned HL-LHC accelerator upgrade around 2025. The ATLAS upgrade includes the replacement of the entire tracking system by an all-silicon detector (Inner Tracker, ITk). The five innermost layers of ITk will be a pixel detector built of new sensor and readout electronics technologies to improve the tracking performance and cope with the severe HL-LHC environment in terms of occupancy and radiation. The total area of the new pixel system could measure up to 14 m2, depending on the final layout choice, which is expected to take place in 2017. In this paper an overview of the ongoing R\\&D activities on modules and electronics for the ATLAS ITk is given including the main developments and achievements in silicon planar and 3D sensor technologies, readout and power challenges.

  2. A High-Resolution In Vivo Atlas of the Human Brain's Serotonin System

    DEFF Research Database (Denmark)

    Beliveau, Vincent; Ganz-Benjaminsen, Melanie; Feng, Ling

    2017-01-01

    The serotonin (5-hydroxytryptamine, 5-HT) system modulates many important brain functions and is critically involved in many neuropsychiatric disorders. Here, we present a high-resolution, multidimensional, in vivo atlas of four of the human brain's 5-HT receptors (5-HT1A, 5-HT1B, 5-HT2A, and 5-HT4...... with postmortem human brain autoradiography outcomes showed a high correlation for the five 5-HT targets and this enabled us to transform the atlas to represent protein densities (in picomoles per milliliter). We also assessed the regional association between protein concentration and mRNA expression in the human...... brain by comparing the 5-HT density across the atlas with data from the Allen Human Brain atlas and identified receptor- and transporter-specific associations that show the regional relation between the two measures. Together, these data provide unparalleled insight into the serotonin system...

  3. Measurement of Single-top Quark Production with ATLAS Data

    CERN Document Server

    Holzbauer, J L

    2011-01-01

    Single-top production processes have been studied using 0.7 fb-1 of data from 7 TeV center-of-mass energy proton-proton collisions collected with the ATLAS detector at the LHC. Single-top is electroweak top production and the standard model includes three production modes. Each contains a Wtb vertex, allowing the possibility of a direct measurement of the CKM matrix element |Vtb|. Single-top could also be sensitive to new physics, such as flavor changing neutral currents or heavy W' bosons. Using cut-based selections, a limit of < 39.1 pb is set for dilepton Wt production and < 26.5 pb for s-channel production. For the t-channel measurement, both cut-based and neural network analyses are performed and the cross-section is measured to be 90 +32 -22 pb, where 65 +28 -19 pb is expected according to standard model.

  4. The ATLAS High Level Trigger Steering Framework and the Trigger 
Configuration System.

    CERN Document Server

    Pérez Cavalcanti, Tiago; The ATLAS collaboration

    2011-01-01

    The ATLAS High Level Trigger Steering Framework and the Trigger 
Configuration System.
 
The ATLAS detector system installed in the Large Hadron Collider (LHC) 
at CERN is designed to study proton-proton and nucleus-nucleus 
collisions with a maximum center of mass energy of 14 TeV at a bunch 
collision rate of 40MHz.  In March 2010 the four LHC experiments saw 
the first proton-proton collisions at 7 TeV. Still within the year a 
collision rate of nearly 10 MHz is expected. At ATLAS, events of 
potential interest for ATLAS physics are selected by a three-level 
trigger system, with a final recording rate of about 200 Hz. The first 
level (L1) is implemented in custom hardware; the two levels of 
the high level trigger (HLT) are software triggers, running on large 
farms of standard computers and network devices. 

Within the ATLAS physics program more than 500 trigger signatures are 
defined. The HLT tests each signature on each L1-accepted event; the 
test outcome is recor...

  5. Study of Multiboson Production with the ATLAS detector

    CERN Document Server

    Kubota, Takashi; The ATLAS collaboration

    2017-01-01

    Measurements of the cross sections of the production of pairs of electroweak gauge bosons at the LHC constitute stringent tests of the electroweak sector of the Standard Model and provide a model-independent means to search for new physics at the TeV scale. The ATLAS collaboration has performed detailed measurements of integrated and differential cross sections of the production of heavy di-boson pairs, such as WW, WZ and ZZ, in the fully-leptonic and partially in the semi-leptonic final states at centre-of-mass energies of 8 and 13 TeV. Moreover, searches for the production of three W bosons or of a W boson and a photon together with a Z or W boson at a center of mass energy of 8 TeV will be presented. These studies are closely connected to the electroweak production of a heavy boson and a photon together with two jets. Evidence has been found for the exclusive production of W boson pairs, which will be presented in this talk.

  6. The Run-2 ATLAS Trigger System: Design, Performance and Plan

    CERN Document Server

    zur Nedden, Martin; The ATLAS collaboration

    2016-01-01

    In high-energy physics experiments, online selection is crucial to select interesting collisions from the large data volume. The ATLAS experiment at the Large Hadron Collider (LHC) utilizes the trigger system that consists of a hardware Level-1 (L1) and a software based high-level trigger (HLT), reducing the event rate from the design bunch-crossing rate of 40 MHz to an average recording rate of about 1000 Hz. In the LHC Run-2 starting from in 2015, the LHC operates at centre-of-mass energy of 13 TeV providing a luminosity up to $1.2 \\cdot 10^{34} {\\rm cm^{-2}s^{-1}}$. The ATLAS trigger system has to cope with these challenges, while maintaining or even improving the efficiency to select relevant physics processes. In this paper, the ATLAS trigger system for LHC Run-2 is reviewed. Secondly, the impressive performance improvements in the HLT trigger algorithms used to identify leptons, hadrons and global event quantities like missing transverse energy is shown. Electron, muon and photon triggers covering trans...

  7. ATLAS utilisation of the Czech national HPC center

    CERN Document Server

    Svatos, Michal; The ATLAS collaboration

    2018-01-01

    The Czech national HPC center IT4Innovations located in Ostrava provides two HPC systems, Anselm and Salomon. The Salomon HPC is amongst the hundred most powerful supercomputers on Earth since its commissioning in 2015. Both clusters were tested for usage by the ATLAS experiment for running simulation jobs. Several thousand core hours were allocated to the project for tests, but the main aim is to use free resources waiting for large parallel jobs of other users. Multiple strategies for ATLAS job execution were tested on the Salomon and Anselm HPCs. The solution described herein is based on the ATLAS experience with other HPC sites. ARC Compute Element (ARC-CE) installed at the grid site in Prague is used for job submission to Salomon. The ATLAS production system submits jobs to the ARC-CE via ARC Control Tower (aCT). The ARC-CE processes job requirements from aCT and creates a script for a batch system which is then executed via ssh. Sshfs is used to share scripts and input files between the site and the HPC...

  8. ATLAS Tile calorimeter calibration and monitoring systems

    CERN Document Server

    Boumediene, Djamel Eddine; The ATLAS collaboration

    2017-01-01

    The ATLAS Tile Calorimeter (TileCal) is the central section of the hadronic calorimeter of the ATLAS experiment and provides important information for reconstruction of hadrons, jets, hadronic decays of tau leptons and missing transverse energy. This sampling calorimeter uses steel plates as absorber and scintillating tiles as active medium. The light produced by the passage of charged particles is transmitted by wavelength shifting fibres to photomultiplier tubes (PMTs). PMT signals are then digitized at 40 MHz and stored on detector and are only transferred off detector once the first level trigger acceptance has been confirmed. The readout is segmented into about 5000 cells (longitudinally and transversally), each of them being read out by two PMTs in parallel. To calibrate and monitor the stability and performance of each part of the readout chain, a set of calibration systems is used. The TileCal calibration system comprises Cesium radioactive sources, laser, charge injection elements and an integrator b...

  9. Measurement of Standard Model VBS/VBF production with the ATLAS and CMS detectors

    CERN Document Server

    Li, Shu; The ATLAS collaboration

    2017-01-01

    Proceeding for the poster presentation at Blois2017, France on the topic of "Measurements of Standard Model VBS/VBF productions with ATLAS+CMS detectors" (ATL-PHYS-SLIDE-2017-333 https://cds.cern.ch/record/2267458) Deadline: 15/10/2017

  10. Complete distributed computing environment for a HEP experiment: experience with ARC-connected infrastructure for ATLAS

    International Nuclear Information System (INIS)

    Read, A; Taga, A; O-Saada, F; Pajchel, K; Samset, B H; Cameron, D

    2008-01-01

    Computing and storage resources connected by the Nordugrid ARC middleware in the Nordic countries, Switzerland and Slovenia are a part of the ATLAS computing Grid. This infrastructure is being commissioned with the ongoing ATLAS Monte Carlo simulation production in preparation for the commencement of data taking in 2008. The unique non-intrusive architecture of ARC, its straightforward interplay with the ATLAS Production System via the Dulcinea executor, and its performance during the commissioning exercise is described. ARC support for flexible and powerful end-user analysis within the GANGA distributed analysis framework is also shown. Whereas the storage solution for this Grid was earlier based on a large, distributed collection of GridFTP-servers, the ATLAS computing design includes a structured SRM-based system with a limited number of storage endpoints. The characteristics, integration and performance of the old and new storage solutions are presented. Although the hardware resources in this Grid are quite modest, it has provided more than double the agreed contribution to the ATLAS production with an efficiency above 95% during long periods of stable operation

  11. Complete distributed computing environment for a HEP experiment: experience with ARC-connected infrastructure for ATLAS

    Energy Technology Data Exchange (ETDEWEB)

    Read, A; Taga, A; O-Saada, F; Pajchel, K; Samset, B H; Cameron, D [Department of Physics, University of Oslo, P.b. 1048 Blindern, N-0316 Oslo (Norway)], E-mail: a.l.read@fys.uio.no

    2008-07-15

    Computing and storage resources connected by the Nordugrid ARC middleware in the Nordic countries, Switzerland and Slovenia are a part of the ATLAS computing Grid. This infrastructure is being commissioned with the ongoing ATLAS Monte Carlo simulation production in preparation for the commencement of data taking in 2008. The unique non-intrusive architecture of ARC, its straightforward interplay with the ATLAS Production System via the Dulcinea executor, and its performance during the commissioning exercise is described. ARC support for flexible and powerful end-user analysis within the GANGA distributed analysis framework is also shown. Whereas the storage solution for this Grid was earlier based on a large, distributed collection of GridFTP-servers, the ATLAS computing design includes a structured SRM-based system with a limited number of storage endpoints. The characteristics, integration and performance of the old and new storage solutions are presented. Although the hardware resources in this Grid are quite modest, it has provided more than double the agreed contribution to the ATLAS production with an efficiency above 95% during long periods of stable operation.

  12. Production and Properties of B-Hadrons with the ATLAS Detector

    CERN Document Server

    Reznicek, Pavel; The ATLAS collaboration

    2017-01-01

    A wide program of studies on heavy flavours is performed with the ATLAS detector. Production cross sections have been measured for hadrons with b and c quark, for quarkonia states, and for associated production J/psi+J/psi, J/psi+muon, J/psi+W and J/psi+Z. The talk will discuss recent results on J/psi+J/psi prompt production (including the extraction of the double-parton-scattering component), and on the kinematical correlation in B hadron pair production, studied through their inclusive decays to J/psi and muons, respectively. Exotic states containing b and c quark have also been studies, and the talk will report on the searches for structures in the invariant mass of J/psi+hadron in the decay products of Lambda_b, and in the mass spectrum of Bs+pion.

  13. Techniques developed for the ATLAS Thin Gap Chambers mass production in Japan

    CERN Document Server

    Tanaka, S; Ishii, K; Iwasaki, H; Arataki, Y; Bando, T; Homma, Y; Ishino, M; Kondo, T; Kobayashi, T; Kurashige, H; Mikenberg, G; Miyazaki, Y; Nakagawa, Y; Nanjo, H; Ikeno, M; Nozaki, M; Ochi, A; Sasaki, O; Shoa, M; Sugimoto, T; Takeda, H; Takeshita, T; Yokoyama, C

    2004-01-01

    The Thin Gap Chambers (TGCs) are used for the muon trigger system in the end-cap regions of the ATLAS detector. The TGC mass production phase at High Energy Accelerator Research Organization (KEK) started in January 2001. As the anode-cathode distance is small, 1.4 mm, chamber flatness is essential to achieve a uniform gas gain over the chamber. In order to perform a stable production with high quality we developed a chamber closing system. When we glue two half-chambers together, we sandwich them between a granite table and an aluminum honeycomb panel to keep the chamber flat from both sides. By using silk screens, we control the quantity of epoxy adhesive that affects the chamber thickness. Due to these developments, we can achieve the flatness of less than 100 µm. Uniformity of detection efficiency of the TGC is measured with a cosmic-ray test bench at Kobe University. So far we have tested 300 TGCs. Position dependence of the efficiency is measured with a granularity of 5mm-by-5mm. The average efficiency...

  14. A Roadmap to Continuous Integration for ATLAS Software Development

    Science.gov (United States)

    Elmsheuser, J.; Krasznahorkay, A.; Obreshkov, E.; Undrus, A.; ATLAS Collaboration

    2017-10-01

    The ATLAS software infrastructure facilitates efforts of more than 1000 developers working on the code base of 2200 packages with 4 million lines of C++ and 1.4 million lines of python code. The ATLAS offline code management system is the powerful, flexible framework for processing new package versions requests, probing code changes in the Nightly Build System, migration to new platforms and compilers, deployment of production releases for worldwide access and supporting physicists with tools and interfaces for efficient software use. It maintains multi-stream, parallel development environment with about 70 multi-platform branches of nightly releases and provides vast opportunities for testing new packages, for verifying patches to existing software and for migrating to new platforms and compilers. The system evolution is currently aimed on the adoption of modern continuous integration (CI) practices focused on building nightly releases early and often, with rigorous unit and integration testing. This paper describes the CI incorporation program for the ATLAS software infrastructure. It brings modern open source tools such as Jenkins and GitLab into the ATLAS Nightly System, rationalizes hardware resource allocation and administrative operations, provides improved feedback and means to fix broken builds promptly for developers. Once adopted, ATLAS CI practices will improve and accelerate innovation cycles and result in increased confidence in new software deployments. The paper reports the status of Jenkins integration with the ATLAS Nightly System as well as short and long term plans for the incorporation of CI practices.

  15. The ATLAS Level-1 Trigger System with 13TeV nominal LHC collisions

    CERN Document Server

    Helary, Louis; The ATLAS collaboration

    2017-01-01

    The Level-1 (L1) Trigger system of the ATLAS experiment at CERN's Large Hadron Collider (LHC) plays a key role in the ATLAS detector data-taking. It is a hardware system that selects in real time events containing physics-motivated signatures. Selection is purely based on calorimetry energy depositions and hits in the muon chambers consistent with muon candidates. The L1 Trigger system has been upgraded to cope with the more challenging run-II LHC beam conditions, including increased centre-of-mass energy, increased instantaneous luminosity and higher levels of pileup. This talk summarises the improvements, commissioning and performance of the L1 ATLAS Trigger for the LHC run-II data period. The acceptance of muon triggers has been improved by increasing the hermiticity of the muon spectrometer. New strategies to obtain a better muon trigger signal purity were designed for certain geometrically difficult transition regions by using the ATLAS hadronic calorimeter. Algorithms to reduce noise spikes in muon trig...

  16. Next Generation PanDA Pilot for ATLAS and Other Experiments

    CERN Document Server

    Nilsson, P; The ATLAS collaboration; Caballero Bejar, J; De, K; Hover, J; Love, P; Maeno, T; Medrano Llamas, R; Walker, R; Wenaus, T

    2013-01-01

    The Production and Distributed Analysis system (PanDA) has been in use in the ATLAS Experiment since 2005. It uses a sophisticated pilot system to execute submitted jobs on the worker nodes. While originally designed for ATLAS, the PanDA Pilot has recently been refactored to facilitate use outside of ATLAS. Experiments are now handled as plug-ins such that a new PanDA Pilot user only has to implement a set of prototyped methods in the plug-in classes, and provide a script that configures and runs the experiment specific payload. We will give an overview of the Next Generation PanDA Pilot system and will present major features and recent improvements including live user payload debugging, data access via the Federated XRootD system, stage-out to alternative storage elements, support for the new ATLAS DDM system (Rucio), and an improved integration with glExec, as well as a description of the experiment specific plug-in classes. The performance of the pilot system in processing LHC data on the OSG, LCG and Nord...

  17. Next Generation PanDA Pilot for ATLAS and Other Experiments

    CERN Document Server

    Nilsson, P; The ATLAS collaboration; Caballero Bejar, J; De, K; Hover, J; Love, P; Maeno, T; Medrano Llamas, R; Walker, R; Wenaus, T

    2014-01-01

    The Production and Distributed Analysis system (PanDA) has been in use in the ATLAS Experiment since 2005. It uses a sophisticated pilot system to execute submitted jobs on the worker nodes. While originally designed for ATLAS, the PanDA Pilot has recently been refactored to facilitate use outside of ATLAS. Experiments are now handled as plug-ins such that a new PanDA Pilot user only has to implement a set of prototyped methods in the plug-in classes, and provide a script that configures and runs the experiment specific payload. We will give an overview of the Next Generation PanDA Pilot system and will present major features and recent improvements including live user payload debugging, data access via the Federated XRootD system, stage-out to alternative storage elements, support for the new ATLAS DDM system (Rucio), and an improved integration with glExec, as well as a description of the experiment specific plug-in classes. The performance of the pilot system in processing LHC data on the OSG, LCG and Nord...

  18. Setup, tests and results for the ATLAS TileCal Read Out Driver production

    CERN Document Server

    Valero, Alberto; Castillo, V; Cuenca, C; Ferrer, A; Fullana, E; González, V; Higón, E; Munar, A; Poveda, J; Ruiz-Martínez, A; Salvachúa, B; Sanchís, E; Solans, C; Soret, J; Torres, J; Valls, J A

    2007-01-01

    In this paper we describe the performance and test results of the production of the 38 ATLAS TileCal Read Out Drivers (RODs). We first describe the basic hardware specifications and firmware functionality of the modules, the test-bench setup used for production and the test procedure to qualify the boards. We then finally show and discuss the performance results.

  19. Status of the ATLAS experiment at CERN

    International Nuclear Information System (INIS)

    Taylor, G.

    2000-01-01

    Full text: The ATLAS experiment, to operate at CERN's Large Hadron Collider (LHC), from 2005 is currently under construction. The Australian HEP Consortium is participating in the development and construction of the semi-conductor tracker (SCT) sub-system of ATLAS. Australian scientists play significant roles in many aspects of the SCT detector module development, including design, prototyping, measurement, beam tests and simulation. Production facilities for construction of two hundred high precision detector modules for the SCT Forward region are well advanced in Melbourne and Sydney laboratories. This talk will give an overview of ATLAS experiment goals and status. It will concentrate on the Australian contribution. The talk will conclude with an outline of the future schedule and plans

  20. The Run-2 ATLAS Trigger System

    International Nuclear Information System (INIS)

    Martínez, A Ruiz

    2016-01-01

    The ATLAS trigger successfully collected collision data during the first run of the LHC between 2009-2013 at different centre-of-mass energies between 900 GeV and 8TeV. The trigger system consists of a hardware Level-1 and a software-based high level trigger (HLT) that reduces the event rate from the design bunch-crossing rate of 40 MHz to an average recording rate of a few hundred Hz. In Run-2, the LHC will operate at centre-of-mass energies of 13 and 14 TeV and higher luminosity, resulting in up to five times higher rates of processes of interest. A brief review of the ATLAS trigger system upgrades that were implemented between Run-1 and Run-2, allowing to cope with the increased trigger rates while maintaining or even improving the efficiency to select physics processes of interest, will be given. This includes changes to the Level-1 calorimeter and muon trigger systems, the introduction of a new Level-1 topological trigger module and the merging of the previously two-level HLT system into a single event processing farm. A few examples will be shown, such as the impressive performance improvements in the HLT trigger algorithms used to identify leptons, hadrons and global event quantities like missing transverse energy. Finally, the status of the commissioning of the trigger system and its performance during the 2015 run will be presented. (paper)

  1. Bringing ATLAS production to HPC resources - A use case with the Hydra supercomputer of the Max Planck Society

    CERN Document Server

    Kennedy, John; The ATLAS collaboration; Mazzaferro, Luca; Walker, Rodney

    2015-01-01

    The possible usage of HPC resources by ATLAS is now becoming viable due to the changing nature of these systems and it is also very attractive due to the need for increasing amounts of simulated data. In recent years the architecture of HPC systems has evolved, moving away from specialized monolithic systems, to a more generic Linux type platform. This change means that the deployment of non HPC specific codes has become much easier. The timing of this evolution perfectly suits the needs of ATLAS and opens a new window of opportunity. The ATLAS experiment at CERN will begin a period of high luminosity data taking in 2015. This high luminosity phase will be accompanied by a need for increasing amounts of simulated data which is expected to exceed the capabilities of the current Grid infrastructure. ATLAS aims to address this need by opportunistically accessing resources such as cloud and HPC systems. This paper presents the results of a pilot project undertaken by ATLAS and the MPP and RZG to provide access to...

  2. Improving ATLAS grid site reliability with functional tests using HammerCloud

    CERN Document Server

    Legger, F; The ATLAS collaboration

    2012-01-01

    With the exponential growth of LHC (Large Hadron Collider) data in 2011, and more coming in 2012, distributed computing has become the established way to analyse collider data. The ATLAS grid infrastructure includes almost 100 sites worldwide, ranging from large national computing centers to smaller university clusters. These facilities are used for data reconstruction and simulation, which are centrally managed by the ATLAS production system, and for distributed user analysis. To ensure the smooth operation of such a complex system, regular tests of all sites are necessary to validate the site capability of successfully executing user and production jobs. We report on the development, optimization and results of an automated functional testing suite using the HammerCloud framework. Functional tests are short light-weight applications covering typical user analysis and production schemes, which are periodically submitted to all ATLAS grid sites. Results from those tests are collected and used to evaluate site...

  3. Improving ATLAS grid site reliability with functional tests using HammerCloud

    CERN Document Server

    Legger, F; The ATLAS collaboration; Medrano Llamas, R; Sciacca, G; Van der Ster, D C

    2012-01-01

    With the exponential growth of LHC (Large Hadron Collider) data in 2011, and more coming in 2012, distributed computing has become the established way to analyse collider data. The ATLAS grid infrastructure includes more than 80 sites worldwide, ranging from large national computing centers to smaller university clusters. These facilities are used for data reconstruction and simulation, which are centrally managed by the ATLAS production system, and for distributed user analysis. To ensure the smooth operation of such a complex system, regular tests of all sites are necessary to validate the site capability of successfully executing user and production jobs. We report on the development, optimization and results of an automated functional testing suite using the HammerCloud framework. Functional tests are short light-weight applications covering typical user analysis and production schemes, which are periodically submitted to all ATLAS grid sites. Results from those tests are collected and used to evaluate si...

  4. Tracking, $b$-Tagging and Measurement of the $b$-Jet Production Cross Section with the ATLAS Detector

    CERN Document Server

    Fleckner, Johanna Elisabeth; Tapprogge, S

    2011-01-01

    The Standard Model of elementary particle physics was developed to describe the fundamental particles which constitute matter and the interactions between them. The Large Hadron Collider (LHC) at CERN in Geneva was built to solve some of the remaining open questions in the Standard Model and to explore physics beyond it, by colliding two proton beams at world-record centre-of-mass energies. The ATLAS experiment is designed to reconstruct particles and their decay products originating from these collisions. The precise reconstruction of particle trajectories plays an important role in the identification of particle jets which originate from bottom quarks (b-tagging). This thesis describes the step-wise commissioning of the ATLAS track reconstruction and b-tagging software and one of the first measurements of the b-jet production cross section in pp collisions at sqrt(s)=7 TeV with the ATLAS detector. The performance of the track reconstruction software was studied in great detail, first using data from cosmi...

  5. Production of configuration tables for the Input Mezzanine and Data Formatter components in the ATLAS Fast Tracker Trigger System

    CERN Document Server

    Poudroux, Jean-Michael

    2014-01-01

    The project revolve around developing configuration tables for two components in the Fast Tracker (FTK) trigger system used in the ATLAS trigger system. These components are Input Mezzanine cards and the Data Formatter. The tables give easy access to different ID's which identify which module the data is originating from and also which tower and what detector-region the data is being processed in.

  6. The Error Reporting in the ATLAS TDAQ system

    CERN Document Server

    Kolos, S; The ATLAS collaboration; Papaevgeniou, L

    2014-01-01

    The ATLAS Error Reporting feature, which is used in the TDAQ environment, provides a service that allows experts and shift crew to track and address errors relating to the data taking components and applications. This service, called the Error Reporting Service(ERS), gives software applications the opportunity to collect and send comprehensive data about errors, happening at run-time, to a place where it can be intercepted in real-time by any other system component. Other ATLAS online control and monitoring tools use the Error Reporting service as one of their main inputs to address system problems in a timely manner and to improve the quality of acquired data. The actual destination of the error messages depends solely on the run-time environment, in which the online applications are operating. When applications send information to ERS, depending on the actual configuration the information may end up in a local file, in a database, in distributed middle-ware, which can transport it to an expert system or dis...

  7. The Error Reporting in the ATLAS TDAQ System

    CERN Document Server

    Kolos, S; The ATLAS collaboration; Papaevgeniou, L

    2015-01-01

    The ATLAS Error Reporting feature, which is used in the TDAQ environment, provides a service that allows experts and shift crew to track and address errors relating to the data taking components and applications. This service, called the Error Reporting Service(ERS), gives software applications the opportunity to collect and send comprehensive data about errors, happening at run-time, to a place where it can be intercepted in real-time by any other system component. Other ATLAS online control and monitoring tools use the Error Reporting service as one of their main inputs to address system problems in a timely manner and to improve the quality of acquired data. The actual destination of the error messages depends solely on the run-time environment, in which the online applications are operating. When applications send information to ERS, depending on the actual configuration the information may end up in a local file, in a database, in distributed middle-ware, which can transport it to an expert system or dis...

  8. Analysis of empty ATLAS pilot jobs

    CERN Document Server

    Love, Peter; The ATLAS collaboration

    2016-01-01

    The pilot model used by the ATLAS production system has been in use for many years. The model has proven to be a success with many advantages over push models. However one of the negative side-effects of using a pilot model is the presence of 'empty pilots' running on sites which consume a small amount of walltime and not running a useful payload job. The impact on a site can be significant with previous studies showing a total 0.5% walltime usage with no benefit to either the site or to ATLAS. Another impact is the number of empty pilots being processed by a site's Compute Element and batch system which can be 5% of the total number of pilots being handled. In this paper we review the latest statistics using both ATLAS and site data and highlight edge cases where the number of empty pilots dominate. We also study the effect of tuning the pilot factories to reduce the number of empty pilots.

  9. Using containers with ATLAS offline software

    CERN Document Server

    Vogel, Marcelo; The ATLAS collaboration

    2017-01-01

    This paper describes the deployment of ATLAS offline software in containers for software development. For this we are using Docker, which is a lightweight virtualization technology that encapsulates a piece of software inside a complete file system. The deployment of offline releases via containers removes the strict requirement of compatibility between the runtime environment needed for job execution and the configuration of worker nodes at computing sites. If these two are decoupled from each other, sites can upgrade their nodes whenever and however they see fit. In this work, ATLAS software is distributed in containers either via the CernVM File System (CVMFS) or by means of a full ATLAS offline release installation. In software development, separating the build and runtime environment from the development environment allows users to take advantage of many modern code development tools that may not be available in production runtime setups like SLC6. It also frees developers from depending on resources lik...

  10. Experimental limits from ATLAS on Standard Model Higgs production.

    CERN Multimedia

    ATLAS, collaboration

    2012-01-01

    Experimental limits from ATLAS on Standard Model Higgs production in the mass range 110-600 GeV. The solid curve reflects the observed experimental limits for the production of a Higgs of each possible mass value (horizontal axis). The region for which the solid curve dips below the horizontal line at the value of 1 is excluded with a 95% confidence level (CL). The dashed curve shows the expected limit in the absence of the Higgs boson, based on simulations. The green and yellow bands correspond (respectively) to 68%, and 95% confidence level regions from the expected limits. Higgs masses in the narrow range 123-130 GeV are the only masses not excluded at 95% CL

  11. Study of an on-line filtering system for the ATLAS detector

    International Nuclear Information System (INIS)

    Fede, E.

    2001-01-01

    The first chapter presents today's knowledge about particle physics and a description of the main decay channels and physical signatures associated to the Higgs boson is given. The second chapter is dedicated to the LHC accelerator with a focus on the ATLAS detector and its sub-detectors. The third chapter presents ATLAS triggering system and its data acquisition system. In the fourth chapter the functionalities required for an adequate event filtering system concerning physics issues and data managing are described. The design of a prototype based on a fleet of PC computers linked through an Ethernet network is presented in the fifth chapter

  12. Searches for electroweak production of supersymmetric gauginos and sleptons with the ATLAS detector

    CERN Document Server

    Kourkoumeli-Charalampidi, Athina; The ATLAS collaboration

    2017-01-01

    The latest results of the electroweak production of Supersymmetric particles is presented. The searches are based on the integrated luminosity of 36.1 fb^{-1} of pp collisions collected at \\sqrt{s} = 13 TeV by the ATLAS experiment at the LHC.

  13. Experience commissioning the ATLAS distributed data management system on top of the WLCG service

    International Nuclear Information System (INIS)

    Campana, S

    2010-01-01

    The ATLAS experiment at CERN developed an automated system for distribution of simulated and detector data. Such system, which partially consists of various ATLAS specific services, strongly relies on the WLCG infrastructure, both at the level of middleware components, service deployment and operations. Because of the complexity of the system and its highly distributed nature, a dedicated effort was put in place to deliver a reliable service for ATLAS data distribution, offering the necessary performance, high availability and accommodating the main use cases. This contribution will describe the various challenges and activities carried on in 2008 for the commissioning of the system, together with the experience distributing simulated data and detector data. The main commissioning activity was concentrated in two Combined Computing Resource Challenges, in February and May 2008, where it was demonstrated that the WLCG service and the ATLAS system could sustain the peak load of data transfer according to the computing model, for several days in a row, concurrently with other LHC experiment activities. This dedicated effort led to the consequential improvements of ATLAS and WLCG services and to daily operation activities throughout the last year. The system has been delivering to WLCG tiers many hundreds of terabytes of simulated data and, since the summer of 2008, more than two petabytes of cosmic and beam data.

  14. ATLAS Facility Description Report

    International Nuclear Information System (INIS)

    Kang, Kyoung Ho; Moon, Sang Ki; Park, Hyun Sik; Cho, Seok; Choi, Ki Yong

    2009-04-01

    A thermal-hydraulic integral effect test facility, ATLAS (Advanced Thermal-hydraulic Test Loop for Accident Simulation), has been constructed at KAERI (Korea Atomic Energy Research Institute). The ATLAS has the same two-loop features as the APR1400 and is designed according to the well-known scaling method suggested by Ishii and Kataoka to simulate the various test scenarios as realistically as possible. It is a half-height and 1/288-volume scaled test facility with respect to the APR1400. The fluid system of the ATLAS consists of a primary system, a secondary system, a safety injection system, a break simulating system, a containment simulating system, and auxiliary systems. The primary system includes a reactor vessel, two hot legs, four cold legs, a pressurizer, four reactor coolant pumps, and two steam generators. The secondary system of the ATLAS is simplified to be of a circulating loop-type. Most of the safety injection features of the APR1400 and the OPR1000 are incorporated into the safety injection system of the ATLAS. In the ATLAS test facility, about 1300 instrumentations are installed to precisely investigate the thermal-hydraulic behavior in simulation of the various test scenarios. This report describes the scaling methodology, the geometric data of the individual component, and the specification and the location of the instrumentations in detail

  15. The ATLAS Trigger System: Ready for Run-2

    CERN Document Server

    Maeda, Junpei; The ATLAS collaboration

    2015-01-01

    The ATLAS trigger has been successfully collecting collision data during the first run of the LHC between 2009-2013 at a centre-of-mass energy between 900 GeV and 8 TeV. The trigger system consists of a hardware Level-1 and a software based high-level trigger that reduces the event rate from the design bunch-crossing rate of 40 MHz to an average recording rate of a few hundred Hz. During the data-taking period of Run-2 the LHC will operate at a centre-of-mass energy of about 13 TeV resulting in roughly five times higher trigger rates. In these proceedings, we briefly review the ATLAS trigger system upgrades that were implemented during the shutdown, allowing us to cope with the increased trigger rates while maintaining or even improving our efficiency to select relevant physics processes. This includes changes to the Level-1 calorimeter and muon trigger system, the introduction of a new Level-1 topological trigger module and themerging of the previously two-level higher-level trigger system into a single even...

  16. The ATLAS Trigger System : Ready for Run-2

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00211007; The ATLAS collaboration

    2016-01-01

    The ATLAS trigger has been successfully collecting collision data during the first run of the LHC between 2009-2013 at a centre-of-mass energy between 900 GeV and 8 TeV. The trigger system consists of a hardware based Level-1 (L1) and a software based high-level trigger (HLT) that reduces the event rate from the design bunch-crossing rate of 40 MHz to an average recording rate of a few hundred Hz. During the course of the ongoing Run-2 data-taking campaign at 13 TeV centre-of-mass energy the trigger rates will be approximately 5 times higher compared to Run-1. In these proceedings we briefly review the ATLAS trigger system upgrades that were implemented during the shutdown, allowing us to cope with the increased trigger rates while maintaining or even improving our efficiency to select relevant physics processes. This includes changes to the L1 calorimeter and muon trigger system, the introduction of a new L1 topological trigger subsystem and the merging of the previously two-level HLT system into a single ev...

  17. The ATLAS liquid argon calorimeter high-voltage system: commissioning, optimisation, and LHC relative luminosity measurement.

    CERN Document Server

    Arfaoui, Samir; Monnier, E

    2011-01-01

    The main goals of the ATLAS scientific programme are the observation or exclusion of physics beyond the Standard Model (SM), as well as the measurement of production cross-sections of SM processes. In oder to do so,it is important to measure the luminosity at the interaction point with great precision. The ATLAS luminosity is extracted using several detectors with varying efficiencies and acceptances. Different methods, such as inclusive - or coincidence - event counting and calorimeter integrated current measurements, are calibrated and cross-compared to provide the most accurate luminosity determination. In order to provide more cross-checks and a better control on the systematic uncertainties, an independent measurement using the liquid argon (LAr) forward calorimeter (FCal), based on the readout current of its high-voltage system, has been developed. This document describes how the LAr calorimeter high-voltage system has been installed and commissioned, as well as its application to a relative luminosity ...

  18. The ATLAS liquid argon calorimeter high-voltage system: commissioning, optimisation and LHC relative luminosity measurement

    International Nuclear Information System (INIS)

    Arfaoui, S.

    2011-10-01

    The main goals of the ATLAS scientific programme are the observation or exclusion of physics beyond the Standard Model (SM), as well as the measurement of production cross-sections of SM processes. In order to do so, it is important to measure the luminosity at the interaction point with great precision. The ATLAS luminosity is extracted using several detectors with varying efficiencies and acceptances. Different methods, such as inclusive - or coincidence - event counting and calorimeter integrated current measurements, are calibrated and cross-compared to provide the most accurate luminosity determination. In order to provide more cross-checks and a better control on the systematic uncertainties, an independent measurement using the liquid argon (LAr) forward calorimeter (FCal), based on the readout current of its high-voltage system, has been developed. This document describes how the LAr calorimeter high-voltage system has been installed and commissioned, as well as its application to a relative luminosity determination. (author)

  19. ATLAS database application enhancements using Oracle 11g

    International Nuclear Information System (INIS)

    Dimitrov, G; Canali, L; Blaszczyk, M; Sorokoletov, R

    2012-01-01

    The ATLAS experiment at LHC relies on databases for detector online data-taking, storage and retrieval of configurations, calibrations and alignments, post data-taking analysis, file management over the grid, job submission and management, condition data replication to remote sites. Oracle Relational Database Management System (RDBMS) has been addressing the ATLAS database requirements to a great extent for many years. Ten database clusters are currently deployed for the needs of the different applications, divided in production, integration and standby databases. The data volume, complexity and demands from the users are increasing steadily with time. Nowadays more than 20 TB of data are stored in the ATLAS production Oracle databases at CERN (not including the index overhead), but the most impressive number is the hosted 260 database schemes (for the most common case each schema is related to a dedicated client application with its own requirements). At the beginning of 2012 all ATLAS databases at CERN have been upgraded to the newest Oracle version at the time: Oracle 11g Release 2. Oracle 11g come with several key improvements compared to previous database engine versions. In this work we present our evaluation of the most relevant new features of Oracle 11g of interest for ATLAS applications and use cases. Notably we report on the performance and scalability enhancements obtained in production since the Oracle 11g deployment during Q1 2012 and we outline plans for future work in this area.

  20. ATLAS Distributed Computing experience and performance during the LHC Run-2

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00081160; The ATLAS collaboration

    2017-01-01

    ATLAS Distributed Computing during LHC Run-1 was challenged by steadily increasing computing, storage and network requirements. In addition, the complexity of processing task workflows and their associated data management requirements led to a new paradigm in the ATLAS computing model for Run-2, accompanied by extensive evolution and redesign of the workflow and data management systems. The new systems were put into production at the end of 2014, and gained robustness and maturity during 2015 data taking. ProdSys2, the new request and task interface; JEDI, the dynamic job execution engine developed as an extension to PanDA; and Rucio, the new data management system, form the core of Run-2 ATLAS distributed computing engine. One of the big changes for Run-2 was the adoption of the Derivation Framework, which moves the chaotic CPU and data intensive part of the user analysis into the centrally organized train production, delivering derived AOD datasets to user groups for final analysis. The effectiveness of the...

  1. ATLAS Distributed Computing experience and performance during the LHC Run-2

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00081160; The ATLAS collaboration

    2016-01-01

    ATLAS Distributed Computing during LHC Run-1 was challenged by steadily increasing computing, storage and network requirements. In addition, the complexity of processing task workflows and their associated data management requirements led to a new paradigm in the ATLAS computing model for Run-2, accompanied by extensive evolution and redesign of the workflow and data management systems. The new systems were put into production at the end of 2014, and gained robustness and maturity during 2015 data taking. ProdSys2, the new request and task interface; JEDI, the dynamic job execution engine developed as an extension to PanDA; and Rucio, the new data management system, form the core of the Run-2 ATLAS distributed computing engine. One of the big changes for Run-2 was the adoption of the Derivation Framework, which moves the chaotic CPU and data intensive part of the user analysis into the centrally organized train production, delivering derived AOD datasets to user groups for final analysis. The effectiveness of...

  2. First Results from the Online Radiation Dose Monitoring System in ATLAS experiment

    CERN Document Server

    Mandić, I; The ATLAS collaboration; Deliyergiyev, M; Gorišek, A; Kramberger, G; Mikuž, M; Franz, S; Hartert, J; Dawson, I; Miyagawa, P S; Nicolas, L

    2011-01-01

    High radiation doses which will accumulate in components of ATLAS experiment during data taking will cause damage to detectors and readout electronics. It is therefore important to continuously monitor the doses to estimate the level of degradation caused by radiation. Online radiation monitoring system measures ionizing dose in SiO2 and fluences of 1-MeV(Si) equivalent neutrons and thermal neutrons at several locations in ATLAS detector. In this paper measurements collected during two years of ATLAS data taking are presented and compared to predictions from radiation background simulations.

  3. ATLAS rewards Russian supplier for scintillating tile production

    CERN Multimedia

    Patrice Loïez

    2001-01-01

    The ATLAS collaboration has awarded Russian firm SIA Luch from Podolsk in the Moscow region an ATLAS Supplier Award. This follows delivery by the company of the final batch of scintillating tiles for the collaboration's tile calorimeter some six months ahead of schedule. Representatives of the firm are seen here receiving the award at a ceremony held in the collaboration's tile calorimeter instrumentation plant at CERN on 30 July. In front of one tile calorimeter module instrumented by scintillating tiles are (left to right) IHEP physicists Evgueni Startchenko and Andrei Karioukhine, Luch Podolsk representatives Igor Karetnikov and Yuri Zaitsev, tile calorimeter project leader Rupert Leitner, ATLAS spokesperson Peter Jenni, and CERN tile calorimeter group leader Ana Henriques-Correia.

  4. Report on container technology for the ATLAS TDAQ system

    CERN Document Server

    Gadirov, Hamid

    2016-01-01

    My summer student project "Container technology for the Upgrade of the ATLAS Trigger and Data Acquisition (TDAQ) system" focused on the research of container-based (operating system-level) virtualization for TDAQ software. Several tests were performed on Docker platform, all of them showed compatibility for TDAQ software.

  5. Recent Results on Quarkonium and Beauty Production in pp collisions with the ATLAS Detector

    CERN Document Server

    Walder, James; The ATLAS collaboration

    2017-01-01

    The latest results from ATLAS on heavy flavour and quarkonium production, including exotic states, are presented. This talk includes the measurement of B-hadron pair production, presented as a function of a variety kinematic variables between the two B-hadrons, providing important inputs to modelling of production via gluon splitting. Additional Insight into QCD models of quarkonium production and double parton scattering is also presented, through the production cross-section measurement of di-Jpsi and effective cross-section from double parton scattering.

  6. Measurement of heavy flavor production and azimuthal anisotropy in small and large systems with ATLAS

    CERN Document Server

    Hu, Qipeng; The ATLAS collaboration

    2018-01-01

    Measurements of heavy flavor hadron production and their correlations in A+A collisions provide insight into the energy loss mechanism and transport properties of heavy quarks in the QGP. From this perspective, the same measurements in $p$+A collisions thus serve as a necessary baseline for understanding the observations in A+A collisions. Additionally, detailed studies of the azimuthal anisotropy for heavy flavor hadrons in $p$+A may help address the question of whether the observed long-range ``ridge'' correlation arises from hard or semi-hard processes, or if it is the result of mechanisms unrelated to the initial hardness scale. This talk presents ATLAS measurements of heavy flavor production, via their semi-leptonic decay to muons in $\\sqrt{s_\\mathrm{NN}} = 2.76$ TeV Pb+Pb and $pp$ collisions, and via identified prompt $D$ mesons in $\\sqrt{s_\\mathrm{NN}} = 8.16$ TeV $p$+Pb collisions. Heavy flavor muon yields are observed to be strongly suppressed in Pb+Pb collisions compared to that in $pp$ collisions. ...

  7. The ATLAS Trigger system upgrade and performance in Run 2

    CERN Document Server

    Shaw, Savanna Marie; The ATLAS collaboration

    2017-01-01

    The ATLAS trigger has been used very successfully for the online event selection during the first part of the LHC Run-2 in 2015/16 at a centre-of-mass energy of 13 TeV. The trigger system is composed of a hardware Level-1 trigger and a software-based high-level trigger; it reduces the event rate from the bunch-crossing rate of 40 MHz to an average recording rate of about 1 kHz. The excellent performance of the ATLAS trigger has been vital for the ATLAS physics program of Run-2, selecting interesting collision events for wide variety of physics signatures with high efficiency. The trigger selection capabilities of ATLAS during Run-2 have been significantly improved compared to Run-1, in order to cope with the higher event rates and pile-up which are the result of the almost doubling of the center-of-mass collision energy and the increase in the instantaneous luminosity of the LHC. In order to prepare for the anticipated further luminosity increase of the LHC in 2017/18, improving the trigger performance remain...

  8. The ATLAS Trigger system upgrade and performance in Run 2

    CERN Document Server

    Shaw, Savanna Marie; The ATLAS collaboration

    2018-01-01

    The ATLAS trigger has been used very successfully for the online event selection during the first part of the second LHC run (Run-2) in 2015/16 at a centre-of-mass energy of 13 TeV. The trigger system is composed of a hardware Level-1 trigger and a software-based high-level trigger; it reduces the event rate from the bunch-crossing rate of 40 MHz to an average recording rate of about 1 kHz. The excellent performance of the ATLAS trigger has been vital for the ATLAS physics program of Run-2, selecting interesting collision events for wide variety of physics signatures with high efficiency. The trigger selection capabilities of ATLAS during Run-2 have been significantly improved compared to Run-1, in order to cope with the higher event rates and pile-up which are the result of the almost doubling of the center-of-mass collision energy and the increase in the instantaneous luminosity of the LHC. At the Level-1 trigger the undertaken improvements resulted in more pile-up robust selection efficiencies and event ra...

  9. Production and spectroscopy of quarkonia states at the ATLAS experiment

    CERN Document Server

    Price, D; The ATLAS collaboration

    2014-01-01

    This presentation will cover ATLAS results in the domain of charmonium production, including J/psi, psi(2s) and chi_c states. The analyses discussed include double-differential production cross-section measurements of the J/psi, psi(2S) and P-wave charmonium states chi_cJ, extending upon previous measurements in precision and kinematic reach. Prompt and non-prompt modes are distinguished, as well as J/psi vs psi(2s) and the contribution to J/psi production from chi_c feed-down. Alongside the latter analysis, a competitive measurement of the branching fraction B± --> chi_c1 + K± was also performed. Results of the measurements are compared with the latest theoretical predictions from a variety of theoretical approaches.

  10. Development and test of the DAQ system for a Micromegas prototype to be installed in the ATLAS experiment

    CERN Document Server

    Zibell, Andre; The ATLAS collaboration; Bianco, Michele; Martoiu, Victor Sorin

    2015-01-01

    A Micromegas (MM) quadruplet prototype with an active area of 0.5 m 2 that adopts the general design foreseen for the upgrade of the innermost forward muon tracking systems (Small Wheels) of the ATLAS detector in 2018-2019, has been built at CERN and is going to be tested in the ATLAS cavern environment during the LHC RUN-II period 2015-2017. The integration of this prototype detector into the ATLAS data acquisition system using custom ATCA equipment is presented. An ATLAS compatible Read Out Driver (ROD) based on the Scalable Readout System (SRS), the Scalable Readout Unit (SRU), will be used in order to transmit the data after generating valid event fragments to the high-level Read Out System (ROS). The SRU will be synchronized with the LHC bunch crossing clock (40.08 MHz) and will receive the Level-1 trigger signals from the Central Trigger Processor (CTP) through the TTCrx receiver ASIC. The configuration of the system will be driven directly from the ATLAS Run Control System. By using the ATLAS TDAQ Soft...

  11. System Description of the Electrical Power Supply System for the ATLAS Integral Test Loop

    International Nuclear Information System (INIS)

    Moon, S. K.; Park, J. K.; Kim, Y. S.; Song, C. H.; Baek, W. P.

    2007-02-01

    An integral effect test loop for pressurized water reactors (PWRs), the ATLAS (Advanced Thermal-hydraulic Test Loop for Accident Simulation), is constructed by Thermal-Hydraulics Safety Research Team in Korea Atomic Energy Research Institute (KAERI). The ATLAS facility has been designed to have the length scale of 1/2 and area scale of 1/144 compared with the reference plant, APR1400. This report describes the design and technical specifications of the electrical power supply system which supplies the electrical powers to core heater rods, other heaters, various pumps and other systems. The electrical power supply system had acquired the final approval on the operation from the Korea Electrical Safety Corporation. During performance tests for the operation and control, the electrical power supply system showed completely acceptable operation and control performance

  12. Role Based Access Control system in the ATLAS experiment

    CERN Document Server

    Valsan, M L; The ATLAS collaboration; Lehmann Miotto, G; Scannicchio, D A; Schlenker, S; Filimonov, V; Khomoutnikov, V; Dumitru, I; Zaytsev, A S; Korol, A A; Bogdantchikov, A; Caramarcu, C; Ballestrero, S; Darlea, G L; Twomey, M; Bujor, F; Avolio, G

    2011-01-01

    The complexity of the ATLAS experiment motivated the deployment of an integrated Access Control System in order to guarantee safe and optimal access for a large number of users to the various software and hardware resources. Such an integrated system was foreseen since the design of the infrastructure and is now central to the operations model. In order to cope with the ever growing needs of restricting access to all resources used within the experiment, the Roles Based Access Control (RBAC) previously developed has been extended and improved. The paper starts with a short presentation of the RBAC design, implementation and the changes made to the system to allow the management and usage of roles to control access to the vast and diverse set of resources. The paper continues with a detailed description of the integration across all areas of the system: local Linux and Windows nodes in the ATLAS Control Network (ATCN), the Linux application gateways offering remote access inside ATCN, the Windows Terminal Serv...

  13. Role Based Access Control System in the ATLAS Experiment

    CERN Document Server

    Valsan, M L; The ATLAS collaboration; Lehmann Miotto, G; Scannicchio, D A; Schlenker, S; Filimonov, V; Khomoutnikov, V; Dumitru, I; Zaytsev, A S; Korol, A A; Bogdantchikov, A; Avolio, G; Caramarcu, C; Ballestrero, S; Darlea, G L; Twomey, M; Bujor, F

    2010-01-01

    The complexity of the ATLAS experiment motivated the deployment of an integrated Access Control System in order to guarantee safe and optimal access for a large number of users to the various software and hardware resources. Such an integrated system was foreseen since the design of the infrastructure and is now central to the operations model. In order to cope with the ever growing needs of restricting access to all resources used within the experiment, the Roles Based Access Control (RBAC) previously developed has been extended and improved. The paper starts with a short presentation of the RBAC design, implementation and the changes made to the system to allow the management and usage of roles to control access to the vast and diverse set of resources. The paper continues with a detailed description of the integration across all areas of the system: local Linux and Windows nodes in the ATLAS Control Network (ATCN), the Linux application gateways offering remote access inside ATCN, the Windows Terminal Serv...

  14. The ATLAS High Level Trigger Steering Framework and the Trigger Configuration System.

    CERN Document Server

    Perez Cavalcanti, Tiago; The ATLAS collaboration

    2011-01-01

    The ATLAS detector system installed in the Large Hadron Collider (LHC) at CERN is designed to study proton-proton and nucleus-nucleus collisions with a maximum centre of mass energy of 14 TeV at a bunch collision rate of 40MHz. In March 2010 the four LHC experiments saw the first proton-proton collisions at 7 TeV. Still within the year a collision rate of nearly 10 MHz is expected. At ATLAS, events of potential interest for ATLAS physics are selected by a three-level trigger system, with a final recording rate of about 200 Hz. The first level (L1) is implemented in custom hardware; the two levels of the high level trigger (HLT) are software triggers, running on large farms of standard computers and network devices. Within the ATLAS physics program more than 500 trigger signatures are defined. The HLT tests each signature on each L1-accepted event; the test outcome is recorded for later analysis. The HLT-Steering is responsible for this. It foremost ensures the independent test of each signature, guarantying u...

  15. Performance of the NorduGrid ARC and the Dulcinea Executor in ATLAS Data Challenge 2

    DEFF Research Database (Denmark)

    Kleist, Josva; Eerola, P; Ekelöf, T.

    2004-01-01

    This talk describes the various stages of ATLAS Data Challenge 2 (DC2) in what concerns usage of resources deployed via NorduGrid's Advanced Resource Connector (ARC). It also describes the integration of these resources with the ATLAS production system using the Dulcinea executor. ATLAS Data...... Challenge 2 (DC2), run in 2004, was designed to be a step forward in the distributed data processing. In particular, much coordination of task assignment to resources was planned to be delegated to Grid in its different flavours. An automatic production management system was designed, to direct the tasks...... participation in ATLAS DC2. This was the first attempt to harness large amounts of strongly heterogeneous resources in various countries for a single collaborative exercise using Grid tools. This talk addresses various issues that arose during different stages of DC2 in this environment: preparation...

  16. Atlas Pulsed Power System: a Driver for Multi-Megagauss Fields

    International Nuclear Information System (INIS)

    Cochrane, J.C.; Bartsch, R.R.; Bennett, G.A.; Bowman, D.W.; Davis, H.A.; Ekdahl, C.A.; Gribble, R.F.; Kimerly, H.J.; Nielsen, K.E.; Parsons, W.M.; Paul, J.D.; Scudder, D.W.; Trainor, R.J.; Thompson, M.C.; Watt, R.G.

    1998-01-01

    Atlas is a pulsed power machine designed for hydrodynamic experiments for the Los Alamos High Energy Density Physics Experimental program. It is presently under construction and should be operational in late 2000. Atlas will store 23 MJ at an erected voltage of 240 kV. This will produce a current of 30 MA into a static load and as much as 32 MA into a dynamic load. The current pulse will have a rise time of approximately5micros and will produce a magnetic field driving the impactor liner of several hundred Tesla at the target radius of one to two centimeters. The collision can produce shock pressures of approximately15 megabars. Design of the pulsed power system will be presented along with data obtained from the Atlas prototype Marx module

  17. Grid site testing for ATLAS with HammerCloud

    International Nuclear Information System (INIS)

    Elmsheuser, J; Hönig, F; Legger, F; LLamas, R Medrano; Sciacca, F G; Ster, D van der

    2014-01-01

    With the exponential growth of LHC (Large Hadron Collider) data in 2012, distributed computing has become the established way to analyze collider data. The ATLAS grid infrastructure includes more than 130 sites worldwide, ranging from large national computing centers to smaller university clusters. HammerCloud was previously introduced with the goals of enabling virtual organisations (VO) and site-administrators to run validation tests of the site and software infrastructure in an automated or on-demand manner. The HammerCloud infrastructure has been constantly improved to support the addition of new test workflows. These new workflows comprise e.g. tests of the ATLAS nightly build system, ATLAS Monte Carlo production system, XRootD federation (FAX) and new site stress test workflows. We report on the development, optimization and results of the various components in the HammerCloud framework.

  18. Grid Site Testing for ATLAS with HammerCloud

    CERN Document Server

    Elmsheuser, J; The ATLAS collaboration; Legger, F; Medrano LLamas, R; Sciacca, G; van der Ster, D

    2014-01-01

    With the exponential growth of LHC (Large Hadron Collider) data in 2012, distributed computing has become the established way to analyze collider data. The ATLAS grid infrastructure includes more than 130 sites worldwide, ranging from large national computing centers to smaller university clusters. HammerCloud was previously introduced with the goals of enabling VO- and site-administrators to run validation tests of the site and software infrastructure in an automated or on-demand manner. The HammerCloud infrastructure has been constantly improved to support the addition of new test work-flows. These new work-flows comprise e.g. tests of the ATLAS nightly build system, ATLAS MC production system, XRootD federation FAX and new site stress test work-flows. We report on the development, optimization and results of the various components in the HammerCloud framework.

  19. Grid Site Testing for ATLAS with HammerCloud

    CERN Document Server

    Elmsheuser, J; The ATLAS collaboration; Legger, F; Medrano LLamas, R; Sciacca, G; van der Ster, D

    2013-01-01

    With the exponential growth of LHC (Large Hadron Collider) data in 2012, distributed computing has become the established way to analyze collider data. The ATLAS grid infrastructure includes more than 130 sites worldwide, ranging from large national computing centers to smaller university clusters. HammerCloud was previously introduced with the goals of enabling VO- and site-administrators to run validation tests of the site and software infrastructure in an automated or on-demand manner. The HammerCloud infrastructure has been constantly improved to support the addition of new test work-flows. These new work-flows comprise e.g. tests of the ATLAS nightly build system, ATLAS MC production system, XRootD federation FAX and new site stress test work-flows. We report on the development, optimization and results of the various components in the HammerCloud framework.

  20. Production of exotic and conventional quarkonia and open beauty/open charm at ATLAS

    CERN Document Server

    C. Bini; The ATLAS collaboration

    2016-01-01

    The ATLAS experiment at LHC is carrying on a wide programme to study the production properties of conventional and exotic quarkonium, beauty, and charm bound states. The latest results on J/$\\psi$, $\\psi$(2s) and X(3872) production at 7, 8, and 13 TeV, together with D meson production with Run-1 are presented. Studies of associated production of charmonium with vector bosons, searches for exotic states in the bottomonium sector and a new measurement of the ratio of b-quark fragmentation functions are also briefly presented.

  1. ATLAS' major cooling project

    CERN Multimedia

    2005-01-01

    In 2005, a considerable effort has been put into commissioning the various units of ATLAS' complex cryogenic system. This is in preparation for the imminent cooling of some of the largest components of the detector in their final underground configuration. The liquid helium and nitrogen ATLAS refrigerators in USA 15. Cryogenics plays a vital role in operating massive detectors such as ATLAS. In many ways the liquefied argon, nitrogen and helium are the life-blood of the detector. ATLAS could not function without cryogens that will be constantly pumped via proximity systems to the superconducting magnets and subdetectors. In recent weeks compressors at the surface and underground refrigerators, dewars, pumps, linkages and all manner of other components related to the cryogenic system have been tested and commissioned. Fifty metres underground The helium and nitrogen refrigerators, installed inside the service cavern, are an important part of the ATLAS cryogenic system. Two independent helium refrigerators ...

  2. Large Scale Software Building with CMake in ATLAS

    CERN Document Server

    Elmsheuser, Johannes; The ATLAS collaboration; Obreshkov, Emil; Undrus, Alexander

    2016-01-01

    The offline software of the ATLAS experiment at the LHC (Large Hadron Collider) serves as the platform for detector data reconstruction, simulation and analysis. It is also used in the detector trigger system to select LHC collision events during data taking. ATLAS offline software consists of several million lines of C++ and Python code organized in a modular design of more than 2000 specialized packages. Because of different workflows many stable numbered releases are in parallel production use. To accommodate specific workflow requests, software patches with modified libraries are distributed on top of existing software releases on a daily basis. The different ATLAS software applications require a flexible build system that strongly supports unit and integration tests. Within the last year this build system was migrated to CMake. A CMake configuration has been developed that allows one to easily set up and build the mentioned software packages. This also makes it possible to develop and test new and modifi...

  3. Large scale software building with CMake in ATLAS

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00218447; The ATLAS collaboration; Elmsheuser, Johannes; Obreshkov, Emil; Undrus, Alexander

    2017-01-01

    The offline software of the ATLAS experiment at the LHC (Large Hadron Collider) serves as the platform for detector data reconstruction, simulation and analysis. It is also used in the detector trigger system to select LHC collision events during data taking. ATLAS offline software consists of several million lines of C++ and Python code organized in a modular design of more than 2000 specialized packages. Because of different workflows many stable numbered releases are in parallel production use. To accommodate specific workflow requests, software patches with modified libraries are distributed on top of existing software releases on a daily basis. The different ATLAS software applications require a flexible build system that strongly supports unit and integration tests. Within the last year this build system was migrated to CMake. A CMake configuration has been developed that allows one to easily set up and build the mentioned software packages. This also makes it possible to develop and test new and modifi...

  4. ATLAS B-physics potential

    International Nuclear Information System (INIS)

    Smizanska, M.

    2001-01-01

    Studies since 1993 have demonstrated the ability of ATLAS to pursue a wide B physics program. This document presents the latest performance studies with special stress on lepton identification. B-decays containing several leptons in ATLAS statistically dominate the high-precision measurements. We present new results on physics simulations of CP violation measurements in the B s 0 → J/Ψphi decay and on a novel ATLAS programme on beauty production in central proton-proton collisions of LHC

  5. ATLAS EventIndex General Dataflow and Monitoring Infrastructure

    CERN Document Server

    Barberis, Dario; The ATLAS collaboration

    2016-01-01

    The ATLAS EventIndex has been running in production since mid-2015, reliably collecting information worldwide about all produced events and storing them in a central Hadoop infrastructure at CERN. A subset of this information is copied to an Oracle relational database for fast datasets discovery, event-picking, crosschecks with other ATLAS systems and checks for event duplication. The system design and its optimization is serving event picking from requests of a few events up to scales of tens of thousand of events, and in addition, data consistency checks are performed for large production campaigns. Detecting duplicate events with a scope of physics collections has recently arisen as an important use case. This paper describes the general architecture of the project and the data flow and operation issues, which are addressed by recent developments to improve the throughput of the overall system. In this direction, the data collection system is reducing the usage of the messaging infrastructure to overcome t...

  6. ATLAS EventIndex general dataflow and monitoring infrastructure

    CERN Document Server

    AUTHOR|(SzGeCERN)638886; The ATLAS collaboration; Barberis, Dario; Favareto, Andrea; Garcia Montoro, Carlos; Gonzalez de la Hoz, Santiago; Hrivnac, Julius; Prokoshin, Fedor; Salt, Jose; Sanchez, Javier; Toebbicke, Rainer; Yuan, Ruijun

    2017-01-01

    The ATLAS EventIndex has been running in production since mid-2015, reliably collecting information worldwide about all produced events and storing them in a central Hadoop infrastructure at CERN. A subset of this information is copied to an Oracle relational database for fast dataset discovery, event-picking, crosschecks with other ATLAS systems and checks for event duplication. The system design and its optimization is serving event picking from requests of a few events up to scales of tens of thousand of events, and in addition, data consistency checks are performed for large production campaigns. Detecting duplicate events with a scope of physics collections has recently arisen as an important use case. This paper describes the general architecture of the project and the data flow and operation issues, which are addressed by recent developments to improve the throughput of the overall system. In this direction, the data collection system is reducing the usage of the messaging infrastructure to overcome th...

  7. Development of mass-production technique of the ATLAS thin gap chamber in Japan

    CERN Document Server

    Tanaka, S; Ishii, K; Iwasaki, H; Arataki, Y; Bando, T; Homma, Y; Ishino, M; Kondo, T; Kobayashi, T; Kurashige, H; Mikenberg, G; Miyazaki, Y; Nakagawa, Y; Nanjo, H; Ikeno, M; Nozaki, M; Ochi, A; Sasaki, O; Shoa, M; Sugimoto, T; Takeda, H; Takeshita, T; Yokoyama, C

    2004-01-01

    The thin gap chambers (TGCs) are used for the muon trigger system in the end-cap regions of the ATLAS detector. As the anode-cathode distance is small, 1.4 mm, chamber flatness is essential to achieve a uniform gas gain over the chamber. The TGC mass production phase at High Energy Accelerator Research Organization (KEK) started in January 2001. In order to perform a stable production with high quality we developed a chamber closing system. When we glue two half- chambers together, we sandwich them with a granite table and an aluminum honeycomb panel that keep the chamber flat from both sides. By using silk screens, we also control the quantity of epoxy adhesive that affects the chamber thickness. Owing to these developments, we can achieve the flatness of less than 100 mu m. Uniformity of detection efficiency of the TGC is measured with a cosmic-ray test bench at Kobe University. We have tested 300 TGCs so far. Position dependence of the efficiency is measured with a granularity of 5 mm- by-5 mm. The average...

  8. Atlas pixel opto-board production and analysis and optolink simulation studies

    International Nuclear Information System (INIS)

    Nderitu, Simon Kirichu

    2007-01-01

    At CERN, a Large collider will collide protons at high energies. There are four experiments being built to study the particle properties from the collision. The ATLAS experiment is the largest. It has many sub detectors among which is the Pixel detector which is the innermost part. The Pixel detector has eighty million channels that have to be read out. An optical link is utilized for the read out. It has optical to electronic interfaces both on the detector and off the detector at the counting room. The component on the detector in called the opto-board. This work discusses the production testing of the opto-boards to be installed on the detector. A total of 300 opto-boards including spares have been produced. The production was done in three laboratories among which is the laboratory at the University of Wuppertal which had the responsibility of Post production testing of all the one third of the total opto-boards. The results are discussed in this work. The analysis of the results from the total production process has been done in the scope of this work as well. In addition to the production, a study by simulation of the communication links optical signal has been done. This has enabled an assessment of the sufficiency of the optical signal against the transmission attenuation and irradiation degradation. A System Test set up has been put up at Wuppertal to enhance general studies for better understanding of the Pixel read out system. Among other studies is the study of the timing parameters behavior of the System which has been done in this work and enhanced by a simulation. These parameters are namely the mark to space ratio and the fine delay and their relatedness during the optolink tuning. A bit error rate test based on the System has also been done which enabled assessment of the transmission quality utilizing the tools inbuilt in the System Test. These results have been presented in this work. (orig.)

  9. Danish heat atlas as a support tool for energy system models

    International Nuclear Information System (INIS)

    Petrovic, Stefan N.; Karlsson, Kenneth B.

    2014-01-01

    Highlights: • The GIS method for calculating costs of district heating expansion is presented. • High socio-economic potential for district heating is identified within urban areas. • The method for coupling a heat atlas and TIMES optimization model is proposed. • Presented methods can be used for any geographical region worldwide. - Abstract: In the past four decades following the global oil crisis in 1973, Denmark has implemented remarkable changes in its energy sector, mainly due to the energy conservation measures on the demand side and the energy efficiency improvements on the supply side. Nowadays, the capital intensive infrastructure investments, such as the expansion of district heating networks and the introduction of significant heat saving measures require highly detailed decision-support tool. A Danish heat atlas provides highly detailed database with extensive information about more than 2.5 million buildings in Denmark. Energy system analysis tools incorporate environmental, economic, energy and engineering analysis of future energy systems and are considered crucial for the quantitative assessment of transitional scenarios towards future milestones, such as EU 2020 goals and Denmark’s goal of achieving fossil free society after 2050. The present paper shows how a Danish heat atlas can be used for providing inputs to energy system models, especially related to the analysis of heat saving measures within building stock and expansion of district heating networks. As a result, marginal cost curves are created, approximated and prepared for the use in optimization energy system model. Moreover, it is concluded that heat atlas can contribute as a tool for data storage and visualisation of results

  10. Supporting ATLAS

    CERN Multimedia

    maximilien brice

    2003-01-01

    Eighteen feet made of stainless steel will support the barrel ATLAS detector in the cavern at Point 1. In total, the ATLAS feet system will carry approximately 6000 tons, and will give the same inclination to the detector as the LHC accelerator.

  11. The zero degree calorimeter for the ATLAS experiment

    International Nuclear Information System (INIS)

    Leite, Marco

    2009-01-01

    Full text. The Zero Degree Calorimeter (ZDC) of the ATLAS experiment at the LHC will measure neutral particles (photons and neutrons) produced at very forward directions in heavy ions and low luminosity p + p collisions. While its main application will be the determination of the centrality of the heavy ions collisions and trigger integration in ATLAS, the design of the ZDC also provides many other interesting heavy ion physics possibilities, like the measurements of the direct flow (by directly measuring the reaction plane formed by the spectator neutrons transverse momentum), ultra-peripheral quarkonia photo-production etc. During low luminosity p+p runs, the ZDC will give valuable information about forward neutron and neutral mesons cross-section production at the LHC energies. The ZDC will also be used in independent luminosity measurements during the early stages of the LHC operation, helping to achieve a better understanding of the standard ATLAS luminosity monitor system (LUCID). The ZDC comprises two sampling calorimeter modules, symmetrically located along the beam line and each one separated 140m from the ATLAS interaction point. This is the region where the accelerator neutral beam absorbers are installed, and the ZDC is strategically inserted inside a slot in these absorbers, extending the ATLAS pseudo-rapidity calorimeter coverage to |η| > 8. Each ZDC module is divided in 4 sections: one electromagnetic followed by three hadronic sections. Built using Tungsten absorber blocs interspersed by quartz fibers for the sampling of the shower, each one of these modules provides energy measurements of the incident particles. The electromagnetic and the first hadronic section can also perform position measurements perpendicular to the projected beam direction due to their segmentation. Instrumenting this realm presents several challenges due to the extremely high radiation levels. To account for the large energy dynamic range (14 bits equivalent), a combination

  12. The ATLAS Trigger System: Ready for Run II

    CERN Document Server

    Czodrowski, Patrick; The ATLAS collaboration

    2015-01-01

    The ATLAS trigger system has been used successfully for data collection in the 2009-2013 Run 1 operation cycle of the CERN Large Hadron Collider (LHC) at center-of-mass energies of up to 8 TeV. With the restart of the LHC for the new Run 2 data-taking period at 13 TeV, the trigger rates are expected to rise by approximately a factor of 5. The trigger system consists of a hardware-based first level (L1) and a software-based high-level trigger (HLT) that reduces the event rate from the design bunch-crossing rate of 40 MHz to an average recording rate of ~ 1kHz. This presentation will give an overview of the upgrades to the ATLAS trigger system that have been implemented during the LHC shutdown period in order to deal with the increased trigger rates while efficiently selecting the physics processes of interest. These upgrades include changes to the L1 calorimeter trigger, the introduction of a new L1 topological trigger module, improvements in the L1 muon system, and the merging of the previously two-level HLT ...

  13. The ATLAS Trigger System: Ready for Run-2

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00211007; The ATLAS collaboration

    2015-01-01

    The ATLAS trigger has been successfully collecting collision data during the first run of the LHC between 2009-2013 at a centre-of-mass energy between 900 GeV and 8 TeV. The trigger system consists of a hardware Level-1 (L1) and a software based high-level trigger (HLT) that reduces the event rate from the design bunch-crossing rate of 40 MHz to an average recording rate of a few hundred Hz. During the next data-taking period starting in 2015 (Run-2) the LHC will operate at a centre-of-mass energy of about 13 TeV resulting in roughly five times higher trigger rates. We will briefly review the ATLAS trigger system upgrades that were implemented during the shutdown, allowing us to cope with the increased trigger rates while maintaining or even improving our efficiency to select relevant physics processes. This includes changes to the L1 calorimeter and muon trigger system, the introduction of a new L1 topological trigger module and the merging of the previously two-level HLT system into a single event filter fa...

  14. Construction and Tests of Modules for the ATLAS Pixel Detector

    CERN Document Server

    AUTHOR|(CDS)2068490

    2003-01-01

    The ATLAS Pixel Detector is the innermost layer of the ATLAS tracking system and will contribute significantly to the ATLAS track and vertex reconstruction. The detector consists of identical sensor-chip-hybrid modules, arranged in three barrels in the centre and three disks on either side for the forward region. The position of the pixel detector near the interaction point requires excellent radiation hardness, mechanical and thermal robustness, good long-term stability, all combined with a low material budget. The pre-production phase of such pixel modules has nearly finished, yielding fully functional modules. Results are presented of tests with these modules.

  15. A thermosiphon for ATLAS

    CERN Multimedia

    Rosaria Marraffino

    2013-01-01

    A new thermosiphon cooling system, designed for the ATLAS silicon detectors by CERN’s EN-CV team in collaboration with the experiment, will replace the current system in the next LHC run in 2015. Using the basic properties of density difference and making gravity do the hard work, the thermosiphon promises to be a very reliable solution that will ensure the long-term stability of the whole system.   Former compressor-based cooling system of the ATLAS inner detectors. The system is currently being replaced by the innovative thermosiphon. (Photo courtesy of Olivier Crespo-Lopez). Reliability is the major issue for the present cooling system of the ATLAS silicon detectors. The system was designed 13 years ago using a compressor-based cooling cycle. “The current cooling system uses oil-free compressors to avoid fluid pollution in the delicate parts of the silicon detectors,” says Michele Battistin, EN-CV-PJ section leader and project leader of the ATLAS thermosiphon....

  16. ATLAS diamond Beam Condition Monitor

    CERN Document Server

    Gorišek, A; Dolenc, I; Frais-Kölbl, H; Griesmayer, E; Kagan, H; Korpar, S; Kramberger, G; Mandic, I; Meyer, M; Mikuz, M; Pernegger, H; Smith, S; Trischuk, W; Weilhammer, P; Zavrtanik, M

    2007-01-01

    The ATLAS experiment has chosen to use diamond for its Beam Condition Monitor (BCM) given its radiation hardness, low capacitance and short charge collection time. In addition, due to low leakage current diamonds do not require cooling. The ATLAS Beam Condition Monitoring system is based on single beam bunch crossing measurements rather than integrating the accumulated particle flux. Its fast electronics will allow separation of LHC collisions from background events such as beam gas interactions or beam accidents. There will be two stations placed symmetrically about the interaction point along the beam axis at . Timing of signals from the two stations will provide almost ideal separation of beam–beam interactions and background events. The ATLAS BCM module consists of diamond pad detectors of area and thickness coupled to a two-stage RF current amplifier. The production of the final detector modules is almost done. A S/N ratio of 10:1 has been achieved with minimum ionizing particles (MIPs) in the test bea...

  17. Recent ATLAS Articles on WLAP

    CERN Multimedia

    J. Herr

    As reported in the September 2004 ATLAS eNews, the Web Lecture Archive Project is a system for the archiving and publishing of multimedia presentations, using the Web as medium. We list here newly available WLAP items relating to ATLAS: Atlas Physics Workshop 6-11 June 2005 June 2005 ATLAS Week Plenary Session Click here to browse WLAP for all ATLAS lectures.

  18. Soft QCD at CMS and ATLAS

    CERN Document Server

    Starovoitov, Pavel; The ATLAS collaboration

    2018-01-01

    A short overview of the recent soft QCD results from the ATLAS and CMS collaborations is presented. The inelastic cross section measurement by CMS at 13 TeV is summarised. The contribution of the diffractive processes to the very forward photon spectra studied by ATLAS and LHCf is discussed. The ATLAS measurements of the exclusive two-photon production of the muon pairs is presented and compared to the previous ATLAS and CMS results.

  19. Searches for electroweak production of supersymmetric gauginos and sleptons with the ATLAS detector

    CERN Document Server

    Lorenz, Jeanette; The ATLAS collaboration

    2018-01-01

    Many supersymmetry models feature gauginos and also sleptons with masses less than a few hundred GeV. These can give rise to direct pair production rates at the LHC that can be observed in the data sample recorded by the ATLAS detector. The talk presents results from searches for gaugino and slepton pair production in final states with leptons, and were performed with pp collisions at a centre-of-mass energy of 13 TeV.

  20. Preparation of Northern Mid-Continent Petroleum Atlas

    Energy Technology Data Exchange (ETDEWEB)

    Lee C. Gerhard; Timothy R. Carr; W. Lynn Watney

    1998-05-01

    As proposed, the third year program will continue and expand upon the Kansas elements of the original program, and provide improved on-line access to the prototype atlas. The third year of the program will result in a digital atlas sufficient to provide a permanent improvement in data access to Kansas operators. The ultimate goal of providing an interactive history-matching interface with a regional database will be demonstrated as the program covers more geographic territory and the database expands. The atlas will expand to include significant reservoirs representing the major plays in Kansas, and North Dakota. Primary products of the third year prototype atlas will be on-line accessible digital databases and technical publications covering two additional petroleum plays in Kansas and one in North Dakota. Regional databases will be supplemented with geological field studies of selected fields in each play. Digital imagery, digital mapping, relational data queries, and geographical information systems will be integral to the field studies and regional data sets. Data sets will have relational links to provide opportunity for history-matching, feasibility, and risk analysis tests on contemplated exploration and development projects. The flexible "web-like" design of the atlas provides ready access to data, and technology at a variety of scales from regional, to field, to lease, and finally to the individual well bore. The digital structure of the atlas permits the operator to access comprehensive reservoir data and customize the interpretative products (e.g., maps and cross-sections) to their needs. The atlas will be accessible in digital form on-line using a World-Wide-Web browser as the graphical user interface. Regional data sets and field studies will be freestanding entities that will be made available on-line through the Internet to users as they are completed. Technology transfer activities will be ongoing from the earliest part of this project, providing

  1. Integration Of PanDA Workload Management System With Supercomputers for ATLAS and Data Intensive Science

    Energy Technology Data Exchange (ETDEWEB)

    De, K [University of Texas at Arlington; Jha, S [Rutgers University; Klimentov, A [Brookhaven National Laboratory (BNL); Maeno, T [Brookhaven National Laboratory (BNL); Nilsson, P [Brookhaven National Laboratory (BNL); Oleynik, D [University of Texas at Arlington; Panitkin, S [Brookhaven National Laboratory (BNL); Wells, Jack C [ORNL; Wenaus, T [Brookhaven National Laboratory (BNL)

    2016-01-01

    The Large Hadron Collider (LHC), operating at the international CERN Laboratory in Geneva, Switzerland, is leading Big Data driven scientific explorations. Experiments at the LHC explore the fundamental nature of matter and the basic forces that shape our universe, and were recently credited for the discovery of a Higgs boson. ATLAS, one of the largest collaborations ever assembled in the sciences, is at the forefront of research at the LHC. To address an unprecedented multi-petabyte data processing challenge, the ATLAS experiment is relying on a heterogeneous distributed computational infrastructure. The ATLAS experiment uses PanDA (Production and Data Analysis) Workload Management System for managing the workflow for all data processing on over 150 data centers. Through PanDA, ATLAS physicists see a single computing facility that enables rapid scientific breakthroughs for the experiment, even though the data centers are physically scattered all over the world. While PanDA currently uses more than 250,000 cores with a peak performance of 0.3 petaFLOPS, LHC data taking runs require more resources than Grid computing can possibly provide. To alleviate these challenges, LHC experiments are engaged in an ambitious program to expand the current computing model to include additional resources such as the opportunistic use of supercomputers. We will describe a project aimed at integration of PanDA WMS with supercomputers in United States, Europe and Russia (in particular with Titan supercomputer at Oak Ridge Leadership Computing Facility (OLCF), MIRA supercomputer at Argonne Leadership Computing Facilities (ALCF), Supercomputer at the National Research Center Kurchatov Institute , IT4 in Ostrava and others). Current approach utilizes modified PanDA pilot framework for job submission to the supercomputers batch queues and local data management, with light-weight MPI wrappers to run single threaded workloads in parallel on LCFs multi-core worker nodes. This implementation

  2. Energy Frontier Research With ATLAS: Final Report

    Energy Technology Data Exchange (ETDEWEB)

    Butler, John [Boston Univ., MA (United States); Black, Kevin [Boston Univ., MA (United States); Ahlen, Steve [Boston Univ., MA (United States)

    2016-06-14

    The Boston University (BU) group is playing key roles across the ATLAS experiment: in detector operations, the online trigger, the upgrade, computing, and physics analysis. Our team has been critical to the maintenance and operations of the muon system since its installation. During Run 1 we led the muon trigger group and that responsibility continues into Run 2. BU maintains and operates the ATLAS Northeast Tier 2 computing center. We are actively engaged in the analysis of ATLAS data from Run 1 and Run 2. Physics analyses we have contributed to include Standard Model measurements (W and Z cross sections, t\\bar{t} differential cross sections, WWW^* production), evidence for the Higgs decaying to \\tau^+\\tau^-, and searches for new phenomena (technicolor, Z' and W', vector-like quarks, dark matter).

  3. ATLAS Detector Control System Data Viewer

    CERN Document Server

    Tsarouchas, Charilaos; Roe, S; Bitenc, U; Fehling-Kaschek, ML; Winkelmann, S; D’Auria, S; Hoffmann, D; Pisano, O

    2011-01-01

    The ATLAS experiment at CERN is one of the four Large Hadron Collider experiments. DCS Data Viewer (DDV) is a web interface application that provides access to historical data of ATLAS Detector Control System [1] (DCS) parameters written to the database (DB). It has a modular andflexible design and is structured using a clientserver architecture. The server can be operated stand alone with a command-line interface to the data while the client offers a user friendly, browser independent interface. The selection of the metadata of DCS parameters is done via a column-tree view or with a powerful search engine. The final visualisation of the data is done using various plugins such as “value over time” charts, data tables, raw ASCII or structured export to ROOT. Excessive access or malicious use of the database is prevented by dedicated protection mechanisms, allowing the exposure of the tool to hundreds of inexperienced users. The metadata selection and data output features can be used separately by XML con...

  4. Top pair production in association with a vector gauge boson in ATLAS

    CERN Document Server

    Sjoelin, Joergen; The ATLAS collaboration

    2017-01-01

    An overview of the latest results for top pair production in association with a vector gauge boson in the ATLAS detector at LHC is presented. The results involving $Z$ and $W$ bosons are recorded at $\\sqrt{13}$ TeV collision energy, while the results involving photons are recorded at $\\sqrt{7}$ TeV and $\\sqrt{8}$ TeV collision energy.

  5. A Fast Hardware Tracker for the ATLAS Trigger System

    CERN Document Server

    Neubauer, Mark S

    2011-01-01

    In hadron collider experiments, triggering the detector to store interesting events for offline analysis is a challenge due to the high rates and multiplicities of particles produced. Maintaining high trigger efficiency for the physics we are most interested in while at the same time suppressing high rate physics from inclusive QCD processes is a difficult but important problem. It is essential that the trigger system be flexible and robust, with sufficient redundancy and operating margin. Providing high quality track reconstruction over the full ATLAS detector by the start of processing at LVL2 is an important element to achieve these needs. As the instantaneous luminosity increases, the computational load on the LVL2 system will significantly increase due to the need for more sophisticated algorithms to suppress backgrounds. The Fast Tracker (FTK) is a proposed upgrade to the ATLAS trigger system. It is designed to enable early rejection of background events and thus leave more LVL2 execution time by moving...

  6. Rucio, the next-generation Data Management system in ATLAS

    CERN Document Server

    Serfon, C; The ATLAS collaboration; Beermann, T; Garonne, V; Goossens, L; Lassnig, M; Nairz, A; Vigne, R

    2014-01-01

    Rucio is the next-generation of Distributed Data Management (DDM) system benefiting from recent advances in cloud and "Big Data" computing to address HEP experiments scaling requirements. Rucio is an evolution of the ATLAS DDM system Don Quijote 2 (DQ2), which has demonstrated very large scale data management capabilities with more than 160 petabytes spread worldwide across 130 sites, and accesses from 1,000 active users. However, DQ2 is reaching its limits in terms of scalability, requiring a large number of support staff to operate and being hard to extend with new technologies. Rucio addresses these issues by relying on new technologies to ensure system scalability, cover new user requirements and employ new automation framework to reduce operational overheads. In this talk, we will present the history of the DDM project and the experience of data management operation in ATLAS computing. Thus, We will show the key concepts of Rucio, including its data organization. The Rucio design, and the technology it e...

  7. Recent test results on the ATLAS SCT detector

    International Nuclear Information System (INIS)

    Pernegger, H.

    2003-01-01

    The ATLAS Semiconductor Tracker (SCT) will be a central part of the tracking system of the ATLAS experiment. The SCT, which is currently under construction, will consist of four concentric barrels of silicon detectors as well as two silicon endcap detectors formed by nine disks each. After an overview of the SCT and the detector module layout, the paper will summarize recent test results obtained from silicon detector modules, which have been extensively tested before starting their large series production. The tests presented here cover electrical performance of individual modules, their performance after irradiation, as well as system tests in a multi-module setup

  8. Measurement of Single-top Quark Production with the ATLAS Detector

    CERN Document Server

    Holzbauer, J L

    2011-01-01

    Single-top production has been studied with the ATLAS detector using 0.7 fb-1 of 2011 data recorded at 7 TeV center-of-mass energy. The measurement of electroweak production of top-quarks allows probes of the Wtb vertex and a direct measurement of the CKM matrix element |V_tb|. It is also expected to be sensitive to new physics such as flavor changing neutral currents or heavy W production. The t-channel cross-section measurements are performed using both a cut-based and neural network approach, while a cut-based selection in the dilepton channel is used to derive a limit on the associated (Wt) production. An observed cross-section of 90 +32 -22 pb (65 +28 -19 pb expected) is obtained for the t-channel, which is consistent with the Standard Model expectation. For the Wt production, an observed limit of < 39.1 pb (40.6 pb expected) is derived, which corresponds to about 2.5 times the Standard Model expectation.

  9. Measurement of photon production cross sections with the ATLAS detector

    CERN Document Server

    Turra, Ruggero; The ATLAS collaboration

    2017-01-01

    The production of prompt isolated photons at hadron colliders provides a stringent test of perturbative QCD and can be used to probe the proton structure. The ATLAS Collaboration has performed precise measurements of the inclusive production of isolated prompt photons at a centre-of-mass energy of 13 TeV, differential in both rapidity and photon transverse momentum. In addition, the integrated and differential cross sections for isolated photon pairs at 8 TeV have been measured. The results are compared with state-of-the-art theory predictions at NLO in QCD and with predictions of several MC generators. The production of prompt photons in association with jets provides an additional testing ground for perturbative QCD (pQCD) with a hard colourless probe less affected by hadronisation effects than jet production. The dynamics of isolated-photon plus jet production in pp collisions at a centre-of-mass energy of 8 and 13 TeV will be presented and discussed.

  10. The ATLAS Software Installation System v2: a highly available system to install and validate Grid and Cloud sites via Panda

    Science.gov (United States)

    De Salvo, A.; Kataoka, M.; Sanchez Pineda, A.; Smirnov, Y.

    2015-12-01

    The ATLAS Installation System v2 is the evolution of the original system, used since 2003. The original tool has been completely re-designed in terms of database backend and components, adding support for submission to multiple backends, including the original Workload Management Service (WMS) and the new PanDA modules. The database engine has been changed from plain MySQL to Galera/Percona and the table structure has been optimized to allow a full High-Availability (HA) solution over Wide Area Network. The servlets, running on each frontend, have been also decoupled from local settings, to allow an easy scalability of the system, including the possibility of an HA system with multiple sites. The clients can also be run in multiple copies and in different geographical locations, and take care of sending the installation and validation jobs to the target Grid or Cloud sites. Moreover, the Installation Database is used as source of parameters by the automatic agents running in CVMFS, in order to install the software and distribute it to the sites. The system is in production for ATLAS since 2013, having as main sites in HA the INFN Roma Tier 2 and the CERN Agile Infrastructure. The Light Job Submission Framework for Installation (LJSFi) v2 engine is directly interfacing with PanDA for the Job Management, the Atlas Grid Information System (AGIS) for the site parameter configurations, and CVMFS for both core components and the installation of the software itself. LJSFi2 is also able to use other plugins, and is essentially Virtual Organization (VO) agnostic, so can be directly used and extended to cope with the requirements of any Grid or Cloud enabled VO. In this work we will present the architecture, performance, status and possible evolutions to the system for the LHC Run2 and beyond.

  11. Software Validation in ATLAS

    International Nuclear Information System (INIS)

    Hodgkinson, Mark; Seuster, Rolf; Simmons, Brinick; Sherwood, Peter; Rousseau, David

    2012-01-01

    The ATLAS collaboration operates an extensive set of protocols to validate the quality of the offline software in a timely manner. This is essential in order to process the large amounts of data being collected by the ATLAS detector in 2011 without complications on the offline software side. We will discuss a number of different strategies used to validate the ATLAS offline software; running the ATLAS framework software, Athena, in a variety of configurations daily on each nightly build via the ATLAS Nightly System (ATN) and Run Time Tester (RTT) systems; the monitoring of these tests and checking the compilation of the software via distributed teams of rotating shifters; monitoring of and follow up on bug reports by the shifter teams and periodic software cleaning weeks to improve the quality of the offline software further.

  12. Measurement of the jet production properties at the LHC with the ATLAS Detector

    CERN Document Server

    Tokar, Stanislav; The ATLAS collaboration

    2015-01-01

    Latest results on the jet production properties obtained by the ATLAS experiment are presented. Jet physics events produced in pp collisions at 7 TeV have been studied. The dijet double differential cross sections as a function of dijet invariant mass for different rapidity regions are reported. Dijet flavour decomposition, a lower limit on the quark compositeness scale, inclusive jet production differential cross sections, three-jet differential production cross sections as well as a measurement of jet shapes for b-jets and light jets from top quark pair events are also reported.

  13. ATLAS B-physics potential

    CERN Document Server

    Smizanska, M

    2001-01-01

    Studies since 1993 have demonstrated the ability of ATLAS to pursue a wide B physics program. This document presents the latest performance studies with special stress on lepton identification. B-decays containing several leptons in ATLAS statistically dominate the high- precision measurements. We present new results on physics simulations of CP violation measurements in the B/sub s//sup 0/ to J/ psi phi decay and on a novel ATLAS programme on beauty production in central proton-proton collisions at the LHC. (7 refs).

  14. The Error Reporting in the ATLAS TDAQ System

    Science.gov (United States)

    Kolos, Serguei; Kazarov, Andrei; Papaevgeniou, Lykourgos

    2015-05-01

    The ATLAS Error Reporting provides a service that allows experts and shift crew to track and address errors relating to the data taking components and applications. This service, called the Error Reporting Service (ERS), gives to software applications the opportunity to collect and send comprehensive data about run-time errors, to a place where it can be intercepted in real-time by any other system component. Other ATLAS online control and monitoring tools use the ERS as one of their main inputs to address system problems in a timely manner and to improve the quality of acquired data. The actual destination of the error messages depends solely on the run-time environment, in which the online applications are operating. When an application sends information to ERS, depending on the configuration, it may end up in a local file, a database, distributed middleware which can transport it to an expert system or display it to users. Thanks to the open framework design of ERS, new information destinations can be added at any moment without touching the reporting and receiving applications. The ERS Application Program Interface (API) is provided in three programming languages used in the ATLAS online environment: C++, Java and Python. All APIs use exceptions for error reporting but each of them exploits advanced features of a given language to simplify the end-user program writing. For example, as C++ lacks language support for exceptions, a number of macros have been designed to generate hierarchies of C++ exception classes at compile time. Using this approach a software developer can write a single line of code to generate a boilerplate code for a fully qualified C++ exception class declaration with arbitrary number of parameters and multiple constructors, which encapsulates all relevant static information about the given type of issues. When a corresponding error occurs at run time, the program just need to create an instance of that class passing relevant values to one

  15. A Fast hardware Tracker for the ATLAS Trigger system

    CERN Document Server

    Pandini, Carlo Enrico; The ATLAS collaboration

    2015-01-01

    The trigger system at the ATLAS experiment is designed to lower the event rate occurring from the nominal bunch crossing at 40 MHz to about 1 kHz for a designed LHC luminosity of 10$^{34}$ cm$^{-2}$ s$^{-1}$. After a very successful data taking run the LHC is expected to run starting in 2015 with much higher instantaneous luminosities and this will increase the load on the High Level Trigger system. More sophisticated algorithms will be needed to achieve higher background rejection while maintaining good efficiency for interesting physics signals, which requires a more extensive use of tracking information. The Fast Tracker (FTK) trigger system, part of the ATLAS trigger upgrade program, is a highly parallel hardware device designed to perform full-scan track-finding at the event rate of 100 kHz. FTK is a dedicated processor based on a mixture of advanced technologies. Modern, powerful, Field Programmable Gate Arrays form an important part of the system architecture, and the combinatorial problem of pattern r...

  16. The GNAM system in the ATLAS online monitoring framework

    Energy Technology Data Exchange (ETDEWEB)

    Salvatore, D. [INFN Cosenza and Dip. di Fisica, Universita della Calabria, ponte P. Bucci 31 C, 87036 Rende (Italy)], E-mail: daniela.salvatore@cern.ch; Adragna, P. [Queen Mary, University of London, London (United Kingdom); Bosman, M. [IFAE, Institut de Fisica de Altes Energies, UAB/Barcelona (Spain); Burckhart, D. [CERN, Geneva (Switzerland); Caprini, M. [National Institute for Physics and Nuclear Engineering, Bucharest (Romania); Corso-Radu, A. [University of California Irvine, Irvine, California (United States); Costa, M.J. [CERN, Geneva (Switzerland); Della Pietra, M. [INFN Sezione diNapoli, Napoli (Italy); Dotti, A. [Universita and INFN Pisa, Pisa (Italy); Eschrich, I. [University of California Irvine, Irvine, California (United States); Ferrari, R. [INFN Sezione di Pavia, Pavia (Italy); Ferrer, M.L. [INFN Laboratori Nazionali di Frascati, Frascati (Italy); Gaudio, G. [INFN Sezione di Pavia, Pavia (Italy); Hadavand, H. [Southern Methodist University, Dallas (United States); Hauschild, M. [CERN, Geneva (Switzerland); Hillier, S. [University of Birmingham, Birmingham (United Kingdom); Kehoe, B. [Southern Methodist University, Dallas (United States); Kolos, S. [University of California Irvine, Irvine, California (United States); Kordas, K. [INFN Laboratori Nazionali di Frascati, Frascati (Italy); Mcpherson, R. [University of Victoria, Vancouver (Canada)] (and others)

    2007-10-15

    ATLAS [ATLAS Collaboration, 'ATLAS Technical Proposal', CERN/LHHCC/94-43, LHCC/P2, CERN, Geneva, Switzerland, 1994] is one of the four experiments under construction along the Large Hadron Collider (LHC) ring, which will produce interactions at a center of mass energy of 14 TeV at 40 MHz rate. The detector consists of more than 140 million electronic channels. The challenging experimental environment and the extreme detector complexity impose the necessity of a common scalable distributed monitoring framework, which can be tuned for the optimal use by different ATLAS detectors at the various levels of the ATLAS data flow.

  17. The GNAM system in the ATLAS online monitoring framework

    Energy Technology Data Exchange (ETDEWEB)

    Salvatore, D. [INFN Cosenza and Dip. di Fisica, Universita della Calabria, ponte P. Bucci 31 C, 87036 Rende (Italy)], E-mail: daniela.salvatore@cern.ch; Adragna, P [Queen Mary, University of London, London (United Kingdom); Bosman, M [IFAE, Institut de Fisica de Altes Energies, UAB/Barcelona (Spain); Burckhart, D [CERN, Geneva (Switzerland); Caprini, M [National Institute for Physics and Nuclear Engineering, Bucharest (Romania); Corso-Radu, A [University of California Irvine, Irvine, California (United States); Costa, M J [CERN, Geneva (Switzerland); Della Pietra, M [INFN Sezione diNapoli, Napoli (Italy); Dotti, A [Universita and INFN Pisa, Pisa (Italy); Eschrich, I [University of California Irvine, Irvine, California (United States); Ferrari, R [INFN Sezione di Pavia, Pavia (Italy); Ferrer, M L [INFN Laboratori Nazionali di Frascati, Frascati (Italy); Gaudio, G [INFN Sezione di Pavia, Pavia (Italy); Hadavand, H [Southern Methodist University, Dallas (United States); Hauschild, M [CERN, Geneva (Switzerland); Hillier, S [University of Birmingham, Birmingham (United Kingdom); Kehoe, B [Southern Methodist University, Dallas (United States); Kolos, S [University of California Irvine, Irvine, California (United States); Kordas, K [INFN Laboratori Nazionali di Frascati, Frascati (Italy); Mcpherson, R [University of Victoria, Vancouver (Canada)

    2007-10-15

    ATLAS [ATLAS Collaboration, 'ATLAS Technical Proposal', CERN/LHHCC/94-43, LHCC/P2, CERN, Geneva, Switzerland, 1994] is one of the four experiments under construction along the Large Hadron Collider (LHC) ring, which will produce interactions at a center of mass energy of 14 TeV at 40 MHz rate. The detector consists of more than 140 million electronic channels. The challenging experimental environment and the extreme detector complexity impose the necessity of a common scalable distributed monitoring framework, which can be tuned for the optimal use by different ATLAS detectors at the various levels of the ATLAS data flow.

  18. The GNAM system in the ATLAS online monitoring framework

    International Nuclear Information System (INIS)

    Salvatore, D.; Adragna, P.; Bosman, M.; Burckhart, D.; Caprini, M.; Corso-Radu, A.; Costa, M.J.; Della Pietra, M.; Dotti, A.; Eschrich, I.; Ferrari, R.; Ferrer, M.L.; Gaudio, G.; Hadavand, H.; Hauschild, M.; Hillier, S.; Kehoe, B.; Kolos, S.; Kordas, K.; Mcpherson, R.

    2007-01-01

    ATLAS [ATLAS Collaboration, 'ATLAS Technical Proposal', CERN/LHHCC/94-43, LHCC/P2, CERN, Geneva, Switzerland, 1994] is one of the four experiments under construction along the Large Hadron Collider (LHC) ring, which will produce interactions at a center of mass energy of 14 TeV at 40 MHz rate. The detector consists of more than 140 million electronic channels. The challenging experimental environment and the extreme detector complexity impose the necessity of a common scalable distributed monitoring framework, which can be tuned for the optimal use by different ATLAS detectors at the various levels of the ATLAS data flow

  19. Studies of Zγ production in association with a high-mass dijet system in pp collisions at $ \\sqrt{s}=8$ TeV with the ATLAS detector

    Energy Technology Data Exchange (ETDEWEB)

    Aaboud, M.; Aad, G.; Abbott, B.; Abdallah, J.; Abdinov, O.; Abeloos, B.; Abidi, S. H.; AbouZeid, O. S.; Abraham, N. L.; Abramowicz, H.; Abreu, H.; Abreu, R.; Abulaiti, Y.; Acharya, B. S.; Adachi, S.; Adamczyk, L.; Adelman, J.; Adersberger, M.; Adye, T.; Affolder, A. A.; Agatonovic-Jovin, T.; Agheorghiesei, C.; Aguilar-Saavedra, J. A.; Ahlen, S. P.; Ahmadov, F.; Aielli, G.; Akatsuka, S.; Akerstedt, H.; Åkesson, T. P. A.; Akimov, A. V.; Alberghi, G. L.; Albert, J.; Alconada Verzini, M. J.; Aleksa, M.; Aleksandrov, I. N.; Alexa, C.; Alexander, G.; Alexopoulos, T.; Alhroob, M.; Ali, B.; Aliev, M.; Alimonti, G.; Alison, J.; Alkire, S. P.; Allbrooke, B. M. M.; Allen, B. W.; Allport, P. P.; Aloisio, A.; Alonso, A.; Alonso, F.; Alpigiani, C.; Alshehri, A. A.; Alstaty, M.; Alvarez Gonzalez, B.; Álvarez Piqueras, D.; Alviggi, M. G.; Amadio, B. T.; Amaral Coutinho, Y.; Amelung, C.; Amidei, D.; Amor Dos Santos, S. P.; Amorim, A.; Amoroso, S.; Amundsen, G.; Anastopoulos, C.; Ancu, L. S.; Andari, N.; Andeen, T.; Anders, C. F.; Anders, J. K.; Anderson, K. J.; Andreazza, A.; Andrei, V.; Angelidakis, S.; Angelozzi, I.; Angerami, A.; Anisenkov, A. V.; Anjos, N.; Annovi, A.; Antel, C.; Antonelli, M.; Antonov, A.; Antrim, D. J.; Anulli, F.; Aoki, M.; Aperio Bella, L.; Arabidze, G.; Arai, Y.; Araque, J. P.; Araujo Ferraz, V.; Arce, A. T. H.; Ardell, R. E.; Arduh, F. A.; Arguin, J-F.; Argyropoulos, S.; Arik, M.; Armbruster, A. J.; Armitage, L. J.; Arnaez, O.; Arnold, H.; Arratia, M.; Arslan, O.; Artamonov, A.; Artoni, G.; Artz, S.; Asai, S.; Asbah, N.; Ashkenazi, A.; Asquith, L.; Assamagan, K.; Astalos, R.; Atkinson, M.; Atlay, N. B.; Augsten, K.; Avolio, G.; Axen, B.; Ayoub, M. K.; Azuelos, G.; Baas, A. E.; Baca, M. J.; Bachacou, H.; Bachas, K.; Backes, M.; Backhaus, M.; Bagiacchi, P.; Bagnaia, P.; Bahrasemani, H.; Baines, J. T.; Bajic, M.; Baker, O. K.; Baldin, E. M.; Balek, P.; Balestri, T.; Balli, F.; Balunas, W. K.; Banas, E.; Banerjee, Sw.; Bannoura, A. A. E.; Barak, L.; Barberio, E. L.; Barberis, D.; Barbero, M.; Barillari, T.; Barisits, M-S; Barklow, T.; Barlow, N.; Barnes, S. L.; Barnett, B. M.; Barnett, R. M.; Barnovska-Blenessy, Z.; Baroncelli, A.; Barone, G.; Barr, A. J.; Barranco Navarro, L.; Barreiro, F.; Barreiro Guimarães da Costa, J.; Bartoldus, R.; Barton, A. E.; Bartos, P.; Basalaev, A.; Bassalat, A.; Bates, R. L.; Batista, S. J.; Batley, J. R.; Battaglia, M.; Bauce, M.; Bauer, F.; Bawa, H. S.; Beacham, J. B.; Beattie, M. D.; Beau, T.; Beauchemin, P. H.; Bechtle, P.; Beck, H. P.; Becker, K.; Becker, M.; Beckingham, M.; Becot, C.; Beddall, A. J.; Beddall, A.; Bednyakov, V. A.; Bedognetti, M.; Bee, C. P.; Beermann, T. A.; Begalli, M.; Begel, M.; Behr, J. K.; Bell, A. S.; Bella, G.; Bellagamba, L.; Bellerive, A.; Bellomo, M.; Belotskiy, K.; Beltramello, O.; Belyaev, N. L.; Benary, O.; Benchekroun, D.; Bender, M.; Bendtz, K.; Benekos, N.; Benhammou, Y.; Benhar Noccioli, E.; Benitez, J.; Benjamin, D. P.; Benoit, M.; Bensinger, J. R.; Bentvelsen, S.; Beresford, L.; Beretta, M.; Berge, D.; Bergeaas Kuutmann, E.; Berger, N.; Beringer, J.; Berlendis, S.; Bernard, N. R.; Bernardi, G.; Bernius, C.; Bernlochner, F. U.; Berry, T.; Berta, P.; Bertella, C.; Bertoli, G.; Bertolucci, F.; Bertram, I. A.; Bertsche, C.; Bertsche, D.; Besjes, G. J.; Bessidskaia Bylund, O.; Bessner, M.; Besson, N.; Betancourt, C.; Bethani, A.; Bethke, S.; Bevan, A. J.; Bianchi, R. M.; Biebel, O.; Biedermann, D.; Bielski, R.; Biesuz, N. V.; Biglietti, M.; Bilbao De Mendizabal, J.; Billoud, T. R. V.; Bilokon, H.; Bindi, M.; Bingul, A.; Bini, C.; Biondi, S.; Bisanz, T.; Bittrich, C.; Bjergaard, D. M.; Black, C. W.; Black, J. E.; Black, K. M.; Blackburn, D.; Blair, R. E.; Blazek, T.; Bloch, I.; Blocker, C.; Blue, A.; Blum, W.; Blumenschein, U.; Blunier, S.; Bobbink, G. J.; Bobrovnikov, V. S.; Bocchetta, S. S.; Bocci, A.; Bock, C.; Boehler, M.; Boerner, D.; Bogavac, D.; Bogdanchikov, A. G.; Bohm, C.; Boisvert, V.; Bokan, P.; Bold, T.; Boldyrev, A. S.; Bolz, A. E.; Bomben, M.; Bona, M.; Boonekamp, M.; Borisov, A.; Borissov, G.; Bortfeldt, J.; Bortoletto, D.; Bortolotto, V.; Boscherini, D.; Bosman, M.; Bossio Sola, J. D.; Boudreau, J.; Bouffard, J.; Bouhova-Thacker, E. V.; Boumediene, D.; Bourdarios, C.; Boutle, S. K.; Boveia, A.; Boyd, J.; Boyko, I. R.; Bracinik, J.; Brandt, A.; Brandt, G.; Brandt, O.; Bratzler, U.; Brau, B.; Brau, J. E.; Breaden Madden, W. D.; Brendlinger, K.; Brennan, A. J.; Brenner, L.; Brenner, R.; Bressler, S.; Briglin, D. L.; Bristow, T. M.; Britton, D.; Britzger, D.; Brochu, F. M.; Brock, I.; Brock, R.; Brooijmans, G.; Brooks, T.; Brooks, W. K.; Brosamer, J.; Brost, E.; Broughton, J. H.; de Renstrom, P. A. Bruckman; Bruncko, D.; Bruni, A.; Bruni, G.; Bruni, L. S.; Brunt, BH; Bruschi, M.; Bruscino, N.; Bryant, P.; Bryngemark, L.; Buanes, T.; Buat, Q.; Buchholz, P.; Buckley, A. G.; Budagov, I. A.; Buehrer, F.; Bugge, M. K.; Bulekov, O.; Bullock, D.; Burckhart, H.; Burdin, S.; Burgard, C. D.; Burger, A. M.; Burghgrave, B.; Burka, K.; Burke, S.; Burmeister, I.; Burr, J. T. P.; Busato, E.; Büscher, D.; Büscher, V.; Bussey, P.; Butler, J. M.; Buttar, C. M.; Butterworth, J. M.; Butti, P.; Buttinger, W.; Buzatu, A.; Buzykaev, A. R.; Cabrera Urbán, S.; Caforio, D.; Cairo, V. M.; Cakir, O.; Calace, N.; Calafiura, P.; Calandri, A.; Calderini, G.; Calfayan, P.; Callea, G.; Caloba, L. P.; Calvente Lopez, S.; Calvet, D.; Calvet, S.; Calvet, T. P.; Camacho Toro, R.; Camarda, S.; Camarri, P.; Cameron, D.; Caminal Armadans, R.; Camincher, C.; Campana, S.; Campanelli, M.; Camplani, A.; Campoverde, A.; Canale, V.; Cano Bret, M.; Cantero, J.; Cao, T.; Capeans Garrido, M. D. M.; Caprini, I.; Caprini, M.; Capua, M.; Carbone, R. M.; Cardarelli, R.; Cardillo, F.; Carli, I.; Carli, T.; Carlino, G.; Carlson, B. T.; Carminati, L.; Carney, R. M. D.; Caron, S.; Carquin, E.; Carrá, S.; Carrillo-Montoya, G. D.; Carvalho, J.; Casadei, D.; Casado, M. P.; Casolino, M.; Casper, D. W.; Castelijn, R.; Castelli, A.; Castillo Gimenez, V.; Castro, N. F.; Catinaccio, A.; Catmore, J. R.; Cattai, A.; Caudron, J.; Cavaliere, V.; Cavallaro, E.; Cavalli, D.; Cavalli-Sforza, M.; Cavasinni, V.; Celebi, E.; Ceradini, F.; Cerda Alberich, L.; Cerqueira, A. S.; Cerri, A.; Cerrito, L.; Cerutti, F.; Cervelli, A.; Cetin, S. A.; Chafaq, A.; Chakraborty, D.; Chan, S. K.; Chan, W. S.; Chan, Y. L.; Chang, P.; Chapman, J. D.; Charlton, D. G.; Chatterjee, A.; Chau, C. C.; Chavez Barajas, C. A.; Che, S.; Cheatham, S.; Chegwidden, A.; Chekanov, S.; Chekulaev, S. V.; Chelkov, G. A.; Chelstowska, M. A.; Chen, C.; Chen, H.; Chen, S.; Chen, S.; Chen, X.; Chen, Y.; Cheng, H. C.; Cheng, H. J.; Cheng, Y.; Cheplakov, A.; Cheremushkina, E.; Cherkaoui El Moursli, R.; Chernyatin, V.; Cheu, E.; Chevalier, L.; Chiarella, V.; Chiarelli, G.; Chiodini, G.; Chisholm, A. S.; Chitan, A.; Chiu, Y. H.; Chizhov, M. V.; Choi, K.; Chomont, A. R.; Chouridou, S.; Chow, B. K. B.; Christodoulou, V.; Chromek-Burckhart, D.; Chu, M. C.; Chudoba, J.; Chuinard, A. J.; Chwastowski, J. J.; Chytka, L.; Ciftci, A. K.; Cinca, D.; Cindro, V.; Cioara, I. A.; Ciocca, C.; Ciocio, A.; Cirotto, F.; Citron, Z. H.; Citterio, M.; Ciubancan, M.; Clark, A.; Clark, B. L.; Clark, M. R.; Clark, P. J.; Clarke, R. N.; Clement, C.; Coadou, Y.; Cobal, M.; Coccaro, A.; Cochran, J.; Colasurdo, L.; Cole, B.; Colijn, A. P.; Collot, J.; Colombo, T.; Conde Muiño, P.; Coniavitis, E.; Connell, S. H.; Connelly, I. A.; Consorti, V.; Constantinescu, S.; Conti, G.; Conventi, F.; Cooke, M.; Cooper, B. D.; Cooper-Sarkar, A. M.; Cormier, F.; Cormier, K. J. R.; Corradi, M.; Corriveau, F.; Cortes-Gonzalez, A.; Cortiana, G.; Costa, G.; Costa, M. J.; Costanzo, D.; Cottin, G.; Cowan, G.; Cox, B. E.; Cranmer, K.; Crawley, S. J.; Creager, R. A.; Cree, G.; Crépé-Renaudin, S.; Crescioli, F.; Cribbs, W. A.; Crispin Ortuzar, M.; Cristinziani, M.; Croft, V.; Crosetti, G.; Cueto, A.; Cuhadar Donszelmann, T.; Cukierman, A. R.; Cummings, J.; Curatolo, M.; Cúth, J.; Czirr, H.; Czodrowski, P.; D’amen, G.; D’Auria, S.; D’Onofrio, M.; Da Cunha Sargedas De Sousa, M. J.; Da Via, C.; Dabrowski, W.; Dado, T.; Dai, T.; Dale, O.; Dallaire, F.; Dallapiccola, C.; Dam, M.; Dandoy, J. R.; Dang, N. P.; Daniells, A. C.; Dann, N. S.; Danninger, M.; Dano Hoffmann, M.; Dao, V.; Darbo, G.; Darmora, S.; Dassoulas, J.; Dattagupta, A.; Daubney, T.; Davey, W.; David, C.; Davidek, T.; Davies, M.; Davison, P.; Dawe, E.; Dawson, I.; De, K.; de Asmundis, R.; De Benedetti, A.; De Castro, S.; De Cecco, S.; De Groot, N.; de Jong, P.; De la Torre, H.; De Lorenzi, F.; De Maria, A.; De Pedis, D.; De Salvo, A.; De Sanctis, U.; De Santo, A.; De Vasconcelos Corga, K.; De Vivie De Regie, J. B.; Dearnaley, W. J.; Debbe, R.; Debenedetti, C.; Dedovich, D. V.; Dehghanian, N.; Deigaard, I.; Del Gaudio, M.; Del Peso, J.; Del Prete, T.; Delgove, D.; Deliot, F.; Delitzsch, C. M.; Dell’Acqua, A.; Dell’Asta, L.; Dell’Orso, M.; Della Pietra, M.; della Volpe, D.; Delmastro, M.; Delporte, C.; Delsart, P. A.; DeMarco, D. A.; Demers, S.; Demichev, M.; Demilly, A.; Denisov, S. P.; Denysiuk, D.; Derendarz, D.; Derkaoui, J. E.; Derue, F.; Dervan, P.; Desch, K.; Deterre, C.; Dette, K.; Deviveiros, P. O.; Dewhurst, A.; Dhaliwal, S.; Di Ciaccio, A.; Di Ciaccio, L.; Di Clemente, W. K.; Di Donato, C.; Di Girolamo, A.; Di Girolamo, B.; Di Micco, B.; Di Nardo, R.; Di Petrillo, K. F.; Di Simone, A.; Di Sipio, R.; Di Valentino, D.; Diaconu, C.; Diamond, M.; Dias, F. A.; Diaz, M. A.; Diehl, E. B.; Dietrich, J.; Díez Cornell, S.; Dimitrievska, A.; Dingfelder, J.; Dita, P.; Dita, S.; Dittus, F.; Djama, F.; Djobava, T.; Djuvsland, J. I.; do Vale, M. A. B.; Dobos, D.; Dobre, M.; Doglioni, C.; Dolejsi, J.; Dolezal, Z.; Donadelli, M.; Donati, S.; Dondero, P.; Donini, J.; Dopke, J.; Doria, A.; Dova, M. T.; Doyle, A. T.; Drechsler, E.; Dris, M.; Du, Y.; Duarte-Campderros, J.; Duchovni, E.; Duckeck, G.; Ducourthial, A.; Ducu, O. A.; Duda, D.; Dudarev, A.; Dudder, A. Chr.; Duffield, E. M.; Duflot, L.; Dührssen, M.; Dumancic, M.; Dumitriu, A. E.; Duncan, A. K.; Dunford, M.; Duran Yildiz, H.; Düren, M.; Durglishvili, A.; Duschinger, D.; Dutta, B.; Dyndal, M.; Eckardt, C.; Ecker, K. M.; Edgar, R. C.; Eifert, T.; Eigen, G.; Einsweiler, K.; Ekelof, T.; El Kacimi, M.; El Kosseifi, R.; Ellajosyula, V.; Ellert, M.; Elles, S.; Ellinghaus, F.; Elliot, A. A.; Ellis, N.; Elmsheuser, J.; Elsing, M.; Emeliyanov, D.; Enari, Y.; Endner, O. C.; Ennis, J. S.; Erdmann, J.; Ereditato, A.; Ernis, G.; Ernst, M.; Errede, S.; Ertel, E.; Escalier, M.; Esch, H.; Escobar, C.; Esposito, B.; Estrada Pastor, O.; Etienvre, A. I.; Etzion, E.; Evans, H.; Ezhilov, A.; Ezzi, M.; Fabbri, F.; Fabbri, L.; Facini, G.; Fakhrutdinov, R. M.; Falciano, S.; Falla, R. J.; Faltova, J.; Fang, Y.; Fanti, M.; Farbin, A.; Farilla, A.; Farina, C.; Farina, E. M.; Farooque, T.; Farrell, S.; Farrington, S. M.; Farthouat, P.; Fassi, F.; Fassnacht, P.; Fassouliotis, D.; Faucci Giannelli, M.; Favareto, A.; Fawcett, W. J.; Fayard, L.; Fedin, O. L.; Fedorko, W.; Feigl, S.; Feligioni, L.; Feng, C.; Feng, E. J.; Feng, H.; Fenyuk, A. B.; Feremenga, L.; Fernandez Martinez, P.; Fernandez Perez, S.; Ferrando, J.; Ferrari, A.; Ferrari, P.; Ferrari, R.; Ferreira de Lima, D. E.; Ferrer, A.; Ferrere, D.; Ferretti, C.; Fiedler, F.; Filipčič, A.; Filipuzzi, M.; Filthaut, F.; Fincke-Keeler, M.; Finelli, K. D.; Fiolhais, M. C. N.; Fiorini, L.; Fischer, A.; Fischer, C.; Fischer, J.; Fisher, W. C.; Flaschel, N.; Fleck, I.; Fleischmann, P.; Fletcher, R. R. M.; Flick, T.; Flierl, B. M.; Flores Castillo, L. R.; Flowerdew, M. J.; Forcolin, G. T.; Formica, A.; Förster, F. A.; Forti, A.; Foster, A. G.; Fournier, D.; Fox, H.; Fracchia, S.; Francavilla, P.; Franchini, M.; Franchino, S.; Francis, D.; Franconi, L.; Franklin, M.; Frate, M.; Fraternali, M.; Freeborn, D.; Fressard-Batraneanu, S. M.; Freund, B.; Froidevaux, D.; Frost, J. A.; Fukunaga, C.; Fusayasu, T.; Fuster, J.; Gabaldon, C.; Gabizon, O.; Gabrielli, A.; Gabrielli, A.; Gach, G. P.; Gadatsch, S.; Gadomski, S.; Gagliardi, G.; Gagnon, L. G.; Gagnon, P.; Galea, C.; Galhardo, B.; Gallas, E. J.; Gallop, B. J.; Gallus, P.; Galster, G.; Gan, K. K.; Ganguly, S.; Gao, J.; Gao, Y.; Gao, Y. S.; Garay Walls, F. M.; García, C.; García Navarro, J. E.; Garcia-Sciveres, M.; Gardner, R. W.; Garelli, N.; Garonne, V.; Gascon Bravo, A.; Gasnikova, K.; Gatti, C.; Gaudiello, A.; Gaudio, G.; Gavrilenko, I. L.; Gay, C.; Gaycken, G.; Gazis, E. N.; Gee, C. N. P.; Geisen, M.; Geisler, M. P.; Gellerstedt, K.; Gemme, C.; Genest, M. H.; Geng, C.; Gentile, S.; Gentsos, C.; George, S.; Gerbaudo, D.; Gershon, A.; Ghasemi, S.; Ghneimat, M.; Giacobbe, B.; Giagu, S.; Giannetti, P.; Gibson, S. M.; Gignac, M.; Gilchriese, M.; Gillberg, D.; Gilles, G.; Gingrich, D. M.; Giokaris, N.; Giordani, M. P.; Giorgi, F. M.; Giraud, P. F.; Giromini, P.; Giugni, D.; Giuli, F.; Giuliani, C.; Giulini, M.; Gjelsten, B. K.; Gkaitatzis, S.; Gkialas, I.; Gkougkousis, E. L.; Gladilin, L. K.; Glasman, C.; Glatzer, J.; Glaysher, P. C. F.; Glazov, A.; Goblirsch-Kolb, M.; Godlewski, J.; Goldfarb, S.; Golling, T.; Golubkov, D.; Gomes, A.; Gonçalo, R.; Goncalves Gama, R.; Goncalves Pinto Firmino Da Costa, J.; Gonella, G.; Gonella, L.; Gongadze, A.; González de la Hoz, S.; Gonzalez-Sevilla, S.; Goossens, L.; Gorbounov, P. A.; Gordon, H. A.; Gorelov, I.; Gorini, B.; Gorini, E.; Gorišek, A.; Goshaw, A. T.; Gössling, C.; Gostkin, M. I.; Goudet, C. R.; Goujdami, D.; Goussiou, A. G.; Govender, N.; Gozani, E.; Graber, L.; Grabowska-Bold, I.; Gradin, P. O. J.; Gramling, J.; Gramstad, E.; Grancagnolo, S.; Gratchev, V.; Gravila, P. M.; Gray, C.; Gray, H. M.; Greenwood, Z. D.; Grefe, C.; Gregersen, K.; Gregor, I. M.; Grenier, P.; Grevtsov, K.; Griffiths, J.; Grillo, A. A.; Grimm, K.; Grinstein, S.; Gris, Ph.; Grivaz, J. -F.; Groh, S.; Gross, E.; Grosse-Knetter, J.; Grossi, G. C.; Grout, Z. J.; Grummer, A.; Guan, L.; Guan, W.; Guenther, J.; Guescini, F.; Guest, D.; Gueta, O.; Gui, B.; Guido, E.; Guillemin, T.; Guindon, S.; Gul, U.; Gumpert, C.; Guo, J.; Guo, W.; Guo, Y.; Gupta, R.; Gupta, S.; Gustavino, G.; Gutierrez, P.; Gutierrez Ortiz, N. G.; Gutschow, C.; Guyot, C.; Guzik, M. P.; Gwenlan, C.; Gwilliam, C. B.; Haas, A.; Haber, C.; Hadavand, H. K.; Haddad, N.; Hadef, A.; Hageböck, S.; Hagihara, M.; Hakobyan, H.; Haleem, M.; Haley, J.; Halladjian, G.; Hallewell, G. D.; Hamacher, K.; Hamal, P.; Hamano, K.; Hamilton, A.; Hamity, G. N.; Hamnett, P. G.; Han, L.; Han, S.; Hanagaki, K.; Hanawa, K.; Hance, M.; Haney, B.; Hanke, P.; Hansen, J. B.; Hansen, J. D.; Hansen, M. C.; Hansen, P. H.; Hara, K.; Hard, A. S.; Harenberg, T.; Hariri, F.; Harkusha, S.; Harrington, R. D.; Harrison, P. F.; Hartmann, N. M.; Hasegawa, M.; Hasegawa, Y.; Hasib, A.; Hassani, S.; Haug, S.; Hauser, R.; Hauswald, L.; Havener, L. B.; Havranek, M.; Hawkes, C. M.; Hawkings, R. J.; Hayakawa, D.; Hayden, D.; Hays, C. P.; Hays, J. M.; Hayward, H. S.; Haywood, S. J.; Head, S. J.; Heck, T.; Hedberg, V.; Heelan, L.; Heidegger, K. K.; Heim, S.; Heim, T.; Heinemann, B.; Heinrich, J. J.; Heinrich, L.; Heinz, C.; Hejbal, J.; Helary, L.; Held, A.; Hellman, S.; Helsens, C.; Henderson, J.; Henderson, R. C. W.; Heng, Y.; Henkelmann, S.; Henriques Correia, A. M.; Henrot-Versille, S.; Herbert, G. H.; Herde, H.; Herget, V.; Hernández Jiménez, Y.; Herten, G.; Hertenberger, R.; Hervas, L.; Herwig, T. C.; Hesketh, G. G.; Hessey, N. P.; Hetherly, J. W.; Higashino, S.; Higón-Rodriguez, E.; Hill, E.; Hill, J. C.; Hiller, K. H.; Hillier, S. J.; Hinchliffe, I.; Hirose, M.; Hirschbuehl, D.; Hiti, B.; Hladik, O.; Hoad, X.; Hobbs, J.; Hod, N.; Hodgkinson, M. C.; Hodgson, P.; Hoecker, A.; Hoeferkamp, M. R.; Hoenig, F.; Hohn, D.; Holmes, T. R.; Homann, M.; Honda, S.; Honda, T.; Hong, T. M.; Hooberman, B. H.; Hopkins, W. H.; Horii, Y.; Horton, A. J.; Hostachy, J-Y.; Hou, S.; Hoummada, A.; Howarth, J.; Hoya, J.; Hrabovsky, M.; Hristova, I.; Hrivnac, J.; Hryn’ova, T.; Hrynevich, A.; Hsu, P. J.; Hsu, S. -C.; Hu, Q.; Hu, S.; Huang, Y.; Hubacek, Z.; Hubaut, F.; Huegging, F.; Huffman, T. B.; Hughes, E. W.; Hughes, G.; Huhtinen, M.; Huo, P.; Huseynov, N.; Huston, J.; Huth, J.; Iacobucci, G.; Iakovidis, G.; Ibragimov, I.; Iconomidou-Fayard, L.; Idrissi, Z.; Iengo, P.; Igonkina, O.; Iizawa, T.; Ikegami, Y.; Ikeno, M.; Ilchenko, Y.; Iliadis, D.; Ilic, N.; Introzzi, G.; Ioannou, P.; Iodice, M.; Iordanidou, K.; Ippolito, V.; Isacson, M. F.; Ishijima, N.; Ishino, M.; Ishitsuka, M.; Issever, C.; Istin, S.; Ito, F.; Iturbe Ponce, J. M.; Iuppa, R.; Iwasaki, H.; Izen, J. M.; Izzo, V.; Jabbar, S.; Jackson, P.; Jain, V.; Jakobi, K. B.; Jakobs, K.; Jakobsen, S.; Jakoubek, T.; Jamin, D. O.; Jana, D. K.; Jansky, R.; Janssen, J.; Janus, M.; Janus, P. A.; Jarlskog, G.; Javadov, N.; Javůrek, T.; Javurkova, M.; Jeanneau, F.; Jeanty, L.; Jejelava, J.; Jelinskas, A.; Jenni, P.; Jeske, C.; Jézéquel, S.; Ji, H.; Jia, J.; Jiang, H.; Jiang, Y.; Jiang, Z.; Jiggins, S.; Jimenez Pena, J.; Jin, S.; Jinaru, A.; Jinnouchi, O.; Jivan, H.; Johansson, P.; Johns, K. A.; Johnson, C. A.; Johnson, W. J.; Jon-And, K.; Jones, R. W. L.; Jones, S. D.; Jones, S.; Jones, T. J.; Jongmanns, J.; Jorge, P. M.; Jovicevic, J.; Ju, X.; Rozas, A. Juste; Köhler, M. K.; Kaczmarska, A.; Kado, M.; Kagan, H.; Kagan, M.; Kahn, S. J.; Kaji, T.; Kajomovitz, E.; Kalderon, C. W.; Kaluza, A.; Kama, S.; Kamenshchikov, A.; Kanaya, N.; Kaneti, S.; Kanjir, L.; Kantserov, V. A.; Kanzaki, J.; Kaplan, B.; Kaplan, L. S.; Kar, D.; Karakostas, K.; Karastathis, N.; Kareem, M. J.; Karentzos, E.; Karpov, S. N.; Karpova, Z. M.; Karthik, K.; Kartvelishvili, V.; Karyukhin, A. N.; Kasahara, K.; Kashif, L.; Kass, R. D.; Kastanas, A.; Kataoka, Y.; Kato, C.; Katre, A.; Katzy, J.; Kawade, K.; Kawagoe, K.; Kawamoto, T.; Kawamura, G.; Kay, E. F.; Kazanin, V. F.; Keeler, R.; Kehoe, R.; Keller, J. S.; Kempster, J. J.; Keoshkerian, H.; Kepka, O.; Kerševan, B. P.; Kersten, S.; Keyes, R. A.; Khader, M.; Khalil-zada, F.; Khanov, A.; Kharlamov, A. G.; Kharlamova, T.; Khodinov, A.; Khoo, T. J.; Khovanskiy, V.; Khramov, E.; Khubua, J.; Kido, S.; Kilby, C. R.; Kim, H. Y.; Kim, S. H.; Kim, Y. K.; Kimura, N.; Kind, O. M.; King, B. T.; Kirchmeier, D.; Kirk, J.; Kiryunin, A. E.; Kishimoto, T.; Kisielewska, D.; Kiuchi, K.; Kivernyk, O.; Kladiva, E.; Klapdor-Kleingrothaus, T.; Klein, M. H.; Klein, M.; Klein, U.; Kleinknecht, K.; Klimek, P.; Klimentov, A.; Klingenberg, R.; Klingl, T.; Klioutchnikova, T.; Kluge, E. -E.; Kluit, P.; Kluth, S.; Knapik, J.; Kneringer, E.; Knoops, E. B. F. G.; Knue, A.; Kobayashi, A.; Kobayashi, D.; Kobayashi, T.; Kobel, M.; Kocian, M.; Kodys, P.; Koffas, T.; Koffeman, E.; Köhler, N. M.; Koi, T.; Kolb, M.; Koletsou, I.; Komar, A. A.; Komori, Y.; Kondo, T.; Kondrashova, N.; Köneke, K.; König, A. C.; Kono, T.; Konoplich, R.; Konstantinidis, N.; Kopeliansky, R.; Koperny, S.; Kopp, A. K.; Korcyl, K.; Kordas, K.; Korn, A.; Korol, A. A.; Korolkov, I.; Korolkova, E. V.; Kortner, O.; Kortner, S.; Kosek, T.; Kostyukhin, V. V.; Kotwal, A.; Koulouris, A.; Kourkoumeli-Charalampidi, A.; Kourkoumelis, C.; Kourlitis, E.; Kouskoura, V.; Kowalewska, A. B.; Kowalewski, R.; Kowalski, T. Z.; Kozakai, C.; Kozanecki, W.; Kozhin, A. S.; Kramarenko, V. A.; Kramberger, G.; Krasnopevtsev, D.; Krasny, M. W.; Krasznahorkay, A.; Krauss, D.; Kravchenko, A.; Kremer, J. A.; Kretz, M.; Kretzschmar, J.; Kreutzfeldt, K.; Krieger, P.; Krizka, K.; Kroeninger, K.; Kroha, H.; Kroll, J.; Kroll, J.; Kroseberg, J.; Krstic, J.; Kruchonak, U.; Krüger, H.; Krumnack, N.; Kruse, M. C.; Kruskal, M.; Kubota, T.; Kucuk, H.; Kuday, S.; Kuechler, J. T.; Kuehn, S.; Kugel, A.; Kuger, F.; Kuhl, T.; Kukhtin, V.; Kukla, R.; Kulchitsky, Y.; Kuleshov, S.; Kulinich, Y. P.; Kuna, M.; Kunigo, T.; Kupco, A.; Kuprash, O.; Kurashige, H.; Kurchaninov, L. L.; Kurochkin, Y. A.; Kurth, M. G.; Kus, V.; Kuwertz, E. S.; Kuze, M.; Kvita, J.; Kwan, T.; Kyriazopoulos, D.; La Rosa, A.; La Rosa Navarro, J. L.; La Rotonda, L.; Lacasta, C.; Lacava, F.; Lacey, J.; Lacker, H.; Lacour, D.; Ladygin, E.; Lafaye, R.; Laforge, B.; Lagouri, T.; Lai, S.; Lammers, S.; Lampl, W.; Lançon, E.; Landgraf, U.; Landon, M. P. J.; Lanfermann, M. C.; Lang, V. S.; Lange, J. C.; Lankford, A. J.; Lanni, F.; Lantzsch, K.; Lanza, A.; Lapertosa, A.; Laplace, S.; Laporte, J. F.; Lari, T.; Lasagni Manghi, F.; Lassnig, M.; Laurelli, P.; Lavrijsen, W.; Law, A. T.; Laycock, P.; Lazovich, T.; Lazzaroni, M.; Le, B.; Le Dortz, O.; Le Guirriec, E.; Le Quilleuc, E. P.; LeBlanc, M.; LeCompte, T.; Ledroit-Guillon, F.; Lee, C. A.; Lee, G. R.; Lee, S. C.; Lee, L.; Lefebvre, B.; Lefebvre, G.; Lefebvre, M.; Legger, F.; Leggett, C.; Lehan, A.; Lehmann Miotto, G.; Lei, X.; Leight, W. A.; Leite, M. A. L.; Leitner, R.; Lellouch, D.; Lemmer, B.; Leney, K. J. C.; Lenz, T.; Lenzi, B.; Leone, R.; Leone, S.; Leonidopoulos, C.; Lerner, G.; Leroy, C.; Lesage, A. A. J.; Lester, C. G.; Levchenko, M.; Levêque, J.; Levin, D.; Levinson, L. J.; Levy, M.; Lewis, D.; Li, B.; Li, C.; Li, H.; Li, L.; Li, Q.; Li, S.; Li, X.; Li, Y.; Liang, Z.; Liberti, B.; Liblong, A.; Lie, K.; Liebal, J.; Liebig, W.; Limosani, A.; Lin, S. C.; Lin, T. H.; Lindquist, B. E.; Lionti, A. E.; Lipeles, E.; Lipniacka, A.; Lisovyi, M.; Liss, T. M.; Lister, A.; Litke, A. M.; Liu, B.; Liu, H.; Liu, H.; Liu, J. K. K.; Liu, J.; Liu, J. B.; Liu, K.; Liu, L.; Liu, M.; Liu, Y. L.; Liu, Y.; Livan, M.; Lleres, A.; Llorente Merino, J.; Lloyd, S. L.; Lo, C. Y.; Sterzo, F. Lo; Lobodzinska, E. M.; Loch, P.; Loebinger, F. K.; Loew, K. M.; Loginov, A.; Lohse, T.; Lohwasser, K.; Lokajicek, M.; Long, B. A.; Long, J. D.; Long, R. E.; Longo, L.; Looper, K. A.; Lopez, J. A.; Lopez Mateos, D.; Lopez Paz, I.; Lopez Solis, A.; Lorenz, J.; Lorenzo Martinez, N.; Losada, M.; Lösel, P. J.; Lou, X.; Lounis, A.; Love, J.; Love, P. A.; Lu, H.; Lu, N.; Lu, Y. J.; Lubatti, H. J.; Luci, C.; Lucotte, A.; Luedtke, C.; Luehring, F.; Lukas, W.; Luminari, L.; Lundberg, O.; Lund-Jensen, B.; Luzi, P. M.; Lynn, D.; Lysak, R.; Lytken, E.; Lyubushkin, V.; Ma, H.; Ma, L. L.; Ma, Y.; Maccarrone, G.; Macchiolo, A.; Macdonald, C. M.; Maček, B.; Machado Miguens, J.; Madaffari, D.; Madar, R.; Maddocks, H. J.; Mader, W. F.; Madsen, A.; Maeda, J.; Maeland, S.; Maeno, T.; Maevskiy, A. S.; Magradze, E.; Mahlstedt, J.; Maiani, C.; Maidantchik, C.; Maier, A. A.; Maier, T.; Maio, A.; Majewski, S.; Makida, Y.; Makovec, N.; Malaescu, B.; Malecki, Pa.; Maleev, V. P.; Malek, F.; Mallik, U.; Malon, D.; Malone, C.; Maltezos, S.; Malyukov, S.; Mamuzic, J.; Mancini, G.; Mandelli, L.; Mandić, I.; Maneira, J.; Manhaes de Andrade Filho, L.; Manjarres Ramos, J.; Mann, A.; Manousos, A.; Mansoulie, B.; Mansour, J. D.; Mantifel, R.; Mantoani, M.; Manzoni, S.; Mapelli, L.; Marceca, G.; March, L.; Marchese, L.; Marchiori, G.; Marcisovsky, M.; Marjanovic, M.; Marley, D. E.; Marroquim, F.; Marsden, S. P.; Marshall, Z.; Martensson, M. U. F.; Marti-Garcia, S.; Martin, C. B.; Martin, T. A.; Martin, V. J.; Martin dit Latour, B.; Martinez, M.; Martinez Outschoorn, V. I.; Martin-Haugh, S.; Martoiu, V. S.; Martyniuk, A. C.; Marzin, A.; Masetti, L.; Mashimo, T.; Mashinistov, R.; Masik, J.; Maslennikov, A. L.; Massa, L.; Mastrandrea, P.; Mastroberardino, A.; Masubuchi, T.; Mättig, P.; Maurer, J.; Maxfield, S. J.; Maximov, D. A.; Mazini, R.; Maznas, I.; Mazza, S. M.; Mc Fadden, N. C.; Mc Goldrick, G.; Mc Kee, S. P.; McCarn, A.; McCarthy, R. L.; McCarthy, T. G.; McClymont, L. I.; McDonald, E. F.; Mcfayden, J. A.; Mchedlidze, G.; McMahon, S. J.; McNamara, P. C.; McPherson, R. A.; Meehan, S.; Megy, T. J.; Mehlhase, S.; Mehta, A.; Meideck, T.; Meier, K.; Meineck, C.; Meirose, B.; Melini, D.; Mellado Garcia, B. R.; Melo, M.; Meloni, F.; Menary, S. B.; Meng, L.; Meng, X. T.; Mengarelli, A.; Menke, S.; Meoni, E.; Mergelmeyer, S.; Mermod, P.; Merola, L.; Meroni, C.; Merritt, F. S.; Messina, A.; Metcalfe, J.; Mete, A. S.; Meyer, C.; Meyer, J-P.; Meyer, J.; Meyer Zu Theenhausen, H.; Miano, F.; Middleton, R. P.; Miglioranzi, S.; Mijović, L.; Mikenberg, G.; Mikestikova, M.; Mikuž, M.; Milesi, M.; Milic, A.; Miller, D. W.; Mills, C.; Milov, A.; Milstead, D. A.; Minaenko, A. A.; Minami, Y.; Minashvili, I. A.; Mincer, A. I.; Mindur, B.; Mineev, M.; Minegishi, Y.; Ming, Y.; Mir, L. M.; Mistry, K. P.; Mitani, T.; Mitrevski, J.; Mitsou, V. A.; Miucci, A.; Miyagawa, P. S.; Mizukami, A.; Mjörnmark, J. U.; Mlynarikova, M.; Moa, T.; Mochizuki, K.; Mogg, P.; Mohapatra, S.; Molander, S.; Moles-Valls, R.; Monden, R.; Mondragon, M. C.; Mönig, K.; Monk, J.; Monnier, E.; Montalbano, A.; Montejo Berlingen, J.; Monticelli, F.; Monzani, S.; Moore, R. W.; Morange, N.; Moreno, D.; Moreno Llácer, M.; Morettini, P.; Morgenstern, S.; Mori, D.; Mori, T.; Morii, M.; Morinaga, M.; Morisbak, V.; Morley, A. K.; Mornacchi, G.; Morris, J. D.; Morvaj, L.; Moschovakos, P.; Mosidze, M.; Moss, H. J.; Moss, J.; Motohashi, K.; Mount, R.; Mountricha, E.; Moyse, E. J. W.; Muanza, S.; Mudd, R. D.; Mueller, F.; Mueller, J.; Mueller, R. S. P.; Muenstermann, D.; Mullen, P.; Mullier, G. A.; Munoz Sanchez, F. J.; Murray, W. J.; Musheghyan, H.; Muškinja, M.; Myagkov, A. G.; Myska, M.; Nachman, B. P.; Nackenhorst, O.; Nagai, K.; Nagai, R.; Nagano, K.; Nagasaka, Y.; Nagata, K.; Nagel, M.; Nagy, E.; Nairz, A. M.; Nakahama, Y.; Nakamura, K.; Nakamura, T.; Nakano, I.; Naranjo Garcia, R. F.; Narayan, R.; Narrias Villar, D. I.; Naryshkin, I.; Naumann, T.; Navarro, G.; Nayyar, R.; Neal, H. A.; Nechaeva, P. Yu.; Neep, T. J.; Negri, A.; Negrini, M.; Nektarijevic, S.; Nellist, C.; Nelson, A.; Nelson, M. E.; Nemecek, S.; Nemethy, P.; Nepomuceno, A. A.; Nessi, M.; Neubauer, M. S.; Neumann, M.; Newman, P. R.; Ng, T. Y.; Nguyen Manh, T.; Nickerson, R. B.; Nicolaidou, R.; Nielsen, J.; Nikolaenko, V.; Nikolic-Audit, I.; Nikolopoulos, K.; Nilsen, J. K.; Nilsson, P.; Ninomiya, Y.; Nisati, A.; Nishu, N.; Nisius, R.; Nobe, T.; Noguchi, Y.; Nomachi, M.; Nomidis, I.; Nomura, M. A.; Nooney, T.; Nordberg, M.; Norjoharuddeen, N.; Novgorodova, O.; Nowak, S.; Nozaki, M.; Nozka, L.; Ntekas, K.; Nurse, E.; Nuti, F.; O’connor, K.; O’Neil, D. C.; O’Rourke, A. A.; O’Shea, V.; Oakham, F. G.; Oberlack, H.; Obermann, T.; Ocariz, J.; Ochi, A.; Ochoa, I.; Ochoa-Ricoux, J. P.; Oda, S.; Odaka, S.; Ogren, H.; Oh, A.; Oh, S. H.; Ohm, C. C.; Ohman, H.; Oide, H.; Okawa, H.; Okumura, Y.; Okuyama, T.; Olariu, A.; Oleiro Seabra, L. F.; Olivares Pino, S. A.; Oliveira Damazio, D.; Olszewski, A.; Olszowska, J.; Onofre, A.; Onogi, K.; Onyisi, P. U. E.; Oreglia, M. J.; Oren, Y.; Orestano, D.; Orlando, N.; Orr, R. S.; Osculati, B.; Ospanov, R.; Otero y Garzon, G.; Otono, H.; Ouchrif, M.; Ould-Saada, F.; Ouraou, A.; Oussoren, K. P.; Ouyang, Q.; Owen, M.; Owen, R. E.; Ozcan, V. E.; Ozturk, N.; Pachal, K.; Pacheco Pages, A.; Pacheco Rodriguez, L.; Padilla Aranda, C.; Pagan Griso, S.; Paganini, M.; Paige, F.; Pais, P.; Palacino, G.; Palazzo, S.; Palestini, S.; Palka, M.; Pallin, D.; St. Panagiotopoulou, E.; Panagoulias, I.; Pandini, C. E.; Panduro Vazquez, J. G.; Pani, P.; Panitkin, S.; Pantea, D.; Paolozzi, L.; Papadopoulou, Th. D.; Papageorgiou, K.; Paramonov, A.; Paredes Hernandez, D.; Parker, A. J.; Parker, M. A.; Parker, K. A.; Parodi, F.; Parsons, J. A.; Parzefall, U.; Pascuzzi, V. R.; Pasner, J. M.; Pasqualucci, E.; Passaggio, S.; Pastore, Fr.; Pataraia, S.; Pater, J. R.; Pauly, T.; Pearce, J.; Pearson, B.; Pedraza Lopez, S.; Pedro, R.; Peleganchuk, S. V.; Penc, O.; Peng, C.; Peng, H.; Penwell, J.; Peralva, B. S.; Perego, M. M.; Perepelitsa, D. V.; Perini, L.; Pernegger, H.; Perrella, S.; Peschke, R.; Peshekhonov, V. D.; Peters, K.; Peters, R. F. Y.; Petersen, B. A.; Petersen, T. C.; Petit, E.; Petridis, A.; Petridou, C.; Petroff, P.; Petrolo, E.; Petrov, M.; Petrucci, F.; Pettersson, N. E.; Peyaud, A.; Pezoa, R.; Phillips, F. H.; Phillips, P. W.; Piacquadio, G.; Pianori, E.; Picazio, A.; Piccaro, E.; Pickering, M. A.; Piegaia, R.; Pilcher, J. E.; Pilkington, A. D.; Pin, A. W. J.; Pinamonti, M.; Pinfold, J. L.; Pirumov, H.; Pitt, M.; Plazak, L.; Pleier, M. -A.; Pleskot, V.; Plotnikova, E.; Pluth, D.; Podberezko, P.; Poettgen, R.; Poggi, R.; Poggioli, L.; Pohl, D.; Polesello, G.; Poley, A.; Policicchio, A.; Polifka, R.; Polini, A.; Pollard, C. S.; Polychronakos, V.; Pommès, K.; Ponomarenko, D.; Pontecorvo, L.; Pope, B. G.; Popeneciu, G. A.; Poppleton, A.; Pospisil, S.; Potamianos, K.; Potrap, I. N.; Potter, C. J.; Poulard, G.; Poveda, J.; Pozo Astigarraga, M. E.; Pralavorio, P.; Pranko, A.; Prell, S.; Price, D.; Price, L. E.; Primavera, M.; Prince, S.; Proklova, N.; Prokofiev, K.; Prokoshin, F.; Protopopescu, S.; Proudfoot, J.; Przybycien, M.; Puddu, D.; Puri, A.; Puzo, P.; Qian, J.; Qin, G.; Qin, Y.; Quadt, A.; Queitsch-Maitland, M.; Quilty, D.; Raddum, S.; Radeka, V.; Radescu, V.; Radhakrishnan, S. K.; Radloff, P.; Rados, P.; Ragusa, F.; Rahal, G.; Raine, J. A.; Rajagopalan, S.; Rangel-Smith, C.; Rashid, T.; Ratti, M. G.; Rauch, D. M.; Rauscher, F.; Rave, S.; Ravenscroft, T.; Ravinovich, I.; Rawling, J. H.; Raymond, M.; Read, A. L.; Readioff, N. P.; Reale, M.; Rebuzzi, D. M.; Redelbach, A.; Redlinger, G.; Reece, R.; Reed, R. G.; Reeves, K.; Rehnisch, L.; Reichert, J.; Reiss, A.; Rembser, C.; Ren, H.; Rescigno, M.; Resconi, S.; Resseguie, E. D.; Rettie, S.; Reynolds, E.; Rezanova, O. L.; Reznicek, P.; Rezvani, R.; Richter, R.; Richter, S.; Richter-Was, E.; Ricken, O.; Ridel, M.; Rieck, P.; Riegel, C. J.; Rieger, J.; Rifki, O.; Rijssenbeek, M.; Rimoldi, A.; Rimoldi, M.; Rinaldi, L.; Ristić, B.; Ritsch, E.; Riu, I.; Rizatdinova, F.; Rizvi, E.; Rizzi, C.; Roberts, R. T.; Robertson, S. H.; Robichaud-Veronneau, A.; Robinson, D.; Robinson, J. E. M.; Robson, A.; Rocco, E.; Roda, C.; Rodina, Y.; Rodriguez Bosca, S.; Rodriguez Perez, A.; Rodriguez Rodriguez, D.; Roe, S.; Rogan, C. S.; Røhne, O.; Roloff, J.; Romaniouk, A.; Romano, M.; Romano Saez, S. M.; Romero Adam, E.; Rompotis, N.; Ronzani, M.; Roos, L.; Rosati, S.; Rosbach, K.; Rose, P.; Rosien, N. -A.; Rossetti, V.; Rossi, E.; Rossi, L. P.; Rosten, J. H. N.; Rosten, R.; Rotaru, M.; Roth, I.; Rothberg, J.; Rousseau, D.; Rozanov, A.; Rozen, Y.; Ruan, X.; Rubbo, F.; Rühr, F.; Ruiz-Martinez, A.; Rurikova, Z.; Rusakovich, N. A.; Russell, H. L.; Rutherfoord, J. P.; Ruthmann, N.; Ryabov, Y. F.; Rybar, M.; Rybkin, G.; Ryu, S.; Ryzhov, A.; Rzehorz, G. F.; Saavedra, A. F.; Sabato, G.; Sacerdoti, S.; Sadrozinski, H. F-W.; Sadykov, R.; Safai Tehrani, F.; Saha, P.; Sahinsoy, M.; Saimpert, M.; Saito, M.; Saito, T.; Sakamoto, H.; Sakurai, Y.; Salamanna, G.; Salazar Loyola, J. E.; Salek, D.; Sales De Bruin, P. H.; Salihagic, D.; Salnikov, A.; Salt, J.; Salvatore, D.; Salvatore, F.; Salvucci, A.; Salzburger, A.; Sammel, D.; Sampsonidis, D.; Sampsonidou, D.; Sánchez, J.; Sanchez Martinez, V.; Sanchez Pineda, A.; Sandaker, H.; Sandbach, R. L.; Sander, C. O.; Sandhoff, M.; Sandoval, C.; Sankey, D. P. C.; Sannino, M.; Sansoni, A.; Santoni, C.; Santonico, R.; Santos, H.; Santoyo Castillo, I.; Sapp, K.; Sapronov, A.; Saraiva, J. G.; Sarrazin, B.; Sasaki, O.; Sato, K.; Sauvan, E.; Savage, G.; Savard, P.; Savic, N.; Sawyer, C.; Sawyer, L.; Saxon, J.; Sbarra, C.; Sbrizzi, A.; Scanlon, T.; Scannicchio, D. A.; Scarcella, M.; Scarfone, V.; Schaarschmidt, J.; Schacht, P.; Schachtner, B. M.; Schaefer, D.; Schaefer, L.; Schaefer, R.; Schaeffer, J.; Schaepe, S.; Schaetzel, S.; Schäfer, U.; Schaffer, A. C.; Schaile, D.; Schamberger, R. D.; Scharf, V.; Schegelsky, V. A.; Scheirich, D.; Schernau, M.; Schiavi, C.; Schier, S.; Schildgen, L. K.; Schillo, C.; Schioppa, M.; Schlenker, S.; Schmidt-Sommerfeld, K. R.; Schmieden, K.; Schmitt, C.; Schmitt, S.; Schmitz, S.; Schnoor, U.; Schoeffel, L.; Schoening, A.; Schoenrock, B. D.; Schopf, E.; Schott, M.; Schouwenberg, J. F. P.; Schovancova, J.; Schramm, S.; Schuh, N.; Schulte, A.; Schultens, M. J.; Schultz-Coulon, H. -C.; Schulz, H.; Schumacher, M.; Schumm, B. A.; Schune, Ph.; Schwartzman, A.; Schwarz, T. A.; Schweiger, H.; Schwemling, Ph.; Schwienhorst, R.; Schwindling, J.; Schwindt, T.; Sciandra, A.; Sciolla, G.; Scuri, F.; Scutti, F.; Searcy, J.; Seema, P.; Seidel, S. C.; Seiden, A.; Seixas, J. M.; Sekhniaidze, G.; Sekhon, K.; Sekula, S. J.; Semprini-Cesari, N.; Senkin, S.; Serfon, C.; Serin, L.; Serkin, L.; Sessa, M.; Seuster, R.; Severini, H.; Sfiligoj, T.; Sforza, F.; Sfyrla, A.; Shabalina, E.; Shaikh, N. W.; Shan, L. Y.; Shang, R.; Shank, J. T.; Shapiro, M.; Shatalov, P. B.; Shaw, K.; Shaw, S. M.; Shcherbakova, A.; Shehu, C. Y.; Shen, Y.; Sherwood, P.; Shi, L.; Shimizu, S.; Shimmin, C. O.; Shimojima, M.; Shipsey, I. P. J.; Shirabe, S.; Shiyakova, M.; Shlomi, J.; Shmeleva, A.; Shoaleh Saadi, D.; Shochet, M. J.; Shojaii, S.; Shope, D. R.; Shrestha, S.; Shulga, E.; Shupe, M. A.; Sicho, P.; Sickles, A. M.; Sidebo, P. E.; Sideras Haddad, E.; Sidiropoulou, O.; Sidorov, D.; Sidoti, A.; Siegert, F.; Sijacki, Dj.; Silva, J.; Silverstein, S. B.; Simak, V.; Simic, Lj.; Simion, S.; Simioni, E.; Simmons, B.; Simon, M.; Sinervo, P.; Sinev, N. B.; Sioli, M.; Siragusa, G.; Siral, I.; Sivoklokov, S. Yu.; Sjölin, J.; Skinner, M. B.; Skubic, P.; Slater, M.; Slavicek, T.; Slawinska, M.; Sliwa, K.; Slovak, R.; Smakhtin, V.; Smart, B. H.; Smiesko, J.; Smirnov, N.; Smirnov, S. Yu.; Smirnov, Y.; Smirnova, L. N.; Smirnova, O.; Smith, J. W.; Smith, M. N. K.; Smith, R. W.; Smizanska, M.; Smolek, K.; Snesarev, A. A.; Snyder, I. M.; Snyder, S.; Sobie, R.; Socher, F.; Soffer, A.; Soh, D. A.; Sokhrannyi, G.; Solans Sanchez, C. A.; Solar, M.; Soldatov, E. Yu.; Soldevila, U.; Solodkov, A. A.; Soloshenko, A.; Solovyanov, O. V.; Solovyev, V.; Sommer, P.; Son, H.; Song, H. Y.; Sopczak, A.; Sosa, D.; Sotiropoulou, C. L.; Soualah, R.; Soukharev, A. M.; South, D.; Sowden, B. C.; Spagnolo, S.; Spalla, M.; Spangenberg, M.; Spanò, F.; Sperlich, D.; Spettel, F.; Spieker, T. M.; Spighi, R.; Spigo, G.; Spiller, L. A.; Spousta, M.; St. Denis, R. D.; Stabile, A.; Stamen, R.; Stamm, S.; Stanecka, E.; Stanek, R. W.; Stanescu, C.; Stanitzki, M. M.; Stapnes, S.; Starchenko, E. A.; Stark, G. H.; Stark, J.; Stark, S. H.; Staroba, P.; Starovoitov, P.; Stärz, S.; Staszewski, R.; Steinberg, P.; Stelzer, B.; Stelzer, H. J.; Stelzer-Chilton, O.; Stenzel, H.; Stewart, G. A.; Stockton, M. C.; Stoebe, M.; Stoicea, G.; Stolte, P.; Stonjek, S.; Stradling, A. R.; Straessner, A.; Stramaglia, M. E.; Strandberg, J.; Strandberg, S.; Strandlie, A.; Strauss, M.; Strizenec, P.; Ströhmer, R.; Strom, D. M.; Stroynowski, R.; Strubig, A.; Stucci, S. A.; Stugu, B.; Styles, N. A.; Su, D.; Su, J.; Suchek, S.; Sugaya, Y.; Suk, M.; Sulin, V. V.; Sultansoy, S.; Sumida, T.; Sun, S.; Sun, X.; Suruliz, K.; Suster, C. J. E.; Sutton, M. R.; Suzuki, S.; Svatos, M.; Swiatlowski, M.; Swift, S. P.; Sykora, I.; Sykora, T.; Ta, D.; Tackmann, K.; Taenzer, J.; Taffard, A.; Tafirout, R.; Taiblum, N.; Takai, H.; Takashima, R.; Takeshita, T.; Takubo, Y.; Talby, M.; Talyshev, A. A.; Tanaka, J.; Tanaka, M.; Tanaka, R.; Tanaka, S.; Tanioka, R.; Tannenwald, B. B.; Tapia Araya, S.; Tapprogge, S.; Tarem, S.; Tartarelli, G. F.; Tas, P.; Tasevsky, M.; Tashiro, T.; Tassi, E.; Tavares Delgado, A.; Tayalati, Y.; Taylor, A. C.; Taylor, G. N.; Taylor, P. T. E.; Taylor, W.; Teixeira-Dias, P.; Temple, D.; Ten Kate, H.; Teng, P. K.; Teoh, J. J.; Tepel, F.; Terada, S.; Terashi, K.; Terron, J.; Terzo, S.; Testa, M.; Teuscher, R. J.; Theveneaux-Pelzer, T.; Thomas, J. P.; Thomas-Wilsker, J.; Thompson, P. D.; Thompson, A. S.; Thomsen, L. A.; Thomson, E.; Tibbetts, M. J.; Ticse Torres, R. E.; Tikhomirov, V. O.; Tikhonov, Yu. A.; Timoshenko, S.; Tipton, P.; Tisserant, S.; Todome, K.; Todorova-Nova, S.; Tojo, J.; Tokár, S.; Tokushuku, K.; Tolley, E.; Tomlinson, L.; Tomoto, M.; Tompkins, L.; Toms, K.; Tong, B.; Tornambe, P.; Torrence, E.; Torres, H.; Torró Pastor, E.; Toth, J.; Touchard, F.; Tovey, D. R.; Treado, C. J.; Trefzger, T.; Tresoldi, F.; Tricoli, A.; Trigger, I. M.; Trincaz-Duvoid, S.; Tripiana, M. F.; Trischuk, W.; Trocmé, B.; Trofymov, A.; Troncon, C.; Trottier-McDonald, M.; Trovatelli, M.; Truong, L.; Trzebinski, M.; Trzupek, A.; Tsang, K. W.; Tseng, J. C-L.; Tsiareshka, P. V.; Tsipolitis, G.; Tsirintanis, N.; Tsiskaridze, S.; Tsiskaridze, V.; Tskhadadze, E. G.; Tsui, K. M.; Tsukerman, I. I.; Tsulaia, V.; Tsuno, S.; Tsybychev, D.; Tu, Y.; Tudorache, A.; Tudorache, V.; Tulbure, T. T.; Tuna, A. N.; Tupputi, S. A.; Turchikhin, S.; Turgeman, D.; Turk Cakir, I.; Turra, R.; Tuts, P. M.; Ucchielli, G.; Ueda, I.; Ughetto, M.; Ukegawa, F.; Unal, G.; Undrus, A.; Unel, G.; Ungaro, F. C.; Unno, Y.; Unverdorben, C.; Urban, J.; Urquijo, P.; Urrejola, P.; Usai, G.; Usui, J.; Vacavant, L.; Vacek, V.; Vachon, B.; Valderanis, C.; Valdes Santurio, E.; Valentinetti, S.; Valero, A.; Valéry, L.; Valkar, S.; Vallier, A.; Valls Ferrer, J. A.; Van Den Wollenberg, W.; van der Graaf, H.; van Eldik, N.; van Gemmeren, P.; Van Nieuwkoop, J.; van Vulpen, I.; van Woerden, M. C.; Vanadia, M.; Vandelli, W.; Vanguri, R.; Vaniachine, A.; Vankov, P.; Vardanyan, G.; Vari, R.; Varnes, E. W.; Varni, C.; Varol, T.; Varouchas, D.; Vartapetian, A.; Varvell, K. E.; Vasquez, J. G.; Vasquez, G. A.; Vazeille, F.; Vazquez Schroeder, T.; Veatch, J.; Veeraraghavan, V.; Veloce, L. M.; Veloso, F.; Veneziano, S.; Ventura, A.; Venturi, M.; Venturi, N.; Venturini, A.; Vercesi, V.; Verducci, M.; Verkerke, W.; Vermeulen, J. C.; Vetterli, M. C.; Viaux Maira, N.; Viazlo, O.; Vichou, I.; Vickey, T.; Vickey Boeriu, O. E.; Viehhauser, G. H. A.; Viel, S.; Vigani, L.; Villa, M.; Villaplana Perez, M.; Vilucchi, E.; Vincter, M. G.; Vinogradov, V. B.; Vishwakarma, A.; Vittori, C.; Vivarelli, I.; Vlachos, S.; Vlasak, M.; Vogel, M.; Vokac, P.; Volpi, G.; von der Schmitt, H.; von Toerne, E.; Vorobel, V.; Vorobev, K.; Vos, M.; Voss, R.; Vossebeld, J. H.; Vranjes, N.; Vranjes Milosavljevic, M.; Vrba, V.; Vreeswijk, M.; Vuillermet, R.; Vukotic, I.; Wagner, P.; Wagner, W.; Wagner-Kuhr, J.; Wahlberg, H.; Wahrmund, S.; Wakabayashi, J.; Walder, J.; Walker, R.; Walkowiak, W.; Wallangen, V.; Wang, C.; Wang, C.; Wang, F.; Wang, H.; Wang, H.; Wang, J.; Wang, J.; Wang, Q.; Wang, R.; Wang, S. M.; Wang, T.; Wang, W.; Wang, W.; Wang, Z.; Wanotayaroj, C.; Warburton, A.; Ward, C. P.; Wardrope, D. R.; Washbrook, A.; Watkins, P. M.; Watson, A. T.; Watson, M. F.; Watts, G.; Watts, S.; Waugh, B. M.; Webb, A. F.; Webb, S.; Weber, M. S.; Weber, S. W.; Weber, S. A.; Webster, J. S.; Weidberg, A. R.; Weinert, B.; Weingarten, J.; Weiser, C.; Weits, H.; Wells, P. S.; Wenaus, T.; Wengler, T.; Wenig, S.; Wermes, N.; Werner, M. D.; Werner, P.; Wessels, M.; Whalen, K.; Whallon, N. L.; Wharton, A. M.; White, A.; White, M. J.; White, R.; Whiteson, D.; Wickens, F. J.; Wiedenmann, W.; Wielers, M.; Wiglesworth, C.; Wiik-Fuchs, L. A. M.; Wildauer, A.; Wilk, F.; Wilkens, H. G.; Williams, H. H.; Williams, S.; Willis, C.; Willocq, S.; Wilson, J. A.; Wingerter-Seez, I.; Winkels, E.; Winklmeier, F.; Winston, O. J.; Winter, B. T.; Wittgen, M.; Wobisch, M.; Wolf, T. M. H.; Wolff, R.; Wolter, M. W.; Wolters, H.; Wong, V. W. S.; Worm, S. D.; Wosiek, B. K.; Wotschack, J.; Wozniak, K. W.; Wu, M.; Wu, S. L.; Wu, X.; Wu, Y.; Wyatt, T. R.; Wynne, B. M.; Xella, S.; Xi, Z.; Xia, L.; Xu, D.; Xu, L.; Yabsley, B.; Yacoob, S.; Yamaguchi, D.; Yamaguchi, Y.; Yamamoto, A.; Yamamoto, S.; Yamanaka, T.; Yamauchi, K.; Yamazaki, Y.; Yan, Z.; Yang, H.; Yang, H.; Yang, Y.; Yang, Z.; Yao, W-M.; Yap, Y. C.; Yasu, Y.; Yatsenko, E.; Yau Wong, K. H.; Ye, J.; Ye, S.; Yeletskikh, I.; Yigitbasi, E.; Yildirim, E.; Yorita, K.; Yoshihara, K.; Young, C.; Young, C. J. S.; Yu, D. R.; Yu, J.; Yu, J.; Yuen, S. P. Y.; Yusuff, I.; Zabinski, B.; Zacharis, G.; Zaidan, R.; Zaitsev, A. M.; Zakharchuk, N.; Zalieckas, J.; Zaman, A.; Zambito, S.; Zanzi, D.; Zeitnitz, C.; Zeman, M.; Zemla, A.; Zeng, J. C.; Zeng, Q.; Zenin, O.; Ženiš, T.; Zerwas, D.; Zhang, D.; Zhang, F.; Zhang, G.; Zhang, H.; Zhang, J.; Zhang, L.; Zhang, L.; Zhang, M.; Zhang, R.; Zhang, R.; Zhang, X.; Zhang, Y.; Zhang, Z.; Zhao, X.; Zhao, Y.; Zhao, Z.; Zhemchugov, A.; Zhong, J.; Zhou, B.; Zhou, C.; Zhou, L.; Zhou, M.; Zhou, M.; Zhou, N.; Zhu, C. G.; Zhu, H.; Zhu, J.; Zhu, Y.; Zhuang, X.; Zhukov, K.; Zibell, A.; Zieminska, D.; Zimine, N. I.; Zimmermann, C.; Zimmermann, S.; Zinonos, Z.; Zinser, M.; Ziolkowski, M.; Živković, L.; Zobernig, G.; Zoccoli, A.; Zou, R.; zur Nedden, M.; Zwalinski, L.

    2017-07-21

    The production of a Z boson and a photon in association with a high-mass dijet system is studied using 20.2 fb-1 of proton-proton collision data at a centre-of-mass energy of $\\sqrt{s}$ = 8 TeV recorded with the ATLAS detector in 2012 at the Large Hadron Collider. Final states with a photon and a Z boson decaying into a pair of either electrons, muons, or neutrinos are analysed. Electroweak and total pp → Zγjj cross-sections are extracted in two fiducial regions with different sensitivities to electroweak production processes. Quartic couplings of vector bosons are studied in regions of phase space with an enhanced contribution from pure electroweak production, sensitive to vector-boson scattering processes VV → Zγ. Finally, no deviations from Standard Model predictions are observed and constraints are placed on anomalous couplings parameterized by higher-dimensional operators using effective field theory.

  20. Measurements of differential cross sections of top quark pair production in association with jets in ${pp}$ collisions at $\\sqrt{s}$=13 TeV using the ATLAS detector

    CERN Document Server

    Aaboud, Morad; ATLAS Collaboration; Abbott, Brad; Abdinov, Ovsat; Abeloos, Baptiste; Abidi, Syed Haider; AbouZeid, Ossama; Abraham, Nicola; Abramowicz, Halina; Abreu, Henso; Abulaiti, Yiming; Acharya, Bobby Samir; Adachi, Shunsuke; Adamczyk, Leszek; Adelman, Jahred; Adersberger, Michael; Adye, Tim; Affolder, Tony; Afik, Yoav; Agheorghiesei, Catalin; Aguilar-Saavedra, Juan Antonio; Ahlen, Steven; Ahmadov, Faig; Aielli, Giulio; Akatsuka, Shunichi; Akerstedt, Henrik; {\\AA}kesson, Torsten Paul Ake; Akilli, Ece; Akimov, Andrei; Alberghi, Gian Luigi; Albert, Justin; Albicocco, Pietro; Alconada Verzini, Maria Josefina; Alderweireldt, Sara; Aleksa, Martin; Aleksandrov, Igor; Alexa, Calin; Alexander, Gideon; Alexopoulos, Theodoros; Alhroob, Muhammad; Ali, Babar; Aliev, Malik; Alimonti, Gianluca; Alison, John; Alkire, Steven Patrick; Allbrooke, Benedict; Allen, Benjamin William; Allport, Phillip; Aloisio, Alberto; Alonso, Alejandro; Alonso, Francisco; Alpigiani, Cristiano; Alshehri, Azzah Aziz; Alstaty, Mahmoud; Alvarez Gonzalez, Barbara; \\'{A}lvarez Piqueras, Dami\\'{a}n; Alviggi, Mariagrazia; Amadio, Brian Thomas; Amaral Coutinho, Yara; Amelung, Christoph; Amidei, Dante; Amor Dos Santos, Susana Patricia; Amoroso, Simone; Anastopoulos, Christos; Ancu, Lucian Stefan; Andari, Nansi; Andeen, Timothy; Anders, Christoph Falk; Anders, John Kenneth; Anderson, Kelby; Andreazza, Attilio; Andrei, George Victor; Angelidakis, Stylianos; Angelozzi, Ivan; Angerami, Aaron; Anisenkov, Alexey; Anjos, Nuno; Annovi, Alberto; Antel, Claire; Antonelli, Mario; Antonov, Alexey; Antrim, Daniel Joseph; Anulli, Fabio; Aoki, Masato; Aperio Bella, Ludovica; Arabidze, Giorgi; Arai, Yasuo; Araque, Juan Pedro; Araujo Ferraz, Victor; Arce, Ayana; Ardell, Rose Elisabeth; Arduh, Francisco Anuar; Arguin, Jean-Francois; Argyropoulos, Spyridon; Arik, Metin; Armbruster, Aaron James; Armitage, Lewis James; Arnaez, Olivier; Arnold, Hannah; Arratia, Miguel; Arslan, Ozan; Artamonov, Andrei; Artoni, Giacomo; Artz, Sebastian; Asai, Shoji; Asbah, Nedaa; Ashkenazi, Adi; Asquith, Lily; Assamagan, Ketevi; Astalos, Robert; Atkinson, Markus; Atlay, Naim Bora; Augsten, Kamil; Avolio, Giuseppe; Axen, Bradley; Ayoub, Mohamad Kassem; Azuelos, Georges; Baas, Alessandra; Baca, Matthew John; Bachacou, Henri; Bachas, Konstantinos; Backes, Moritz; Bagnaia, Paolo; Bahmani, Marzieh; Bahrasemani, Sina; Baines, John; Bajic, Milena; Baker, Oliver Keith; Bakker, Pepijn Johannes; Baldin, Evgenii; Balek, Petr; Balli, Fabrice; Balunas, William Keaton; Banas, Elzbieta; Bandyopadhyay, Anjishnu; Banerjee, Swagato; Bannoura, Arwa A E; Barak, Liron; Barberio, Elisabetta Luigia; Barberis, Dario; Barbero, Marlon; Barillari, Teresa; Barisits, Martin-Stefan; Barkeloo, Jason Tyler Colt; Barklow, Timothy; Barlow, Nick; Barnes, Sarah Louise; Barnett, Bruce; Barnett, Michael; Barnovska-Blenessy, Zuzana; Baroncelli, Antonio; Barone, Gaetano; Barr, Alan; Barranco Navarro, Laura; Barreiro, Fernando; Barreiro Guimar\\~{a}es da Costa, Jo\\~{a}o; Bartoldus, Rainer; Barton, Adam Edward; Bartos, Pavol; Basalaev, Artem; Bassalat, Ahmed; Bates, Richard; Batista, Santiago Juan; Batley, Richard; Battaglia, Marco; Bauce, Matteo; Bauer, Florian; Bauer, Kevin Thomas; Bawa, Harinder Singh; Beacham, James; Beattie, Michael David; Beau, Tristan; Beauchemin, Pierre-Hugues; Bechtle, Philip; Beck, Hans~Peter; Beck, Helge Christoph; Becker, Kathrin; Becker, Maurice; Becot, Cyril; Beddall, Andrew; Beddall, Ayda; Bednyakov, Vadim; Bedognetti, Matteo; Bee, Christopher; Beermann, Thomas; Begalli, Marcia; Begel, Michael; Behr, Janna Katharina; Bell, Andrew Stuart; Bella, Gideon; Bellagamba, Lorenzo; Bellerive, Alain; Bellomo, Massimiliano; Belotskiy, Konstantin; Beltramello, Olga; Belyaev, Nikita; Benary, Odette; Benchekroun, Driss; Bender, Michael; Benekos, Nektarios; Benhammou, Yan; Benhar Noccioli, Eleonora; Benitez, Jose; Benjamin, Douglas; Benoit, Mathieu; Bensinger, James; Bentvelsen, Stan; Beresford, Lydia; Beretta, Matteo; Berge, David; Bergeaas Kuutmann, Elin; Berger, Nicolas; Bergsten, Laura Jean; Beringer, J\\"urg; Berlendis, Simon; Bernard, Nathan Rogers; Bernardi, Gregorio; Bernius, Catrin; Bernlochner, Florian Urs; Berry, Tracey; Berta, Peter; Bertella, Claudia; Bertoli, Gabriele; Bertram, Iain Alexander; Bertsche, Carolyn; Besjes, Geert-Jan; Bessidskaia Bylund, Olga; Bessner, Martin Florian; Besson, Nathalie; Bethani, Agni; Bethke, Siegfried; Betti, Alessandra; Bevan, Adrian John; Beyer, Julien-christopher; Bianchi, Riccardo-Maria; Biebel, Otmar; Biedermann, Dustin; Bielski, Rafal; Bierwagen, Katharina; Biesuz, Nicolo Vladi; Biglietti, Michela; Billoud, Thomas Remy Victor; Bilokon, Halina; Bindi, Marcello; Bingul, Ahmet; Bini, Cesare; Biondi, Silvia; Bisanz, Tobias; Bittrich, Carsten; Bjergaard, David Martin; Black, James; Black, Kevin; Blair, Robert; Blazek, Tomas; Bloch, Ingo; Blocker, Craig; Blue, Andrew; Blumenschein, Ulrike; Blunier, Sylvain; Bobbink, Gerjan; Bobrovnikov, Victor; Bocchetta, Simona Serena; Bocci, Andrea; Bock, Christopher; Boehler, Michael; Boerner, Daniela; Bogavac, Danijela; Bogdanchikov, Alexander; Bohm, Christian; Boisvert, Veronique; Bokan, Petar; Bold, Tomasz; Boldyrev, Alexey; Bolz, Arthur Eugen; Bomben, Marco; Bona, Marcella; Boonekamp, Maarten; Borisov, Anatoly; Borissov, Guennadi; Bortfeldt, Jonathan; Bortoletto, Daniela; Bortolotto, Valerio; Boscherini, Davide; Bosman, Martine; Bossio Sola, Jonathan David; Boudreau, Joseph; Bouhova-Thacker, Evelina Vassileva; Boumediene, Djamel Eddine; Bourdarios, Claire; Boutle, Sarah Kate; Boveia, Antonio; Boyd, James; Boyko, Igor; Bozson, Adam James; Bracinik, Juraj; Brandt, Andrew; Brandt, Gerhard; Brandt, Oleg; Braren, Frued; Bratzler, Uwe; Brau, Benjamin; Brau, James; Breaden Madden, William Dmitri; Brendlinger, Kurt; Brennan, Amelia Jean; Brenner, Lydia; Brenner, Richard; Bressler, Shikma; Briglin, Daniel Lawrence; Bristow, Timothy Michael; Britton, Dave; Britzger, Daniel; Brochu, Frederic; Brock, Ian; Brock, Raymond; Brooijmans, Gustaaf; Brooks, Timothy; Brooks, William; Brost, Elizabeth; Broughton, James; Bruckman de Renstrom, Pawel; Bruncko, Dusan; Bruni, Alessia; Bruni, Graziano; Bruni, Lucrezia Stella; Bruno, Salvatore; Brunt, Benjamin; Bruschi, Marco; Bruscino, Nello; Bryant, Patrick; Bryngemark, Lene; Buanes, Trygve; Buat, Quentin; Buchholz, Peter; Buckley, Andrew; Budagov, Ioulian; Buehrer, Felix; Bugge, Magnar Kopangen; Bulekov, Oleg; Bullock, Daniel; Burch, Tyler James; Burdin, Sergey; Burgard, Carsten Daniel; Burger, Angela Maria; Burghgrave, Blake; Burka, Klaudia; Burke, Stephen; Burmeister, Ingo; Burr, Jonathan Thomas Peter; B\\"uscher, Daniel; B\\"uscher, Volker; Buschmann, Eric; Bussey, Peter; Butler, John; Buttar, Craig; Butterworth, Jonathan; Butti, Pierfrancesco; Buttinger, William; Buzatu, Adrian; Buzykaev, Aleksey; C-Q, Changqiao; Cabrera Urb\\'an, Susana; Caforio, Davide; Cai, Huacheng; Cairo, Valentina; Cakir, Orhan; Calace, Noemi; Calafiura, Paolo; Calandri, Alessandro; Calderini, Giovanni; Calfayan, Philippe; Callea, Giuseppe; Caloba, Luiz; Calvente Lopez, Sergio; Calvet, David; Calvet, Samuel; Calvet, Thomas Philippe; Camacho Toro, Reina; Camarda, Stefano; Camarri, Paolo; Cameron, David; Caminal Armadans, Roger; Camincher, Clement; Campana, Simone; Campanelli, Mario; Camplani, Alessandra; Campoverde, Angel; Canale, Vincenzo; Cano Bret, Marc; Cantero, Josu; Cao, Tingting; Capeans Garrido, Maria Del Mar; Caprini, Irinel; Caprini, Mihai; Capua, Marcella; Carbone, Ryne Michael; Cardarelli, Roberto; Cardillo, Fabio; Carli, Ina; Carli, Tancredi; Carlino, Gianpaolo; Carlson, Benjamin Taylor; Carminati, Leonardo; Carney, Rebecca; Caron, Sascha; Carquin, Edson; Carr\\'a, Sonia; Carrillo-Montoya, German D; Casadei, Diego; Casado, Maria Pilar; Casha, Albert Francis; Casolino, Mirkoantonio; Casper, David William; Castelijn, Remco; Castillo Gimenez, Victoria; Castro, Nuno Filipe; Catinaccio, Andrea; Catmore, James; Cattai, Ariella; Caudron, Julien; Cavaliere, Viviana; Cavallaro, Emanuele; Cavalli, Donatella; Cavalli-Sforza, Matteo; Cavasinni, Vincenzo; Celebi, Emre; Ceradini, Filippo; Cerda Alberich, Leonor; Santiago Cerqueira, Augusto; Cerri, Alessandro; Cerrito, Lucio; Cerutti, Fabio; Cervelli, Alberto; Cetin, Serkant Ali; Chafaq, Aziz; Chakraborty, Dhiman; Chan, Stephen Kam-wah; Chan, Wing Sheung; Chan, Yat Long; Chang, Philip; Chapman, John Derek; Charlton, Dave; Chau, Chav Chhiv; Chavez Barajas, Carlos Alberto; Che, Siinn; Cheatham, Susan; Chegwidden, Andrew; Chekanov, Sergei; Chekulaev, Sergey; Chelkov, Gueorgui; Chelstowska, Magda Anna; Chen, Cheng; Chen, Chunhui; Chen, Hucheng; Chen, Jing; Chen, Shenjian; Chen, Shion; Chen, Xin; Chen, Ye; Cheng, Hok Chuen; Cheng, Huajie; Cheplakov, Alexander; Cheremushkina, Evgeniya; Cherkaoui El Moursli, Rajaa; Cheu, Elliott; Cheung, Kingman; Chevalier, Laurent; Chiarella, Vitaliano; Chiarelli, Giorgio; Chiodini, Gabriele; Chisholm, Andrew; Chitan, Adrian; Chiu, Yu Him Justin; Chizhov, Mihail; Choi, Kyungeon; Chomont, Arthur Rene; Chouridou, Sofia; Chow, Yun Sang; Christodoulou, Valentinos; Chu, Ming Chung; Chudoba, Jiri; Chuinard, Annabelle Julia; Chwastowski, Janusz; Chytka, Ladislav; Ciftci, Abbas Kenan; Cinca, Diane; Cindro, Vladimir; Cioara, Irina Antonela; Ciocio, Alessandra; Cirotto, Francesco; Citron, Zvi Hirsh; Citterio, Mauro; Ciubancan, Mihai; Clark, Allan G; Clark, Michael; Clark, Philip James; Clarke, Robert; Clement, Christophe; Coadou, Yann; Cobal, Marina; Coccaro, Andrea; Cochran, James H; Colasurdo, Luca; Cole, Brian; Colijn, Auke-Pieter; Collot, Johann; Colombo, Tommaso; Conde Mui\\~no, Patricia; Coniavitis, Elias; Connell, Simon Henry; Connelly, Ian; Constantinescu, Serban; Conti, Geraldine; Conventi, Francesco; Cooper-Sarkar, Amanda; Cormier, Felix; Cormier, Kyle James Read; Corradi, Massimo; Corrigan, Eric Edward; Corriveau, Francois; Cortes-Gonzalez, Arely; Costa, Giuseppe; Costa, Mar\\'ia Jos\\'e; Costanzo, Davide; Cottin, Giovanna; Cowan, Glen; Cox, Brian; Cranmer, Kyle; Crawley, Samuel Joseph; Creager, Rachael; Cree, Graham; Cr\\'ep\\'e-Renaudin, Sabine; Crescioli, Francesco; Cribbs, Wayne Allen; Cristinziani, Markus; Croft, Vince; Crosetti, Giovanni; Cueto, Ana; Cuhadar Donszelmann, Tulay; Cukierman, Aviv Ruben; Cummings, Jane; Curatolo, Maria; C\\'uth, Jakub; Czekierda, Sabina; Czodrowski, Patrick; D'amen, Gabriele; D'Auria, Saverio; D'eramo, Louis; D'Onofrio, Monica; Da Cunha Sargedas De Sousa, Mario Jose; Da Via, Cinzia; Dabrowski, Wladyslaw; Dado, Tomas; Dai, Tiesheng; Dale, Orjan; Dallaire, Frederick; Dallapiccola, Carlo; Dam, Mogens; Dandoy, Jeffrey; Daneri, Maria Florencia; Dang, Nguyen Phuong; Dann, Nicholas Stuart; Danninger, Matthias; Dano Hoffmann, Maria; Dao, Valerio; Darbo, Giovanni; Darmora, Smita; Dassoulas, James; Dattagupta, Aparajita; Daubney, Thomas; Davey, Will; David, Claire; Davidek, Tomas; Davis, Douglas; Davison, Peter; Dawe, Edmund; Dawson, Ian; De, Kaushik; de Asmundis, Riccardo; De Benedetti, Abraham; De Castro, Stefano; De Cecco, Sandro; De Groot, Nicolo; de Jong, Paul; De la Torre, Hector; De Lorenzi, Francesco; De Maria, Antonio; De Pedis, Daniele; De Salvo, Alessandro; De Sanctis, Umberto; De Santo, Antonella; De Vasconcelos Corga, Kevin; De Vivie De Regie, Jean-Baptiste; Debbe, Ramiro; Debenedetti, Chiara; Dedovich, Dmitri; Dehghanian, Nooshin; Deigaard, Ingrid; Del Gaudio, Michela; Del Peso, Jose; Delgove, David; Deliot, Frederic; Delitzsch, Chris Malena; Dell'Acqua, Andrea; Dell'Asta, Lidia; Della Pietra, Massimo; della Volpe, Domenico; Delmastro, Marco; Delporte, Charles; Delsart, Pierre-Antoine; DeMarco, David; Demers, Sarah; Demichev, Mikhail; Demilly, Aurelien; Denisov, Sergey; Denysiuk, Denys; Derendarz, Dominik; Derkaoui, Jamal Eddine; Derue, Frederic; Dervan, Paul; Desch, Klaus Kurt; Deterre, Cecile; Dette, Karola; Devesa, Maria Roberta; Deviveiros, Pier-Olivier; Dewhurst, Alastair; Dhaliwal, Saminder; Di Bello, Francesco Armando; Di Ciaccio, Anna; Di Ciaccio, Lucia; Di Clemente, William Kennedy; Di Donato, Camilla; Di Girolamo, Alessandro; Di Girolamo, Beniamino; Di Micco, Biagio; Di Nardo, Roberto; Di Petrillo, Karri Folan; Di Simone, Andrea; Di Sipio, Riccardo; Di Valentino, David; Diaconu, Cristinel; Diamond, Miriam; Dias, Flavia; Diaz, Marco Aurelio; Dickinson, Jennet; Diehl, Edward; Dietrich, Janet; D\\'iez Cornell, Sergio; Dimitrievska, Aleksandra; Dingfelder, Jochen; Dita, Petre; Dita, Sanda; Dittus, Fridolin; Djama, Fares; Djobava, Tamar; Djuvsland, Julia Isabell; Barros do Vale, Maria Aline; Dobre, Monica; Dodsworth, David; Doglioni, Caterina; Dolejsi, Jiri; Dolezal, Zdenek; Donadelli, Marisilvia; Donati, Simone; Donini, Julien; Dopke, Jens; Doria, Alessandra; Dova, Maria-Teresa; Doyle, Tony; Drechsler, Eric; Dris, Manolis; Du, Yanyan; Duarte-Campderros, Jorge; Dubinin, Filipp; Dubreuil, Arnaud; Duchovni, Ehud; Duckeck, Guenter; Ducourthial, Audrey; Ducu, Otilia Anamaria; Duda, Dominik; Dudarev, Alexey; Dudder, Andreas Christian; Duffield, Emily Marie; Duflot, Laurent; D\\"uhrssen, Michael; Dulsen, Carsten; Dumancic, Mirta; Dumitriu, Ana Elena; Duncan, Anna Kathryn; Dunford, Monica; Duperrin, Arnaud; Duran Yildiz, Hatice; D\\"uren, Michael; Durglishvili, Archil; Duschinger, Dirk; Dutta, Baishali; Duvnjak, Damir; Dyndal, Mateusz; Dziedzic, Bartosz Sebastian; Eckardt, Christoph; Ecker, Katharina Maria; Edgar, Ryan Christopher; Eifert, Till; Eigen, Gerald; Einsweiler, Kevin; Ekelof, Tord; El Kacimi, Mohamed; El Kosseifi, Rima; Ellajosyula, Venugopal; Ellert, Mattias; Elles, Sabine; Ellinghaus, Frank; Elliot, Alison; Ellis, Nicolas; Elmsheuser, Johannes; Elsing, Markus; Emeliyanov, Dmitry; Enari, Yuji; Ennis, Joseph Stanford; Epland, Matthew Berg; Erdmann, Johannes; Ereditato, Antonio; Ernst, Michael; Errede, Steven; Escalier, Marc; Escobar, Carlos; Esposito, Bellisario; Estrada Pastor, Oscar; Etienvre, Anne-Isabelle; Etzion, Erez; Evans, Hal; Ezhilov, Alexey; Ezzi, Mohammed; Fabbri, Federica; Fabbri, Laura; Fabiani, Veronica; Facini, Gabriel; Fakhrutdinov, Rinat; Falciano, Speranza; Falla, Rebecca Jane; Faltova, Jana; Fang, Yaquan; Fanti, Marcello; Farbin, Amir; Farilla, Addolorata; Farina, Edoardo Maria; Farooque, Trisha; Farrell, Steven; Farrington, Sinead; Farthouat, Philippe; Fassi, Farida; Fassnacht, Patrick; Fassouliotis, Dimitrios; Faucci Giannelli, Michele; Favareto, Andrea; Fawcett, William James; Fayard, Louis; Fedin, Oleg; Fedorko, Wojciech; Feigl, Simon; Feligioni, Lorenzo; Feng, Cunfeng; Feng, Eric; Feng, Minyu; Fenton, Michael James; Fenyuk, Alexander; Feremenga, Last; Fernandez Martinez, Patricia; Ferrando, James; Ferrari, Arnaud; Ferrari, Pamela; Ferrari, Roberto; Ferreira de Lima, Danilo Enoque; Ferrer, Antonio; Ferrere, Didier; Ferretti, Claudio; Fiedler, Frank; Filip\\v{c}i\\v{c}, Andrej; Filipuzzi, Marco; Filthaut, Frank; Fincke-Keeler, Margret; Finelli, Kevin Daniel; Fiolhais, Miguel; Fiorini, Luca; Fischer, Cora; Fischer, Julia; Fisher, Wade Cameron; Flaschel, Nils; Fleck, Ivor; Fleischmann, Philipp; Fletcher, Rob Roy MacGregor; Flick, Tobias; Flierl, Bernhard Matthias; Flores Castillo, Luis; Forcolin, Giulio Tiziano; Formica, Andrea; F\\"orster, Fabian Alexander; Forti, Alessandra; Foster, Andrew Geoffrey; Fournier, Daniel; Fox, Harald; Fracchia, Silvia; Francavilla, Paolo; Franchini, Matteo; Franchino, Silvia; Francis, David; Franconi, Laura; Franklin, Melissa; Frate, Meghan; Fraternali, Marco; Freeborn, David; Fressard-Batraneanu, Silvia; Freund, Benjamin; Froidevaux, Daniel; Frost, James; Fukunaga, Chikara; Fusayasu, Takahiro; Fuster, Juan; Gabizon, Ofir; Gabrielli, Alessandro; Gabrielli, Andrea; Gach, Grzegorz; Gadatsch, Stefan; Gadomski, Szymon; Gagliardi, Guido; Gagnon, Louis Guillaume; Galea, Cristina; Galhardo, Bruno; Gallas, Elizabeth; Gallop, Bruce; Gallus, Petr; Galster, Gorm Aske Gram Krohn; Gan, KK; Ganguly, Sanmay; Gao, Yanyan; Gao, Yongsheng; Garay Walls, Francisca; Garc\\'ia, Carmen; Garc\\'ia Navarro, Jos\\'e Enrique; Garc\\'ia Pascual, Juan Antonio; Garcia-Sciveres, Maurice; Gardner, Robert; Garelli, Nicoletta; Garonne, Vincent; Gascon Bravo, Alberto; Gasnikova, Ksenia; Gatti, Claudio; Gaudiello, Andrea; Gaudio, Gabriella; Gavrilenko, Igor; Gay, Colin; Gaycken, Goetz; Gazis, Evangelos; Gee, Norman; Geisen, Jannik; Geisen, Marc; Geisler, Manuel Patrice; Gellerstedt, Karl; Gemme, Claudia; Genest, Marie-H\\'el\\`ene; Geng, Cong; Gentile, Simonetta; Gentsos, Christos; George, Simon; Gerbaudo, Davide; Ge\\ss{}ner, Gregor; Ghasemi, Sara; Ghneimat, Mazuza; Giacobbe, Benedetto; Giagu, Stefano; Giangiacomi, Nico; Giannetti, Paola; Gibson, Stephen; Gignac, Matthew; Gilchriese, Murdock; Gillberg, Dag; Gilles, Geoffrey; Gingrich, Douglas; Giordani, MarioPaolo; Giorgi, Filippo Maria; Giraud, Pierre-Francois; Giromini, Paolo; Giugliarelli, Gilberto; Giugni, Danilo; Giuli, Francesco; Giulini, Maddalena; Gjelsten, B{\\o}rge Kile; Gkaitatzis, Stamatios; Gkialas, Ioannis; Gkougkousis, Evangelos Leonidas; Gkountoumis, Panagiotis; Gladilin, Leonid; Glasman, Claudia; Glatzer, Julian; Glaysher, Paul; Glazov, Alexandre; Goblirsch-Kolb, Maximilian; Godlewski, Jan; Goldfarb, Steven; Golling, Tobias; Golubkov, Dmitry; Gomes, Agostinho; Gon\\c calo, Ricardo; Goncalves Gama, Rafael; Goncalves Pinto Firmino Da Costa, Joao; Gonella, Giulia; Gonella, Laura; Gongadze, Alexi; Gonnella, Francesco; Gonski, Julia; Gonz\\'alez de la Hoz, Santiago; Gonzalez-Sevilla, Sergio; Goossens, Luc; Gorbounov, Petr Andreevich; Gordon, Howard; Gorini, Benedetto; Gorini, Edoardo; Gori\\v{s}ek, Andrej; Goshaw, Alfred; G\\"ossling, Claus; Gostkin, Mikhail Ivanovitch; Gottardo, Carlo Alberto; Goudet, Christophe Raymond; Goujdami, Driss; Goussiou, Anna; Govender, Nicolin; Gozani, Eitan; Grabowska-Bold, Iwona; Gradin, Per Olov Joakim; Graham, Emily Charlotte; Gramling, Johanna; Gramstad, Eirik; Grancagnolo, Sergio; Gratchev, Vadim; Gravila, Paul Mircea; Gray, Chloe; Gray, Heather; Greenwood, Zeno Dixon; Grefe, Christian; Gregersen, Kristian; Gregor, Ingrid-Maria; Grenier, Philippe; Grevtsov, Kirill; Griffiths, Justin; Grillo, Alexander; Grimm, Kathryn; Grinstein, Sebastian; Gris, Philippe Luc Yves; Grivaz, Jean-Francois; Groh, Sabrina; Gross, Eilam; Grosse-Knetter, Joern; Grossi, Giulio Cornelio; Grout, Zara Jane; Grummer, Aidan; Guan, Liang; Guan, Wen; Guenther, Jaroslav; Guescini, Francesco; Guest, Daniel; Gueta, Orel; Gui, Bin; Guido, Elisa; Guillemin, Thibault; Guindon, Stefan; Gul, Umar; Gumpert, Christian; Guo, Jun; Guo, Wen; Guo, Yicheng; Gupta, Ruchi; Gurbuz, Saime; Gustavino, Giuliano; Gutelman, Benjamin Jacque; Gutierrez, Phillip; Gutierrez Ortiz, Nicolas Gilberto; Gutschow, Christian; Guyot, Claude; Guzik, Marcin Pawel; Gwenlan, Claire; Gwilliam, Carl; Haas, Andy; Haber, Carl; Hadavand, Haleh Khani; Haddad, Nacim; Hadef, Asma; Hageb\\"ock, Stephan; Hagihara, Mutsuto; Hakobyan, Hrachya; Haleem, Mahsana; Haley, Joseph; Halladjian, Garabed; Hallewell, Gregory David; Hamacher, Klaus; Hamal, Petr; Hamano, Kenji; Hamilton, Andrew; Hamity, Guillermo Nicolas; Hamnett, Phillip George; Han, Kunlin; Han, Liang; Han, Shuo; Hanagaki, Kazunori; Hanawa, Keita; Hance, Michael; Handl, David Michael; Haney, Bijan; Hanke, Paul; Hansen, J{\\o}rgen Beck; Hansen, Jorn Dines; Hansen, Maike Christina; Hansen, Peter Henrik; Hara, Kazuhiko; Hard, Andrew; Harenberg, Torsten; Hariri, Faten; Harkusha, Siarhei; Harrison, Paul Fraser; Hartmann, Nikolai Marcel; Hasegawa, Yoji; Hasib, Ahmed; Hassani, Samira; Haug, Sigve; Hauser, Reiner; Hauswald, Lorenz; Havener, Laura Brittany; Havranek, Miroslav; Hawkes, Christopher; Hawkings, Richard John; Hayden, Daniel; Hays, Chris; Hays, Jonathan Michael; Hayward, Helen; Haywood, Stephen; Heck, Tobias; Hedberg, Vincent; Heelan, Louise; Heer, Sebastian; Heidegger, Kim Katrin; Heim, Sarah; Heim, Timon; Heinemann, Beate; Heinrich, Jochen Jens; Heinrich, Lukas; Heinz, Christian; Hejbal, Jiri; Helary, Louis; Held, Alexander; Hellman, Sten; Helsens, Clement; Henderson, Robert; Heng, Yang; Henkelmann, Steffen; Henriques Correia, Ana Maria; Henrot-Versille, Sophie; Herbert, Geoffrey Henry; Herde, Hannah; Herget, Verena; Hern\\'andez Jim\\'enez, Yesenia; Herr, Holger; Herten, Gregor; Hertenberger, Ralf; Hervas, Luis; Herwig, Theodor Christian; Hesketh, Gavin Grant; Hessey, Nigel; Hetherly, Jeffrey Wayne; Higashino, Satoshi; Hig\\'on-Rodriguez, Emilio; Hildebrand, Kevin; Hill, Ewan; Hill, John; Hiller, Karl Heinz; Hillier, Stephen; Hils, Maximilian; Hinchliffe, Ian; Hirose, Minoru; Hirschbuehl, Dominic; Hiti, Bojan; Hladik, Ondrej; Hlaluku, Dingane Reward; Hoad, Xanthe; Hobbs, John; Hod, Noam; Hodgkinson, Mark; Hodgson, Paul; Hoecker, Andreas; Hoeferkamp, Martin; Hoenig, Friedrich; Hohn, David; Holmes, Tova Ray; Holzbock, Michael; Homann, Michael; Honda, Shunsuke; Honda, Takuya; Hong, Tae Min; Hooberman, Benjamin Henry; Hopkins, Walter; Horii, Yasuyuki; Horton, Arthur James; Hostachy, Jean-Yves; Hostiuc, Alexandru; Hou, Suen; Hoummada, Abdeslam; Howarth, James; Hoya, Joaquin; Hrabovsky, Miroslav; Hrdinka, Julia; Hristova, Ivana; Hrivnac, Julius; Hryn'ova, Tetiana; Hrynevich, Aliaksei; Hsu, Pai-hsien Jennifer; Hsu, Shih-Chieh; Hu, Qipeng; Hu, Shuyang; Huang, Yanping; Hubacek, Zdenek; Hubaut, Fabrice; Huegging, Fabian; Huffman, Todd Brian; Hughes, Emlyn; Huhtinen, Mika; Hunter, Robert Francis Holub; Huo, Peng; Huseynov, Nazim; Huston, Joey; Huth, John; Hyneman, Rachel; Iacobucci, Giuseppe; Iakovidis, Georgios; Ibragimov, Iskander; Iconomidou-Fayard, Lydia; Idrissi, Zineb; Iengo, Paolo; Igonkina, Olga; Iizawa, Tomoya; Ikegami, Yoichi; Ikeno, Masahiro; Ilchenko, Yuriy; Iliadis, Dimitrios; Ilic, Nikolina; Iltzsche, Franziska; Introzzi, Gianluca; Ioannou, Pavlos; Iodice, Mauro; Iordanidou, Kalliopi; Ippolito, Valerio; Isacson, Max Fredrik; Ishijima, Naoki; Ishino, Masaya; Ishitsuka, Masaki; Issever, Cigdem; Istin, Serhat; Ito, Fumiaki; Iturbe Ponce, Julia Mariana; Iuppa, Roberto; Iwasaki, Hiroyuki; Izen, Joseph; Izzo, Vincenzo; Jabbar, Samina; Jackson, Paul; Jacobs, Ruth Magdalena; Jain, Vivek; Jakobi, Katharina Bianca; Jakobs, Karl; Jakobsen, Sune; Jakoubek, Tomas; Jamin, David Olivier; Jana, Dilip; Jansky, Roland; Janssen, Jens; Janus, Michel; Janus, Piotr Andrzej; Jarlskog, G\\"oran; Javadov, Namig; Jav\\r{u}rek, Tom\\'{a}\\v{s}; Javurkova, Martina; Jeanneau, Fabien; Jeanty, Laura; Jejelava, Juansher; Jelinskas, Adomas; Jenni, Peter; Jeske, Carl; J\\'ez\\'equel, St\\'ephane; Ji, Haoshuang; Jia, Jiangyong; Jiang, Hai; Jiang, Yi; Jiang, Zihao; Jiggins, Stephen; Jimenez Pena, Javier; Jin, Shan; Jinaru, Adam; Jinnouchi, Osamu; Jivan, Harshna; Johansson, Per; Johns, Kenneth; Johnson, Christian; Johnson, William Joseph; Jon-And, Kerstin; Jones, Roger; Jones, Samuel David; Jones, Sarah; Jones, Tim; Jongmanns, Jan; Jorge, Pedro; Jovicevic, Jelena; Ju, Xiangyang; Juste Rozas, Aurelio; K\\"{o}hler, Markus Konrad; Kaczmarska, Anna; Kado, Marumi; Kagan, Harris; Kagan, Michael; Kahn, Sebastien Jonathan; Kaji, Toshiaki; Kajomovitz, Enrique; Kalderon, Charles William; Kaluza, Adam; Kama, Sami; Kamenshchikov, Andrey; Kanaya, Naoko; Kanjir, Luka; Kantserov, Vadim; Kanzaki, Junichi; Kaplan, Benjamin; Kaplan, Laser Seymour; Kar, Deepak; Karakostas, Konstantinos; Karastathis, Nikolaos; Kareem, Mohammad Jawad; Karentzos, Efstathios; Karpov, Sergey; Karpova, Zoya; Kartvelishvili, Vakhtang; Karyukhin, Andrey; Kasahara, Kota; Kashif, Lashkar; Kass, Richard; Kastanas, Alex; Kataoka, Yousuke; Kato, Chikuma; Katre, Akshay; Katzy, Judith; Kawade, Kentaro; Kawagoe, Kiyotomo; Kawamoto, Tatsuo; Kawamura, Gen; Kay, Ellis; Kazanin, Vassili; Keeler, Richard; Kehoe, Robert; Keller, John; Kellermann, Edgar; Kempster, Jacob Julian; Kendrick, James; Keoshkerian, Houry; Kepka, Oldrich; Ker\\v{s}evan, Borut Paul; Kersten, Susanne; Keyes, Robert; Khader, Mazin; Khalil-zada, Farkhad; Khanov, Alexander; Kharlamov, Alexey; Kharlamova, Tatyana; Khodinov, Alexander; Khoo, Teng Jian; Khovanskiy, Valery; Khramov, Evgeniy; Khubua, Jemal; Kido, Shogo; Kiehn, Moritz; Kilby, Callum; Kim, Hee Yeun; Kim, Shinhong; Kim, Young-Kee; Kimura, Naoki; Kind, Oliver Maria; King, Barry; Kirchmeier, David; Kirk, Julie; Kiryunin, Andrey; Kishimoto, Tomoe; Kisielewska, Danuta; Kitali, Vincent; Kivernyk, Oleh; Kladiva, Eduard; Klapdor-Kleingrothaus, Thorwald; Klein, Matthew Henry; Klein, Max; Klein, Uta; Kleinknecht, Konrad; Klimek, Pawel; Klimentov, Alexei; Klingenberg, Reiner; Klingl, Tobias; Klioutchnikova, Tatiana; Klitzner, Felix Fidelio; Kluge, Eike-Erik; Kluit, Peter; Kluth, Stefan; Kneringer, Emmerich; Knoops, Edith; Knue, Andrea; Kobayashi, Aine; Kobayashi, Dai; Kobayashi, Tomio; Kobel, Michael; Kocian, Martin; Kodys, Peter; Koffas, Thomas; Koffeman, Els; K\\"ohler, Nicolas Maximilian; Koi, Tatsumi; Kolb, Mathis; Koletsou, Iro; Kondo, Takahiko; Kondrashova, Nataliia; K\\"oneke, Karsten; K\\"onig, Adriaan; Kono, Takanori; Konoplich, Rostislav; Konstantinidis, Nikolaos; Konya, Balazs; Kopeliansky, Revital; Koperny, Stefan; Korcyl, Krzysztof; Kordas, Kostantinos; Korn, Andreas; Korolkov, Ilya; Korolkova, Elena; Kortner, Oliver; Kortner, Sandra; Kosek, Tomas; Kostyukhin, Vadim; Kotwal, Ashutosh; Koulouris, Aimilianos; Kourkoumeli-Charalampidi, Athina; Kourkoumelis, Christine; Kourlitis, Evangelos; Kouskoura, Vasiliki; Kowalewska, Anna Bozena; Kowalewski, Robert Victor; Kowalski, Tadeusz; Kozakai, Chihiro; Kozanecki, Witold; Kozhin, Anatoly; Kramarenko, Viktor; Kramberger, Gregor; Krasnopevtsev, Dimitrii; Krasny, Mieczyslaw Witold; Krasznahorkay, Attila; Krauss, Dominik; Kremer, Jakub Andrzej; Kretzschmar, Jan; Kreutzfeldt, Kristof; Krieger, Peter; Krizka, Karol; Kroeninger, Kevin; Kroha, Hubert; Kroll, Jiri; Kroll, Joe; Kroseberg, Juergen; Krstic, Jelena; Kruchonak, Uladzimir; Kr\\"uger, Hans; Krumnack, Nils; Kruse, Mark; Kubota, Takashi; Kucuk, Hilal; Kuday, Sinan; Kuechler, Jan Thomas; Kuehn, Susanne; Kugel, Andreas; Kuger, Fabian; Kuhl, Thorsten; Kukhtin, Victor; Kukla, Romain; Kulchitsky, Yuri; Kuleshov, Sergey; Kulinich, Yakov Petrovich; Kuna, Marine; Kunigo, Takuto; Kupco, Alexander; Kupfer, Tobias; Kuprash, Oleg; Kurashige, Hisaya; Kurchaninov, Leonid; Kurochkin, Yurii; Kurth, Matthew Glenn; Kuwertz, Emma Sian; Kuze, Masahiro; Kvita, Jiri; Kwan, Tony; Kyriazopoulos, Dimitrios; La Rosa, Alessandro; La Rosa Navarro, Jose Luis; La Rotonda, Laura; La Ruffa, Francesco; Lacasta, Carlos; Lacava, Francesco; Lacey, James; Lack, David Philip John; Lacker, Heiko; Lacour, Didier; Ladygin, Evgueni; Lafaye, Remi; Laforge, Bertrand; Lai, Stanley; Lammers, Sabine; Lampl, Walter; Lan\\c con, Eric; Landgraf, Ulrich; Landon, Murrough; Lanfermann, Marie Christine; Lang, Valerie Susanne; Lange, J \\"{o}rn Christian; Langenberg, Robert Johannes; Lankford, Andrew; Lanni, Francesco; Lantzsch, Kerstin; Lanza, Agostino; Lapertosa, Alessandro; Laplace, Sandrine; Laporte, Jean-Francois; Lari, Tommaso; Lasagni Manghi, Federico; Lassnig, Mario; Lau, Tak Shun; Laurelli, Paolo; Lavrijsen, Wim; Law, Alexander; Laycock, Paul; Lazovich, Tomo; Lazzaroni, Massimo; Le, Brian; Le Dortz, Olivier; Le Guirriec, Emmanuel; Le Quilleuc, Eloi; LeBlanc, Matthew Edgar; LeCompte, Thomas; Ledroit-Guillon, Fabienne; Lee, Claire Alexandra; Lee, Graham Richard; Lee, Shih-Chang; Lee, Lawrence; Lefebvre, Benoit; Lefebvre, Guillaume; Lefebvre, Michel; Legger, Federica; Leggett, Charles; Lehmann Miotto, Giovanna; Lei, Xiaowen; Leight, William Axel; Leite, Marco Aurelio Lisboa; Leitner, Rupert; Lellouch, Daniel; Lemmer, Boris; Leney, Katharine; Lenz, Tatjana; Lenzi, Bruno; Leone, Robert; Leone, Sandra; Leonidopoulos, Christos; Lerner, Giuseppe; Leroy, Claude; Les, Robert; Lesage, Arthur; Lester, Christopher; Levchenko, Mikhail; Lev\\^eque, Jessica; Levin, Daniel; Levinson, Lorne; Levy, Mark; Lewis, Dave; Li, Bing; Li, Haifeng; Li, Liang; Li, Qi; Li, Quanyin; Li, Shu; Li, Xingguo; Li, Yichen; Liang, Zhijun; Liberti, Barbara; Liblong, Aaron; Lie, Ki; Limosani, Antonio; Lin, Chiao-ying; Lin, Kuan-yu; Lin, Simon; Lin, Tai-Hua; Linck, Rebecca Anne; Lindquist, Brian Edward; Lionti, Anthony Eric; Lipeles, Elliot; Lipniacka, Anna; Lisovyi, Mykhailo; Liss, Tony; Lister, Alison; Litke, Alan; Liu, Bo; Liu, Hao; Liu, Hongbin; Liu, Jesse; Liu, Jian; Liu, Jianbei; Liu, Kun; Liu, Lulu; Liu, Minghui; Liu, Yanlin; Liu, Yanwen; Livan, Michele; Lleres, Annick; Llorente Merino, Javier; Lloyd, Stephen; Lo, Cheuk Yee; Lo Sterzo, Francesco; Lobodzinska, Ewelina Maria; Loch, Peter; Loebinger, Fred; Loesle, Alena; Loew, Kevin Michael; Lohse, Thomas; Lohwasser, Kristin; Lokajicek, Milos; Long, Brian Alexander; Long, Jonathan David; Long, Robin Eamonn; Longo, Luigi; Looper, Kristina Anne; Lopez, Jorge; Lopez Paz, Ivan; Lopez Solis, Alvaro; Lorenz, Jeanette; Lorenzo Martinez, Narei; Losada, Marta; L{\\"o}sel, Philipp Jonathan; Lou, XinChou; Lounis, Abdenour; Love, Jeremy; Love, Peter; Lu, Haonan; Lu, Nan; Lu, Yun-Ju; Lubatti, Henry; Luci, Claudio; Lucotte, Arnaud; Luedtke, Christian; Luehring, Frederick; Lukas, Wolfgang; Luminari, Lamberto; Lund-Jensen, Bengt; Lutz, Margaret Susan; Luzi, Pierre Marc; Lynn, David; Lysak, Roman; Lytken, Else; Lyu, Feng; Lyubushkin, Vladimir; Ma, Hong; Ma, Lian Liang; Ma, Yanhui; Maccarrone, Giovanni; Macchiolo, Anna; Macdonald, Calum Michael; Ma\\v{c}ek, Bo\\v{s}tjan; Machado Miguens, Joana; Madaffari, Daniele; Madar, Romain; Mader, Wolfgang; Madsen, Alexander; Madysa, Nico; Maeda, Junpei; Maeland, Steffen; Maeno, Tadashi; Maevskiy, Artem; Magerl, Veronika; Maiani, Camilla; Maidantchik, Carmen; Maier, Thomas; Maio, Am\\'elia; Majersky, Oliver; Majewski, Stephanie; Makida, Yasuhiro; Makovec, Nikola; Malaescu, Bogdan; Malecki, Pawel; Maleev, Victor; Malek, Fairouz; Mallik, Usha; Malon, David; Malone, Claire; Maltezos, Stavros; Malyukov, Sergei; Mamuzic, Judita; Mancini, Giada; Mandi\\'{c}, Igor; Maneira, Jos\\'e; Manhaes de Andrade Filho, Luciano; Manjarres Ramos, Joany; Mankinen, Katja Hannele; Mann, Alexander; Manousos, Athanasios; Mansoulie, Bruno; Mansour, Jason Dhia; Mantifel, Rodger; Mantoani, Matteo; Manzoni, Stefano; Mapelli, Livio; Marceca, Gino; March, Luis; Marchese, Luigi; Marchiori, Giovanni; Marcisovsky, Michal; Marin Tobon, Cesar Augusto; Marjanovic, Marija; Marley, Daniel; Marroquim, Fernando; Marsden, Stephen Philip; Marshall, Zach; Martensson, Mikael; Marti-Garcia, Salvador; Martin, Christopher Blake; Martin, Tim; Martin, Victoria Jane; Martin dit Latour, Bertrand; Martinez, Mario; Martinez Outschoorn, Verena; Martin-Haugh, Stewart; Martoiu, Victor Sorin; Martyniuk, Alex; Marzin, Antoine; Masetti, Lucia; Mashimo, Tetsuro; Mashinistov, Ruslan; Masik, Jiri; Maslennikov, Alexey; Mason, Lara Hannan; Massa, Lorenzo; Mastrandrea, Paolo; Mastroberardino, Anna; Masubuchi, Tatsuya; M\\"attig, Peter; Maurer, Julien; Maxfield, Stephen; Maximov, Dmitriy; Mazini, Rachid; Maznas, Ioannis; Mazza, Simone Michele; Mc Fadden, Neil Christopher; Mc Goldrick, Garrin; Mc Kee, Shawn Patrick; McCarn, Allison; McCarthy, Robert; McCarthy, Tom; McClymont, Laurie; McDonald, Emily; Mcfayden, Josh; Mchedlidze, Gvantsa; McMahon, Steve; McNamara, Peter Charles; McNicol, Christopher John; McPherson, Robert; Meehan, Samuel; Megy, Theo Jean; Mehlhase, Sascha; Mehta, Andrew; Meideck, Thomas; Meier, Karlheinz; Meirose, Bernhard; Melini, Davide; Mellado Garcia, Bruce Rafael; Mellenthin, Johannes Donatus; Melo, Matej; Meloni, Federico; Melzer, Alexander; Menary, Stephen Burns; Meng, Lingxin; Meng, Xiangting; Mengarelli, Alberto; Menke, Sven; Meoni, Evelin; Mergelmeyer, Sebastian; Merlassino, Claudia; Mermod, Philippe; Merola, Leonardo; Meroni, Chiara; Merritt, Frank; Messina, Andrea; Metcalfe, Jessica; Mete, Alaettin Serhan; Meyer, Christopher; Meyer, Jean-Pierre; Meyer, Jochen; Meyer Zu Theenhausen, Hanno; Miano, Fabrizio; Middleton, Robin; Miglioranzi, Silvia; Mijovi\\'{c}, Liza; Mikenberg, Giora; Mikestikova, Marcela; Miku\\v{z}, Marko; Milesi, Marco; Milic, Adriana; Millar, Declan Andrew; Miller, David; Mills, Corrinne; Milov, Alexander; Milstead, David; Minaenko, Andrey; Minami, Yuto; Minashvili, Irakli; Mincer, Allen; Mindur, Bartosz; Mineev, Mikhail; Minegishi, Yuji; Ming, Yao; Mir, Lluisa-Maria; Mirto, Alessandro; Mistry, Khilesh; Mitani, Takashi; Mitrevski, Jovan; Mitsou, Vasiliki A; Miucci, Antonio; Miyagawa, Paul; Mizukami, Atsushi; Mj\\"ornmark, Jan-Ulf; Mkrtchyan, Tigran; Mlynarikova, Michaela; Moa, Torbjoern; Mochizuki, Kazuya; Mogg, Philipp; Mohapatra, Soumya; Molander, Simon; Moles-Valls, Regina; Mondragon, Matthew Craig; M\\"onig, Klaus; Monk, James; Monnier, Emmanuel; Montalbano, Alyssa; Montejo Berlingen, Javier; Monticelli, Fernando; Monzani, Simone; Moore, Roger; Morange, Nicolas; Moreno, Deywis; Moreno Ll\\'acer, Mar\\'ia; Morettini, Paolo; Morgenstern, Marcus; Morgenstern, Stefanie; Mori, Daniel; Mori, Tatsuya; Morii, Masahiro; Morinaga, Masahiro; Morisbak, Vanja; Morley, Anthony Keith; Mornacchi, Giuseppe; Morris, John; Morvaj, Ljiljana; Moschovakos, Paris; Mosidze, Maia; Moss, Harry James; Moss, Josh; Motohashi, Kazuki; Mount, Richard; Mountricha, Eleni; Moyse, Edward; Muanza, Steve; Mueller, Felix; Mueller, James; Mueller, Ralph Soeren Peter; Muenstermann, Daniel; Mullen, Paul; Mullier, Geoffrey; Munoz Sanchez, Francisca Javiela; Murray, Bill; Musheghyan, Haykuhi; Mu\\v{s}kinja, Miha; Mwewa, Chilufya; Myagkov, Alexey; Myska, Miroslav; Nachman, Benjamin Philip; Nackenhorst, Olaf; Nagai, Koichi; Nagai, Ryo; Nagano, Kunihiro; Nagasaka, Yasushi; Nagata, Kazuki; Nagel, Martin; Nagy, Elemer; Nairz, Armin Michael; Nakahama, Yu; Nakamura, Koji; Nakamura, Tomoaki; Nakano, Itsuo; Naranjo Garcia, Roger Felipe; Narayan, Rohin; Narrias Villar, Daniel Isaac; Naryshkin, Iouri; Naumann, Thomas; Navarro, Gabriela; Nayyar, Ruchika; Neal, Homer; Nechaeva, Polina; Neep, Thomas James; Negri, Andrea; Negrini, Matteo; Nektarijevic, Snezana; Nellist, Clara; Nelson, Andrew; Nelson, Michael Edward; Nemecek, Stanislav; Nemethy, Peter; Nessi, Marzio; Neubauer, Mark; Neumann, Manuel; Newman, Paul; Ng, Tsz Yu; Ng, Sam Yanwing; Nguyen Manh, Tuan; Nickerson, Richard; Nicolaidou, Rosy; Nielsen, Jason; Nikiforou, Nikiforos; Nikolaenko, Vladimir; Nikolic-Audit, Irena; Nikolopoulos, Konstantinos; Nilsson, Paul; Ninomiya, Yoichi; Nisati, Aleandro; Nishu, Nishu; Nisius, Richard; Nitsche, Isabel; Nitta, Tatsumi; Nobe, Takuya; Noguchi, Yohei; Nomachi, Masaharu; Nomidis, Ioannis; Nomura, Marcelo Ayumu; Nooney, Tamsin; Nordberg, Markus; Norjoharuddeen, Nurfikri; Novgorodova, Olga; Nozaki, Mitsuaki; Nozka, Libor; Ntekas, Konstantinos; Nurse, Emily; Nuti, Francesco; O'connor, Kelsey; O'Neil, Dugan; O'Rourke, Abigail Alexandra; O'Shea, Val; Oakham, Gerald; Oberlack, Horst; Obermann, Theresa; Ocariz, Jose; Ochi, Atsuhiko; Ochoa, Ines; Ochoa-Ricoux, Juan Pedro; Oda, Susumu; Odaka, Shigeru; Oh, Alexander; Oh, Seog; Ohm, Christian; Ohman, Henrik; Oide, Hideyuki; Okawa, Hideki; Okumura, Yasuyuki; Okuyama, Toyonobu; Olariu, Albert; Oleiro Seabra, Luis Filipe; Olivares Pino, Sebastian Andres; Oliveira Damazio, Denis; Olsson, Joakim; Olszewski, Andrzej; Olszowska, Jolanta; Onofre, Ant\\'onio; Onogi, Kouta; Onyisi, Peter; Oppen, Henrik; Oreglia, Mark; Oren, Yona; Orestano, Domizia; Orgill, Emily Claire; Orlando, Nicola; Orr, Robert; Osculati, Bianca; Ospanov, Rustem; Otero y Garzon, Gustavo; Otono, Hidetoshi; Ouchrif, Mohamed; Ould-Saada, Farid; Ouraou, Ahmimed; Oussoren, Koen Pieter; Ouyang, Qun; Owen, Mark; Owen, Rhys Edward; Ozcan, Veysi Erkcan; Ozturk, Nurcan; Pachal, Katherine; Pacheco Pages, Andres; Pacheco Rodriguez, Laura; Padilla Aranda, Cristobal; Pagan Griso, Simone; Paganini, Michela; Paige, Frank; Palacino, Gabriel; Palazzo, Serena; Palestini, Sandro; Palka, Marek; Pallin, Dominique; Panagiotopoulou, Evgenia; Panagoulias, Ilias; Pandini, Carlo Enrico; Panduro Vazquez, William; Pani, Priscilla; Panitkin, Sergey; Pantea, Dan; Paolozzi, Lorenzo; Papadopoulou, Theodora; Papageorgiou, Konstantinos; Paramonov, Alexander; Paredes Hernandez, Daniela; Parker, Adam Jackson; Parker, Michael Andrew; Parker, Kerry Ann; Parodi, Fabrizio; Parsons, John; Parzefall, Ulrich; Pascuzzi, Vincent; Pasner, Jacob Martin; Pasqualucci, Enrico; Passaggio, Stefano; Pastore, Francesca; Pataraia, Sophio; Pater, Joleen; Pauly, Thilo; Pearson, Benjamin; Pedraza Lopez, Sebastian; Pedro, Rute; Peleganchuk, Sergey; Penc, Ondrej; Peng, Cong; Peng, Haiping; Penwell, John; Peralva, Bernardo; Perego, Marta Maria; Perepelitsa, Dennis; Peri, Francesco; Perini, Laura; Pernegger, Heinz; Perrella, Sabrina; Peschke, Richard; Peshekhonov, Vladimir; Peters, Krisztian; Peters, Yvonne; Petersen, Brian; Petersen, Troels; Petit, Elisabeth; Petridis, Andreas; Petridou, Chariclia; Petroff, Pierre; Petrolo, Emilio; Petrov, Mariyan; Petrucci, Fabrizio; Pettersson, Nora Emilia; Peyaud, Alan; Pezoa, Raquel; Phillips, Forrest Hays; Phillips, Peter William; Piacquadio, Giacinto; Pianori, Elisabetta; Picazio, Attilio; Pickering, Mark Andrew; Piegaia, Ricardo; Pilcher, James; Pilkington, Andrew; Pinamonti, Michele; Pinfold, James; Pirumov, Hayk; Pitt, Michael; Plazak, Lukas; Pleier, Marc-Andre; Pleskot, Vojtech; Plotnikova, Elena; Pluth, Daniel; Podberezko, Pavel; Poettgen, Ruth; Poggi, Riccardo; Poggioli, Luc; Pogrebnyak, Ivan; Pohl, David-leon; Pokharel, Ishan; Polesello, Giacomo; Poley, Anne-luise; Policicchio, Antonio; Polifka, Richard; Polini, Alessandro; Pollard, Christopher Samuel; Polychronakos, Venetios; Pomm\\`es, Kathy; Ponomarenko, Daniil; Pontecorvo, Ludovico; Popeneciu, Gabriel Alexandru; Portillo Quintero, Dilia Mar\\'ia; Pospisil, Stanislav; Potamianos, Karolos; Potrap, Igor; Potter, Christina; Potti, Harish; Poulsen, Trine; Poveda, Joaquin; Pozo Astigarraga, Mikel Eukeni; Pralavorio, Pascal; Pranko, Aliaksandr; Prell, Soeren; Price, Darren; Primavera, Margherita; Prince, Sebastien; Proklova, Nadezda; Prokofiev, Kirill; Prokoshin, Fedor; Protopopescu, Serban; Proudfoot, James; Przybycien, Mariusz; Puri, Akshat; Puzo, Patrick; Qian, Jianming; Qin, Yang; Quadt, Arnulf; Queitsch-Maitland, Michaela; Quilty, Donnchadha; Raddum, Silje; Radeka, Veljko; Radescu, Voica; Radhakrishnan, Sooraj Krishnan; Radloff, Peter; Rados, Pere; Ragusa, Francesco; Rahal, Ghita; Raine, John Andrew; Rajagopalan, Srinivasan; Rashid, Tasneem; Raspopov, Sergii; Ratti, Maria Giulia; Rauch, Daniel; Rauscher, Felix; Rave, Stefan; Ravinovich, Ilia; Rawling, Jacob Henry; Raymond, Michel; Read, Alexander Lincoln; Readioff, Nathan Peter; Reale, Marilea; Rebuzzi, Daniela; Redelbach, Andreas; Redlinger, George; Reece, Ryan; Reed, Robert; Reeves, Kendall; Rehnisch, Laura; Reichert, Joseph; Reiss, Andreas; Rembser, Christoph; Ren, Huan; Rescigno, Marco; Resconi, Silvia; Resseguie, Elodie Deborah; Rettie, Sebastien; Reynolds, Elliot; Rezanova, Olga; Reznicek, Pavel; Rezvani, Reyhaneh; Richter, Robert; Richter, Stefan; Richter-Was, Elzbieta; Ricken, Oliver; Ridel, Melissa; Rieck, Patrick; Riegel, Christian Johann; Rieger, Julia; Rifki, Othmane; Rijssenbeek, Michael; Rimoldi, Adele; Rimoldi, Marco; Rinaldi, Lorenzo; Ripellino, Giulia; Risti\\'{c}, Branislav; Ritsch, Elmar; Riu, Imma; Rizatdinova, Flera; Rizvi, Eram; Rizzi, Chiara; Roberts, Rhys Thomas; Robertson, Steven; Robichaud-Veronneau, Andree; Robinson, Dave; Robinson, James; Robson, Aidan; Rocco, Elena; Roda, Chiara; Rodina, Yulia; Rodriguez Bosca, Sergi; Rodriguez Perez, Andrea; Rodriguez Rodriguez, Daniel; Roe, Shaun; Rogan, Christopher Sean; R{\\o}hne, Ole; Roloff, Jennifer; Romaniouk, Anatoli; Romano, Marino; Romano Saez, Silvestre Marino; Romero Adam, Elena; Rompotis, Nikolaos; Ronzani, Manfredi; Roos, Lydia; Rosati, Stefano; Rosbach, Kilian; Rose, Peyton; Rosien, Nils-Arne; Rossi, Elvira; Rossi, Leonardo Paolo; Rosten, Jonatan; Rosten, Rachel; Rotaru, Marina; Rothberg, Joseph; Rousseau, David; Roy, Debarati; Rozanov, Alexandre; Rozen, Yoram; Ruan, Xifeng; Rubbo, Francesco; R\\"uhr, Frederik; Ruiz-Martinez, Aranzazu; Rurikova, Zuzana; Rusakovich, Nikolai; Russell, Heather; Rutherfoord, John; Ruthmann, Nils; R{\\"u}ttinger, Elias Michael; Ryabov, Yury; Rybar, Martin; Rybkin, Grigori; Ryu, Soo; Ryzhov, Andrey; Rzehorz, Gerhard Ferdinand; Saavedra, Aldo; Sabato, Gabriele; Sacerdoti, Sabrina; Sadrozinski, Hartmut; Sadykov, Renat; Safai Tehrani, Francesco; Saha, Puja; Sahinsoy, Merve; Saimpert, Matthias; Saito, Masahiko; Saito, Tomoyuki; Sakamoto, Hiroshi; Sakurai, Yuki; Salamanna, Giuseppe; Salazar Loyola, Javier Esteban; Salek, David; Sales De Bruin, Pedro Henrique; Salihagic, Denis; Salnikov, Andrei; Salt, Jos\\'e; Salvatore, Daniela; Salvatore, Pasquale Fabrizio; Salvucci, Antonio; Salzburger, Andreas; Sammel, Dirk; Sampsonidis, Dimitrios; Sampsonidou, Despoina; S\\'anchez, Javier; Sanchez Pineda, Arturo Rodolfo; Sandaker, Heidi; Sandbach, Ruth Laura; Sander, Christian Oliver; Sandhoff, Marisa; Sandoval, Carlos; Sankey, Dave; Sannino, Mario; Sano, Yuta; Sansoni, Andrea; Santoni, Claudio; Santos, Helena; Santoyo Castillo, Itzebelt; Sapronov, Andrey; Saraiva, Jo\\~ao; Sasaki, Osamu; Sato, Koji; Sauvan, Emmanuel; Savage, Graham; Savard, Pierre; Savic, Natascha; Sawyer, Craig; Sawyer, Lee; Sbarra, Carla; Sbrizzi, Antonio; Scanlon, Tim; Scannicchio, Diana; Schaarschmidt, Jana; Schacht, Peter; Schachtner, Balthasar Maria; Schaefer, Douglas; Schaefer, Leigh; Schaeffer, Jan; Schaepe, Steffen; Sch\\"afer, Uli; Schaffer, Arthur; Schaile, Dorothee; Schamberger, R Dean; Schegelsky, Valery; Scheirich, Daniel; Schenck, Ferdinand; Schernau, Michael; Schiavi, Carlo; Schier, Sheena; Schildgen, Lara Katharina; Schillo, Christian; Schioppa, Marco; Schlenker, Stefan; Schmidt-Sommerfeld, Korbinian Ralf; Schmieden, Kristof; Schmitt, Christian; Schmitt, Stefan; Schmitz, Simon; Schnoor, Ulrike; Schoeffel, Laurent; Schoening, Andre; Schoenrock, Bradley Daniel; Schopf, Elisabeth; Schott, Matthias; Schouwenberg, Jeroen; Schovancova, Jaroslava; Schramm, Steven; Schuh, Natascha; Schulte, Alexandra; Schultens, Martin Johannes; Schultz-Coulon, Hans-Christian; Schumacher, Markus; Schumm, Bruce; Schune, Philippe; Schwartzman, Ariel; Schwarz, Thomas Andrew; Schweiger, Hansdieter; Schwemling, Philippe; Schwienhorst, Reinhard; Schwindling, Jerome; Sciandra, Andrea; Sciolla, Gabriella; Scornajenghi, Matteo; Scuri, Fabrizio; Scutti, Federico; Searcy, Jacob; Seema, Pienpen; Seidel, Sally; Seiden, Abraham; Seixas, Jos\\'e; Sekhniaidze, Givi; Sekhon, Karishma; Sekula, Stephen; Semprini-Cesari, Nicola; Senkin, Sergey; Serfon, Cedric; Serin, Laurent; Serkin, Leonid; Sessa, Marco; Seuster, Rolf; Severini, Horst; \\v{S}filigoj, Tina; Sforza, Federico; Sfyrla, Anna; Shabalina, Elizaveta; Shaikh, Nabila Wahab; Shan, Lianyou; Shang, Ruo-yu; Shank, James; Shapiro, Marjorie; Shatalov, Pavel; Shaw, Kate; Shaw, Savanna Marie; Shcherbakova, Anna; Shehu, Ciwake Yusufu; Shen, Yu-Ting; Sherafati, Nima; Sherman, Alexander David; Sherwood, Peter; Shi, Liaoshan; Shimizu, Shima; Shimmin, Chase Owen; Shimojima, Makoto; Shipsey, Ian Peter Joseph; Shirabe, Shohei; Shiyakova, Mariya; Shlomi, Jonathan; Shmeleva, Alevtina; Shoaleh Saadi, Diane; Shochet, Mel; Shojaii, Seyed Ruhollah; Shope, David Richard; Shrestha, Suyog; Shulga, Evgeny; Shupe, Michael; Sicho, Petr; Sickles, Anne Marie; Sidebo, Per Edvin; Sideras Haddad, Elias; Sidiropoulou, Ourania; Sidoti, Antonio; Siegert, Frank; Sijacki, Djordje; Silva, Jos\\'e; Silverstein, Samuel; Simak, Vladislav; Simic, Ljiljana; Simion, Stefan; Simioni, Eduard; Simmons, Brinick; Simon, Manuel; Sinervo, Pekka; Sinev, Nikolai; Sioli, Maximiliano; Siragusa, Giovanni; Siral, Ismet; Sivoklokov, Serguei; Sj\\"{o}lin, J\\"{o}rgen; Skinner, Malcolm Bruce; Skubic, Patrick; Slater, Mark; Slavicek, Tomas; Slawinska, Magdalena; Sliwa, Krzysztof; Slovak, Radim; Smakhtin, Vladimir; Smart, Ben; Smiesko, Juraj; Smirnov, Nikita; Smirnov, Sergei; Smirnov, Yury; Smirnova, Lidia; Smirnova, Oxana; Smith, Joshua Wyatt; Smith, Matthew; Smith, Russell; Smizanska, Maria; Smolek, Karel; Snesarev, Andrei; Snyder, Ian Michael; Snyder, Scott; Sobie, Randall; Socher, Felix; Soffer, Abner; S{\\o}gaard, Andreas; Soh, Dart-yin; Sokhrannyi, Grygorii; Solans Sanchez, Carlos; Solar, Michael; Soldatov, Evgeny; Soldevila, Urmila; Solodkov, Alexander; Soloshenko, Alexei; Solovyanov, Oleg; Solovyev, Victor; Sommer, Philip; Son, Hyungsuk; Sopczak, Andre; Sosa, David; Sotiropoulou, Calliope Louisa; Sottocornola, Simone; Soualah, Rachik; Soukharev, Andrey; South, David; Sowden, Benjamin; Spagnolo, Stefania; Spalla, Margherita; Spangenberg, Martin; Span\\`o, Francesco; Sperlich, Dennis; Spettel, Fabian; Spieker, Thomas Malte; Spighi, Roberto; Spigo, Giancarlo; Spiller, Laurence Anthony; Spousta, Martin; St Denis, Richard Dante; Stabile, Alberto; Stamen, Rainer; Stamm, Soren; Stanecka, Ewa; Stanek, Robert; Stanescu, Cristian; Stanitzki, Marcel Michael; Stapf, Birgit Sylvia; Stapnes, Steinar; Starchenko, Evgeny; Stark, Giordon; Stark, Jan; Stark, Simon Holm; Staroba, Pavel; Starovoitov, Pavel; St\\"arz, Steffen; Staszewski, Rafal; Stegler, Martin; Steinberg, Peter; Stelzer, Bernd; Stelzer, Harald Joerg; Stelzer-Chilton, Oliver; Stenzel, Hasko; Stevenson, Thomas James; Stewart, Graeme; Stockton, Mark; Stoebe, Michael; Stoicea, Gabriel; Stolte, Philipp; Stonjek, Stefan; Stradling, Alden; Straessner, Arno; Stramaglia, Maria Elena; Strandberg, Jonas; Strandberg, Sara; Strauss, Michael; Strizenec, Pavol; Str\\"ohmer, Raimund; Strom, David; Stroynowski, Ryszard; Strubig, Antonia; Stucci, Stefania Antonia; Stugu, Bjarne; Styles, Nicholas Adam; Su, Dong; Su, Jun; Suchek, Stanislav; Sugaya, Yorihito; Suk, Michal; Sulin, Vladimir; Sultan, D M S; Sultansoy, Saleh; Sumida, Toshi; Sun, Siyuan; Sun, Xiaohu; Suruliz, Kerim; Suster, Carl; Sutton, Mark; Suzuki, Shota; Svatos, Michal; Swiatlowski, Maximilian; Swift, Stewart Patrick; Sykora, Ivan; Sykora, Tomas; Ta, Duc; Tackmann, Kerstin; Taenzer, Joe; Taffard, Anyes; Tafirout, Reda; Tahirovic, Elvedin; Taiblum, Nimrod; Takai, Helio; Takashima, Ryuichi; Takasugi, Eric Hayato; Takeda, Kosuke; Takeshita, Tohru; Takubo, Yosuke; Talby, Mossadek; Talyshev, Alexey; Tanaka, Junichi; Tanaka, Masahiro; Tanaka, Reisaburo; Tanioka, Ryo; Tannenwald, Benjamin Bordy; Tapia Araya, Sebastian; Tapprogge, Stefan; Tarem, Shlomit; Tartarelli, Giuseppe Francesco; Tas, Petr; Tasevsky, Marek; Tashiro, Takuya; Tassi, Enrico; Tavares Delgado, Ademar; Tayalati, Yahya; Taylor, Aaron; Taylor, Alan James; Taylor, Geoffrey; Taylor, Pierre Thor Elliot; Taylor, Wendy; Teixeira-Dias, Pedro; Temple, Darren; Ten Kate, Herman; Teng, Ping-Kun; Teoh, Jia Jian; Tepel, Fabian-Phillipp; Terada, Susumu; Terashi, Koji; Terron, Juan; Terzo, Stefano; Testa, Marianna; Teuscher, Richard; Thais, Savannah Jennifer; Theveneaux-Pelzer, Timoth\\'ee; Thiele, Fabian; Thomas, Juergen; Thomas-Wilsker, Joshuha; Thompson, Paul; Thompson, Stan; Thomsen, Lotte Ansgaard; Thomson, Evelyn; Tian, Yun; Tibbetts, Mark James; Ticse Torres, Royer Edson; Tikhomirov, Vladimir; Tikhonov, Yury; Timoshenko, Sergey; Tipton, Paul; Tisserant, Sylvain; Todome, Kazuki; Todorova-Nova, Sharka; Todt, Stefanie; Tojo, Junji; Tok\\'ar, Stanislav; Tokushuku, Katsuo; Tolley, Emma; Tomlinson, Lee; Tomoto, Makoto; Tompkins, Lauren; Toms, Konstantin; Tong, Baojia(Tony); Tornambe, Peter; Torrence, Eric; Torres, Heberth; Torr\\'o Pastor, Emma; Toth, Jozsef; Touchard, Francois; Tovey, Daniel; Treado, Colleen Jennifer; Trefzger, Thomas; Tresoldi, Fabio; Tricoli, Alessandro; Trigger, Isabel Marian; Trincaz-Duvoid, Sophie; Tripiana, Martin; Trischuk, William; Trocm\\'e, Benjamin; Trofymov, Artur; Troncon, Clara; Trovatelli, Monica; Truong, Loan; Trzebinski, Maciej; Trzupek, Adam; Tsang, Ka Wa; Tseng, Jeffrey; Tsiareshka, Pavel; Tsirintanis, Nikolaos; Tsiskaridze, Shota; Tsiskaridze, Vakhtang; Tskhadadze, Edisher; Tsukerman, Ilya; Tsulaia, Vakhtang; Tsuno, Soshi; Tsybychev, Dmitri; Tu, Yanjun; Tudorache, Alexandra; Tudorache, Valentina; Tulbure, Traian Tiberiu; Tuna, Alexander Naip; Turchikhin, Semen; Turgeman, Daniel; Turk Cakir, Ilkay; Turra, Ruggero; Tuts, Michael; Ucchielli, Giulia; Ueda, Ikuo; Ughetto, Michael; Ukegawa, Fumihiko; Unal, Guillaume; Undrus, Alexander; Unel, Gokhan; Ungaro, Francesca; Unno, Yoshinobu; Uno, Kenta; Urban, Jozef; Urquijo, Phillip; Urrejola, Pedro; Usai, Giulio; Usui, Junya; Vacavant, Laurent; Vacek, Vaclav; Vachon, Brigitte; Vadla, Knut Oddvar Hoie; Vaidya, Amal; Valderanis, Chrysostomos; Valdes Santurio, Eduardo; Valente, Marco; Valentinetti, Sara; Valero, Alberto; Val\\'ery, Lo\\"ic; Vallier, Alexis; Valls Ferrer, Juan Antonio; Van Den Wollenberg, Wouter; van der Graaf, Harry; van Gemmeren, Peter; Van Nieuwkoop, Jacobus; van Vulpen, Ivo; van Woerden, Marius Cornelis; Vanadia, Marco; Vandelli, Wainer; Vaniachine, Alexandre; Vankov, Peter; Vardanyan, Gagik; Vari, Riccardo; Varnes, Erich; Varni, Carlo; Varol, Tulin; Varouchas, Dimitris; Vartapetian, Armen; Varvell, Kevin; Vasquez, Jared Gregory; Vasquez, Gerardo; Vazeille, Francois; Vazquez Furelos, David; Vazquez Schroeder, Tamara; Veatch, Jason; Veeraraghavan, Venkatesh; Veloce, Laurelle Maria; Veloso, Filipe; Veneziano, Stefano; Ventura, Andrea; Venturi, Manuela; Venturi, Nicola; Vercesi, Valerio; Verducci, Monica; Verkerke, Wouter; Vermeulen, Ambrosius Thomas; Vermeulen, Jos; Vetterli, Michel; Viaux Maira, Nicolas; Viazlo, Oleksandr; Vichou, Irene; Vickey, Trevor; Vickey Boeriu, Oana Elena; Viehhauser, Georg; Viel, Simon; Vigani, Luigi; Villa, Mauro; Villaplana Perez, Miguel; Vilucchi, Elisabetta; Vincter, Manuella; Vinogradov, Vladimir; Vishwakarma, Akanksha; Vittori, Camilla; Vivarelli, Iacopo; Vlachos, Sotirios; Vogel, Marcelo; Vokac, Petr; Volpi, Guido; von Buddenbrock, Stefan; von der Schmitt, Hans; von Toerne, Eckhard; Vorobel, Vit; Vorobev, Konstantin; Vos, Marcel; Voss, Rudiger; Vossebeld, Joost; Vranjes, Nenad; Vranjes Milosavljevic, Marija; Vrba, Vaclav; Vreeswijk, Marcel; Vuillermet, Raphael; Vukotic, Ilija; Wagner, Peter; Wagner, Wolfgang; Wagner-Kuhr, Jeannine; Wahlberg, Hernan; Wahrmund, Sebastian; Wakamiya, Kotaro; Walder, James; Walker, Rodney; Walkowiak, Wolfgang; Wallangen, Veronica; Wang, Ann Miao; Wang, Chao; Wang, Fuquan; Wang, Haichen; Wang, Hulin; Wang, Jike; Wang, Jin; Wang, Qing; Wang, Renjie; Wang, Rui; Wang, Song-Ming; Wang, Tingting; Wang, Wei; Wang, Wenxiao; Wang, Zirui; Wanotayaroj, Chaowaroj; Warburton, Andreas; Ward, Patricia; Wardrope, David Robert; Washbrook, Andrew; Watkins, Peter; Watson, Alan; Watson, Miriam; Watts, Gordon; Watts, Stephen; Waugh, Ben; Webb, Aaron Foley; Webb, Samuel; Weber, Michele; Weber, Sebastian Mario; Weber, Stephen; Webster, Jordan S; Weidberg, Anthony; Weinert, Benjamin; Weingarten, Jens; Weirich, Marcel; Weiser, Christian; Wells, Phillippa; Wenaus, Torre; Wengler, Thorsten; Wenig, Siegfried; Wermes, Norbert; Werner, Michael David; Werner, Per; Wessels, Martin; Weston, Thomas; Whalen, Kathleen; Whallon, Nikola Lazar; Wharton, Andrew Mark; White, Aaron; White, Andrew; White, Martin; White, Ryan; Whiteson, Daniel; Whitmore, Ben William; Wickens, Fred; Wiedenmann, Werner; Wielers, Monika; Wiglesworth, Craig; Wiik-Fuchs, Liv Antje Mari; Wildauer, Andreas; Wilk, Fabian; Wilkens, Henric George; Williams, Hugh; Williams, Sarah; Willis, Christopher; Willocq, Stephane; Wilson, John; Wingerter-Seez, Isabelle; Winkels, Emma; Winklmeier, Frank; Winston, Oliver James; Winter, Benedict Tobias; Wittgen, Matthias; Wobisch, Markus; Wolf, Anton; Wolf, Tim Michael Heinz; Wolff, Robert; Wolter, Marcin Wladyslaw; Wolters, Helmut; Wong, Vincent Wai Sum; Woods, Natasha Lee; Worm, Steven; Wosiek, Barbara; Wotschack, Jorg; Wozniak, Krzysztof; Wu, Miles; Wu, Sau Lan; Wu, Xin; Wu, Yusheng; Wyatt, Terry Richard; Wynne, Benjamin; Xella, Stefania; Xi, Zhaoxu; Xia, Ligang; Xu, Da; Xu, Lailin; Xu, Tairan; Xu, Wenhao; Yabsley, Bruce; Yacoob, Sahal; Yajima, Kazuki; Yamaguchi, Daiki; Yamaguchi, Yohei; Yamamoto, Akira; Yamamoto, Shimpei; Yamanaka, Takashi; Yamane, Fumiya; Yamatani, Masahiro; Yamazaki, Tomohiro; Yamazaki, Yuji; Yan, Zhen; Yang, Haijun; Yang, Hongtao; Yang, Yi; Yang, Zongchang; Yao, Weiming; Yap, Yee Chinn; Yasu, Yoshiji; Yatsenko, Elena; Yau Wong, Kaven Henry; Ye, Jingbo; Ye, Shuwei; Yeletskikh, Ivan; Yigitbasi, Efe; Yildirim, Eda; Yorita, Kohei; Yoshihara, Keisuke; Young, Charles; Young, Christopher John; Yu, Jaehoon; Yu, Jie; Yuen, Stephanie P; Yusuff, Imran; Zabinski, Bartlomiej; Zacharis, Georgios; Zaidan, Remi; Zaitsev, Alexander; Zakharchuk, Nataliia; Zalieckas, Justas; Zaman, Aungshuman; Zambito, Stefano; Zanzi, Daniele; Zeitnitz, Christian; Zemaityte, Gabija; Zeng, Jian Cong; Zeng, Qi; Zenin, Oleg; \\v{Z}eni\\v{s}, Tibor; Zerwas, Dirk; Zhang, Dengfeng; Zhang, Dongliang; Zhang, Fangzhou; Zhang, Guangyi; Zhang, Huijun; Zhang, Jinlong; Zhang, Lei; Zhang, Liqing; Zhang, Matt; Zhang, Peng; Zhang, Rui; Zhang, Ruiqi; Zhang, Xueyao; Zhang, Yu; Zhang, Zhiqing; Zhao, Xiandong; Zhao, Yongke; Zhao, Zhengguo; Zhemchugov, Alexey; Zhou, Bing; Zhou, Chen; Zhou, Li; Zhou, Maosen; Zhou, Mingliang; Zhou, Ning; Zhou, You; Zhu, Cheng Guang; Zhu, Hongbo; Zhu, Junjie; Zhu, Yingchun; Zhuang, Xuai; Zhukov, Konstantin; Zibell, Andre; Zieminska, Daria; Zimine, Nikolai; Zimmermann, Stephanie; Zinonos, Zinonas; Zinser, Markus; Ziolkowski, Michael; \\v{Z}ivkovi\\'{c}, Lidija; Zobernig, Georg; Zoccoli, Antonio; Zou, Rui; zur Nedden, Martin; Zwalinski, Lukasz

    2018-01-01

    Measurements of differential cross sections of top quark pair production in association with jets by the ATLAS experiment at the LHC are presented. The measurements are performed as functions of the top quark transverse momentum, the transverse momentum of the top quark-antitop quark system and the out-of-plane transverse momentum using data from $pp$ collisions at $\\sqrt{s}$=13 TeV collected by the ATLAS detector at the LHC in 2015 and corresponding to an integrated luminosity of 3.2 fb$^{-1}$. The top quark pair events are selected in the lepton (electron or muon) + jets channel. The measured cross sections, which are compared to several predictions, allow a detailed study of top quark production.

  1. Measurements of the vector boson production with the ATLAS detector

    Science.gov (United States)

    Lapertosa, A.

    2018-01-01

    Measurements of the Drell-Yan production of W and Z bosons at the LHC provide a benchmark of our understanding of perturbative QCD and probe the proton structure in a unique way. The ATLAS collaboration has performed new high precision measurements at a center-of-mass energy of 7 TeV. The measurements are performed for W+, W- and Z bosons integrated and as a function of the boson or lepton rapidity and the Z mass. Unprecedented precision is reached and strong constraints on Parton Distribution Functions, in particular the strange density are found. Z boson cross sections are also measured at center-of-mass energies of 8 TeV and 13 TeV, and cross-section ratios to the top-quark pair production have been derived. This ratio measurement leads to a cancellation of systematic effects and allows for a high precision comparison to the theory predictions. The production of jets in association with vector bosons is a further important process to study perturbative QCD in a multi-scale environment. The ATLAS collaboration has performed new measurements of Z boson plus jets cross sections, differential in several kinematic variables, in proton-proton collision data taken at a center-of-mass energy of 13 TeV. The measurements are compared to state-of-the art theory predictions. They are sensitive to higher-order pQCD effects, probe flavour and mass schemes and can be used to constrain the proton structure. In addition, a new measurement of the splitting scales of the kt jet-clustering algorithm for final states containing a Z boson candidate at a center-of-mass energy of 8 TeV is presented.

  2. Technical Design Report for the Phase-I Upgrade of the ATLAS TDAQ System

    CERN Document Server

    AUTHOR|(CDS)2069742; Abbott, Brad; Abdallah, Jalal; Abdel Khalek, Samah; Abdinov, Ovsat; Aben, Rosemarie; Abi, Babak; AbouZeid, Ossama; Abramowicz, Halina; Abreu, Henso; Abreu, Ricardo; Abulaiti, Yiming; Acharya, Bobby Samir; Achenbach, Ralf; Adamczyk, Leszek; Adams, David; Adelman, Jahred; Adomeit, Stefanie; Adye, Tim; Aefsky, Scott; Agatonovic-Jovin, Tatjana; Aguilar-Saavedra, Juan Antonio; Agustoni, Marco; Ahlen, Steven; Ahmad, Ashfaq; Ahmadov, Faig; Aielli, Giulio; Åkesson, Torsten Paul Ake; Akimoto, Ginga; Akimov, Andrei; Alam, Muhammad Aftab; Albert, Justin; Albrand, Solveig; Alconada Verzini, Maria Josefina; Aleksa, Martin; Aleksandrov, Igor; Alexa, Calin; Alexander, Gideon; Alexandre, Gauthier; Alexandrov, Evgeny; Alexopoulos, Theodoros; Alhroob, Muhammad; Alimonti, Gianluca; Alio, Lion; Alison, John; Allbrooke, Benedict; Allison, Lee John; Allport, Phillip; Allwood-Spiers, Sarah; Almond, John; Aloisio, Alberto; Alon, Raz; Alonso, Alejandro; Alonso, Francisco; Altheimer, Andrew David; Alvarez Gonzalez, Barbara; Alviggi, Mariagrazia; Amaral Coutinho, Yara; Amelung, Christoph; Amor Dos Santos, Susana Patricia; Amoroso, Simone; Amram, Nir; Amundsen, Glenn; Anastopoulos, Christos; Ancu, Lucian Stefan; Andari, Nansi; Andeen, Timothy; Anders, Christoph Falk; Anders, Gabriel; Anderson, John Thomas; Anderson, Kelby; Andreazza, Attilio; Andrei, George Victor; Anduaga, Xabier; Angelidakis, Stylianos; Angelozzi, Ivan; Anger, Philipp; Angerami, Aaron; Anisenkov, Alexey; Anjos, Nuno; Annovi, Alberto; Antonaki, Ariadni; Antonelli, Mario; Antonov, Alexey; Antos, Jaroslav; Anulli, Fabio; Aoki, Masato; Aperio Bella, Ludovica; Apolle, Rudi; Arabidze, Giorgi; Aracena, Ignacio; Arai, Yasuo; Araujo Ferraz, Victor; Arce, Ayana; Arguin, Jean-Francois; Argyropoulos, Spyridon; Arik, Metin; Armbruster, Aaron James; Arnaez, Olivier; Arnal, Vanessa; Arslan, Ozan; Artamonov, Andrei; Artoni, Giacomo; Asai, Shoji; Asbah, Nedaa; Ask, Stefan; Åsman, Barbro; Asquith, Lily; Assamagan, Ketevi; Astalos, Robert; Atkinson, Markus; Atlay, Naim Bora; Auerbach, Benjamin; Augsten, Kamil; Augusto, José; Aurousseau, Mathieu; Avolio, Giuseppe; Azuelos, Georges; Azuma, Yuya; Baak, Max; Baas, Alessandra; Bach, Andre; Bachacou, Henri; Bachas, Konstantinos; Backes, Moritz; Backhaus, Malte; Backus Mayes, John; Badescu, Elisabeta; Bagiacchi, Paolo; Bagnaia, Paolo; Bai, Yu; Bain, Travis; Baines, John; Baker, Oliver Keith; Baker, Sarah; Balek, Petr; Ballestrero, Sergio; Balli, Fabrice; Banas, Elzbieta; Banerjee, Swagato; Bangert, Andrea Michelle; Bansal, Vikas; Bansil, Hardeep Singh; Barak, Liron; Barber, Tom; Barberio, Elisabetta Luigia; Barberis, Dario; Barbero, Marlon; Barillari, Teresa; Barisonzi, Marcello; Barklow, Timothy; Barlow, Nick; Barnett, Bruce; Barnett, Michael; Baroncelli, Antonio; Barone, Gaetano; Barr, Alan; Barreiro, Fernando; Barreiro Guimarães da Costa, João; Bartoldus, Rainer; Barton, Adam Edward; Bartos, Pavol; Bartsch, Valeria; Bassalat, Ahmed; Basye, Austin; Bates, Richard; Batkova, Lucia; Batley, Richard; Batraneanu, Silvia; Battistin, Michele; Bauer, Florian; Bauss, Bruno; Bawa, Harinder Singh; Beacham, James Baker; Beau, Tristan; Beauchemin, Pierre-Hugues; Bechtle, Philip; Beck, Hans Peter; Becker, Anne Kathrin; Becker, Sebastian; Beckingham, Matthew; Beddall, Andrew; Beddall, Ayda; Bedikian, Sourpouhi; Bednyakov, Vadim; Bee, Christopher; Beemster, Lars; Beermann, Thomas; Begel, Michael; Behr, Katharina; Belanger-Champagne, Camille; Bell, Paul; Bell, William; Bella, Gideon; Bellagamba, Lorenzo; Bellerive, Alain; Bellomo, Massimiliano; Belloni, Alberto; Belotskiy, Konstantin; Beltramello, Olga; Benary, Odette; Benchekroun, Driss; Bendtz, Katarina; Benekos, Nektarios; Benhammou, Yan; Benhar Noccioli, Eleonora; Benitez Garcia, Jorge-Armando; Benjamin, Douglas; Bensinger, James; Bentvelsen, Stan; Beretta, Matteo; Berge, David; Bergeaas Kuutmann, Elin; Berger, Nicolas; Berghaus, Frank; Berglund, Elina; Beringer, Jürg; Bernard, Clare; Bernat, Pauline; Bernius, Catrin; Bernlochner, Florian Urs; Berry, Tracey; Berta, Peter; Bertella, Claudia; Bertelsen, Henrik; Bertolucci, Federico; Besana, Maria Ilaria; Besjes, Geert-Jan; Bessidskaia Bylund, Olga; Besson, Nathalie; Betancourt, Christopher; Bethke, Siegfried; Bhimji, Wahid; Bianchi, Riccardo-Maria; Bianchini, Louis; Bianco, Michele; Biebel, Otmar; Bieniek, Stephen Paul; Bierwagen, Katharina; Biesiada, Jed; Biglietti, Michela; Bilbao De Mendizabal, Javier; Bilokon, Halina; Bindi, Marcello; Binet, Sebastien; Bingul, Ahmet; Bini, Cesare; Bittner, Bernhard; Black, Curtis; Black, James; Black, Kevin; Blackburn, Daniel; Blair, Robert; Blanchard, Jean-Baptiste; Blazek, Tomas; Bloch, Ingo; Blocker, Craig; Blum, Walter; Blumenschein, Ulrike; Bobbink, Gerjan; Bobrovnikov, Victor; Bocchetta, Simona Serena; Bocci, Andrea; Boddy, Christopher Richard; Boehler, Michael; Boek, Jennifer; Boek, Thorsten Tobias; Bogdan, Mircea Arghir; Bogdanchikov, Alexander; Bohm, Christian; Boisvert, Veronique; Bold, Tomasz; Boldyrev, Alexey; Bolnet, Nayanka Myriam; Bomben, Marco; Bona, Marcella; Boonekamp, Maarten; Borga, Andrea; Borisov, Anatoly; Borissov, Guennadi; Borri, Marcello; Borroni, Sara; Bortfeldt, Jonathan; Bortolotto, Valerio; Bos, Kors; Boscherini, Davide; Bosman, Martine; Boterenbrood, Hendrik; Bouchami, Jihene; Boudreau, Joseph; Bouhova-Thacker, Evelina Vassileva; Boumediene, Djamel Eddine; Bourdarios, Claire; Bousson, Nicolas; Boutouil, Sara; Boveia, Antonio; Boyd, James; Boyko, Igor; Bozovic-Jelisavcic, Ivanka; Bracinik, Juraj; Brandt, Andrew; Brandt, Gerhard; Brandt, Oleg; Bratzler, Uwe; Brau, Benjamin; Brau, James; Brawn, Ian; Brazzale, Simone Federico; Brelier, Bertrand; Brendlinger, Kurt; Brennan, Amelia Jean; Brenner, Richard; Bressler, Shikma; Bristow, Kieran; Bristow, Timothy Michael; Britton, Dave; Brochu, Frederic; Brock, Ian; Brock, Raymond; Bronner, Johanna; Brooijmans, Gustaaf; Brooks, Timothy; Brooks, William; Brosamer, Jacquelyn; Brost, Elizabeth; Brown, Gareth; Brown, Jonathan; Bruckman de Renstrom, Pawel; Bruncko, Dusan; Bruneliere, Renaud; Brunet, Sylvie; Bruni, Alessia; Bruni, Graziano; Bruschi, Marco; Bryngemark, Lene; Buanes, Trygve; Buat, Quentin; Bucci, Francesca; Buchholz, Peter; Buckingham, Ryan; Buckley, Andrew; Buda, Stelian Ioan; Budagov, Ioulian; Budick, Burton; Buehrer, Felix; Bugge, Lars; Bugge, Magnar Kopangen; Bulekov, Oleg; Bundock, Aaron Colin; Bunse, Moritz; Burdin, Sergey; Burghgrave, Blake; Burke, Stephen; Burmeister, Ingo; Busato, Emmanuel; Büscher, Volker; Bussey, Peter; Buszello, Claus-Peter; Butler, Bart; Butler, John; Butt, Aatif Imtiaz; Buttar, Craig; Butterworth, Jonathan; Buttinger, William; Buzatu, Adrian; Byszewski, Marcin; Cabrera Urbán, Susana; Caforio, Davide; Cakir, Orhan; Calafiura, Paolo; Calderini, Giovanni; Calfayan, Philippe; Calkins, Robert; Caloba, Luiz; Calvet, David; Calvet, Samuel; Camacho Toro, Reina; Cameron, David; Caminada, Lea Michaela; Caminal Armadans, Roger; Campana, Simone; Campanelli, Mario; Campoverde, Angel; Canale, Vincenzo; Canelli, Florencia; Canepa, Anadi; Cantero, Josu; Cantrill, Robert; Cao, Tingting; Capeans Garrido, Maria Del Mar; Caprini, Irinel; Caprini, Mihai; Capua, Marcella; Caputo, Regina; Cardarelli, Roberto; Carli, Tancredi; Carlino, Gianpaolo; Carminati, Leonardo; Caron, Sascha; Carquin, Edson; Carrillo-Montoya, German D; Carter, Janet; Casadei, Diego; Casado, Maria Pilar; Castaneda-Miranda, Elizabeth; Castelli, Angelantonio; Castillo Gimenez, Victoria; Castro, Nuno Filipe; Catastini, Pierluigi; Catinaccio, Andrea; Catmore, James; Cattai, Ariella; Cattani, Giordano; Caughron, Seth; Cavaliere, Viviana; Cavalli, Donatella; Cavalli-Sforza, Matteo; Cavasinni, Vincenzo; Ceradini, Filippo; Cerio, Benjamin; Cerny, Karel; Cerqueira, Augusto Santiago; Cerri, Alessandro; Cerrito, Lucio; Cerutti, Fabio; Cerv, Matevz; Cervelli, Alberto; Cetin, Serkant Ali; Chafaq, Aziz; Chakraborty, Dhiman; Chalupkova, Ina; Chan, Kevin; Chang, Philip; Chapleau, Bertrand; Chapman, John Derek; Charfeddine, Driss; Charlton, Dave; Chavda, Vikash; Chavez Barajas, Carlos Alberto; Cheatham, Susan; Chekanov, Sergei; Chekulaev, Sergey; Chelkov, Gueorgui; Chelstowska, Magda Anna; Chen, Chunhui; Chen, Hucheng; Chen, Karen; Chen, Liming; Chen, Shenjian; Chen, Xin; Chen, Yujiao; Cheng, Hok Chuen; Cheng, Yangyang; Cheplakov, Alexander; Cherkaoui El Moursli, Rajaa; Chernyatin, Valeriy; Cheu, Elliott; Chevalier, Laurent; Chiarella, Vitaliano; Chiefari, Giovanni; Childers, John Taylor; Chilingarov, Alexandre; Chiodini, Gabriele; Chisholm, Andrew; Chislett, Rebecca Thalatta; Chitan, Adrian; Chizhov, Mihail; Chouridou, Sofia; Chow, Bonnie Kar Bo; Christidi, Ilektra-Athanasia; Chudoba, Jiri; Ciapetti, Guido; Ciftci, Abbas Kenan; Ciftci, Rena; Cinca, Diane; Cindro, Vladimir; Ciocio, Alessandra; Ciodaro Xavier, Thiago; Cirkovic, Predrag; Citraro, Saverio; Citron, Zvi Hirsh; Citterio, Mauro; Ciubancan, Mihai; Clark, Allan G; Clark, Philip James; Clarke, Robert; Clemens, Jean-Claude; Clement, Benoit; Clement, Christophe; Coadou, Yann; Cobal, Marina; Coccaro, Andrea; Cochran, James H; Coffey, Laurel; Cogan, Joshua Godfrey; Coggeshall, James; Cole, Brian; Cole, Stephen; Colijn, Auke-Pieter; Collins-Tooth, Christopher; Collot, Johann; Colombo, Tommaso; Colon, German; Compostella, Gabriele; Conde Muiño, Patricia; Coniavitis, Elias; Conidi, Maria Chiara; Connelly, Ian; Consonni, Sofia Maria; Consorti, Valerio; Constantinescu, Serban; Conti, Geraldine; Conventi, Francesco; Cooke, Mark; Cooper, Ben; Cooper-Sarkar, Amanda; Cooper-Smith, Neil; Copic, Katherine; Cornelissen, Thijs; Corradi, Massimo; Corriveau, Francois; Corso-Radu, Alina; Cortes-Gonzalez, Arely; Cortiana, Giorgio; Costa, Giuseppe; Costa, María José; Costanzo, Davide; Côté, David; Cottin, Giovanna; Coura Torres, Rodrigo; Cowan, Glen; Cox, Brian; Cranmer, Kyle; Cree, Graham; Crépé-Renaudin, Sabine; Crescioli, Francesco; Crispin Ortuzar, Mireia; Cristinziani, Markus; Crone, Gordon Jeremy; Crosetti, Giovanni; Cuciuc, Constantin-Mihai; Cuenca Almenar, Cristóbal; Cuhadar Donszelmann, Tulay; Cummings, Jane; Curatolo, Maria; Cuthbert, Cameron; Czirr, Hendrik; Czodrowski, Patrick; Czyczula, Zofia; D'Auria, Saverio; D'Onofrio, Monica; D'Orazio, Alessia; Da Cunha Sargedas De Sousa, Mario Jose; Da Via, Cinzia; Dabrowski, Wladyslaw; Dafinca, Alexandru; Dai, Tiesheng; Dale, Orjan; Dallaire, Frederick; Dallapiccola, Carlo; Dam, Mogens; Daniells, Andrew Christopher; Dano Hoffmann, Maria; Dao, Valerio; Darbo, Giovanni; Darlea, Georgiana Lavinia; Darmora, Smita; Dassoulas, James; Davey, Will; David, Claire; Davidek, Tomas; Davies, Eleanor; Davies, Merlin; Davignon, Olivier; Davison, Adam; Davygora, Yuriy; Dawe, Edmund; Dawson, Ian; Daya-Ishmukhametova, Rozmin; De, Kaushik; de Asmundis, Riccardo; De Castro, Stefano; De Cecco, Sandro; de Graat, Julien; De Groot, Nicolo; de Jong, Paul; De la Torre, Hector; De Lorenzi, Francesco; De Nooij, Lucie; De Pedis, Daniele; De Salvo, Alessandro; De Sanctis, Umberto; De Santo, Antonella; De Vivie De Regie, Jean-Baptiste; Dearnaley, William James; Debbe, Ramiro; Debenedetti, Chiara; Dechenaux, Benjamin; Dedovich, Dmitri; Degenhardt, James; Deigaard, Ingrid; Del Peso, Jose; Del Prete, Tarcisio; Delemontex, Thomas; Deliot, Frederic; Deliyergiyev, Maksym; Dell'Acqua, Andrea; Dell'Asta, Lidia; Dell'Orso, Mauro; Della Pietra, Massimo; della Volpe, Domenico; Delmastro, Marco; Delsart, Pierre-Antoine; Deluca, Carolina; Demers, Sarah; Demichev, Mikhail; Demilly, Aurelien; Denisov, Sergey; Derendarz, Dominik; Derkaoui, Jamal Eddine; Derue, Frederic; Dervan, Paul; Desch, Klaus Kurt; Deviveiros, Pier-Olivier; Dewhurst, Alastair; Dhaliwal, Saminder; Di Ciaccio, Anna; Di Ciaccio, Lucia; Di Domenico, Antonio; Di Donato, Camilla; Di Girolamo, Alessandro; Di Girolamo, Beniamino; Di Mattia, Alessandro; Di Micco, Biagio; Di Nardo, Roberto; Di Simone, Andrea; Di Sipio, Riccardo; Di Valentino, David; Diaz, Marco Aurelio; Diehl, Edward; Dietrich, Janet; Dietzsch, Thorsten; Diglio, Sara; Dimitrievska, Aleksandra; Dingfelder, Jochen; Dionisi, Carlo; Dita, Petre; Dita, Sanda; Dittus, Fridolin; Djama, Fares; Djobava, Tamar; do Vale, Maria Aline Barros; Doan, Thi Kieu Oanh; Dobos, Daniel; Dobson, Ellie; Doglioni, Caterina; Doherty, Tom; Dohmae, Takeshi; Dolejsi, Jiri; Dolezal, Zdenek; Donadelli, Marisilvia; Donati, Simone; Dondero, Paolo; Donini, Julien; Dopke, Jens; Doria, Alessandra; Dova, Maria-Teresa; Doyle, Tony; Drake, Gary; Dris, Manolis; Dubbert, Jörg; Dube, Sourabh; Dubreuil, Emmanuelle; Duchovni, Ehud; Duckeck, Guenter; Ducu, Otilia Anamaria; Duda, Dominik; Dudarev, Alexey; Dudziak, Fanny; Duflot, Laurent; Duguid, Liam; Dührssen, Michael; Dunford, Monica; Duran Yildiz, Hatice; Düren, Michael; Dwuznik, Michal; Ebke, Johannes; Edmunds, Daniel; Edson, William; Edwards, Clive; Edwards, Nicholas Charles; Ehrenfeld, Wolfgang; Eifert, Till; Eigen, Gerald; Einsweiler, Kevin; Ekelof, Tord; El Kacimi, Mohamed; Ellert, Mattias; Elles, Sabine; Ellinghaus, Frank; Ellis, Nicolas; Elmsheuser, Johannes; Elsing, Markus; Emeliyanov, Dmitry; Enari, Yuji; Endner, Oliver Chris; Endo, Masaki; Erdmann, Johannes; Ereditato, Antonio; Ermoline, Iouri; Ernis, Gunar; Ernst, Jesse; Ernst, Michael; Ernwein, Jean; Errede, Deborah; Errede, Steven; Ertel, Eugen; Escalier, Marc; Esch, Hendrik; Escobar, Carlos; Esposito, Bellisario; Etienvre, Anne-Isabelle; Etzion, Erez; Evangelakou, Despoina; Evans, Hal; Fabbri, Laura; Facini, Gabriel; Fakhrutdinov, Rinat; Falciano, Speranza; Faltova, Jana; Fang, Yaquan; Fanti, Marcello; Farbin, Amir; Farilla, Addolorata; Farooque, Trisha; Farrell, Steven; Farrington, Sinead; Farthouat, Philippe; Fassi, Farida; Fassnacht, Patrick; Fassouliotis, Dimitrios; Fatholahzadeh, Baharak; Faulkner, Peter; Favareto, Andrea; Fayard, Louis; Federic, Pavol; Fedin, Oleg; Fedorko, Wojciech; Fehling-Kaschek, Mirjam; Feigl, Simon; Feligioni, Lorenzo; Feng, Cunfeng; Feng, Eric; Feng, Haolu; Fenyuk, Alexander; Fernandez Perez, Sonia; Ferrag, Samir; Ferrando, James; Ferrara, Valentina; Ferrari, Arnaud; Ferrari, Pamela; Ferrari, Roberto; Ferreira de Lima, Danilo Enoque; Ferrer, Antonio; Ferrere, Didier; Ferretti, Claudio; Ferretto Parodi, Andrea; Fiascaris, Maria; Fiedler, Frank; Filipčič, Andrej; Filipuzzi, Marco; Filthaut, Frank; Fincke-Keeler, Margret; Finelli, Kevin Daniel; Fiolhais, Miguel; Fiorini, Luca; Firan, Ana; Fischer, Julia; Fisher, Matthew; Fitzgerald, Eric Andrew; Flechl, Martin; Fleck, Ivor; Fleischmann, Philipp; Fleischmann, Sebastian; Fletcher, Gareth Thomas; Fletcher, Gregory; Flick, Tobias; Floderus, Anders; Flores Castillo, Luis; Florez Bustos, Andres Carlos; Flowerdew, Michael; Formica, Andrea; Forti, Alessandra; Fortin, Dominique; Fournier, Daniel; Fox, Harald; Francavilla, Paolo; Franchini, Matteo; Franchino, Silvia; Francis, David; Franklin, Melissa; Franz, Sebastien; Fraternali, Marco; Fratina, Sasa; French, Sky; Friedrich, Conrad; Friedrich, Felix; Froidevaux, Daniel; Front, David Moris; Frost, James; Fukunaga, Chikara; Fullana Torregrosa, Esteban; Fulsom, Bryan Gregory; Fusayasu, Takahiro; Fuster, Juan; Gabaldon, Carolina; Gabizon, Ofir; Gabrielli, Alessandro; Gabrielli, Andrea; Gadatsch, Stefan; Gadfort, Thomas; Gadomski, Szymon; Gagliardi, Guido; Gagnon, Pauline; Galea, Cristina; Galhardo, Bruno; Gallas, Elizabeth; Gallo, Valentina Santina; Gallop, Bruce; Gallus, Petr; Galster, Gorm Aske Gram Krohn; Gan, KK; Gandrajula, Reddy Pratap; Gao, Jun; Gao, Yongsheng; Garay Walls, Francisca; Garberson, Ford; García, Carmen; García Navarro, José Enrique; Garcia-Sciveres, Maurice; Gardner, Robert; Garelli, Nicoletta; Garonne, Vincent; Gatti, Claudio; Gaudio, Gabriella; Gaur, Bakul; Gauthier, Lea; Gavrilenko, Igor; Gay, Colin; Gaycken, Goetz; Gazis, Evangelos; Ge, Peng; Gecse, Zoltan; Gee, Norman; Geerts, Daniël Alphonsus Adrianus; Geich-Gimbel, Christoph; Gellerstedt, Karl; Gemme, Claudia; Gemmell, Alistair; Genest, Marie-Hélène; Gentile, Simonetta; Gentsos, Christos; George, Matthias; George, Simon; Gerbaudo, Davide; Gershon, Avi; Ghibaudi, Marco; Ghodbane, Nabil; Giacobbe, Benedetto; Giagu, Stefano; Giangiobbe, Vincent; Giannetti, Paola; Gianotti, Fabiola; Gibson, Stephen; Gillam, Thomas; Gillberg, Dag; Gingrich, Douglas; Giokaris, Nikos; Giordani, MarioPaolo; Giordano, Raffaele; Giorgi, Francesco Michelangelo; Giovannini, Paola; Giraud, Pierre-Francois; Giugni, Danilo; Giuliani, Claudia; Giulini, Maddalena; Giunta, Michele; Gjelsten, Børge Kile; Gkaitatzis, Stamatios; Gkialas, Ioannis; Gladilin, Leonid; Glasman, Claudia; Glatzer, Julian; Glazov, Alexandre; Glonti, George; Goblirsch-Kolb, Maximilian; Goddard, Jack Robert; Godfrey, Jennifer; Godlewski, Jan; Goeringer, Christian; Goldfarb, Steven; Golling, Tobias; Golubkov, Dmitry; Gomes, Agostinho; Gomez Fajardo, Luz Stella; Gonçalo, Ricardo; Goncalves Gama, Rafael; Goncalves Pinto Firmino Da Costa, Joao; Gonella, Laura; González de la Hoz, Santiago; Gonzalez Parra, Garoe; Gonzalez Silva, Laura; Gonzalez-Sevilla, Sergio; Goossens, Luc; Gorbounov, Petr Andreevich; Gordon, Howard; Gorelov, Igor; Gorini, Benedetto; Gorini, Edoardo; Gorišek, Andrej; Gornicki, Edward; Goshaw, Alfred; Gössling, Claus; Gostkin, Mikhail Ivanovitch; Gouighri, Mohamed; Goujdami, Driss; Goulette, Marc Phillippe; Goussiou, Anna; Goy, Corinne; Gozpinar, Serdar; Grabas, Herve Marie Xavier; Graber, Lars; Grabowska-Bold, Iwona; Grafström, Per; Grahn, Karl-Johan; Gramling, Johanna; Gramstad, Eirik; Grancagnolo, Sergio; Grassi, Valerio; Gratchev, Vadim; Gray, Heather; Gray, Julia Ann; Graziani, Enrico; Grebenyuk, Oleg; Green, Barry; Greenwood, Zeno Dixon; Gregersen, Kristian; Gregor, Ingrid-Maria; Grenier, Philippe; Griffiths, Justin; Grillo, Alexander; Grimm, Kathryn; Grinstein, Sebastian; Gris, Philippe Luc Yves; Grishkevich, Yaroslav; Grivaz, Jean-Francois; Grohs, Johannes Philipp; Grohsjean, Alexander; Gross, Eilam; Grosse-Knetter, Joern; Grossi, Giulio Cornelio; Groth-Jensen, Jacob; Grout, Zara Jane; Grybel, Kai; Guan, Liang; Guescini, Francesco; Guest, Daniel; Gueta, Orel; Guicheney, Christophe; Guido, Elisa; Guillemin, Thibault; Guindon, Stefan; Gul, Umar; Gumpert, Christian; Gunther, Jaroslav; Guo, Jun; Gupta, Shaun; Gutierrez, Phillip; Gutierrez Ortiz, Nicolas Gilberto; Gutschow, Christian; Guttman, Nir; Guyot, Claude; Gwenlan, Claire; Gwilliam, Carl; Haas, Andy; Haas, Stefan; Haber, Carl; Hadavand, Haleh Khani; Haefner, Petra; Hageböck, Stephan; Hakobyan, Hrachya; Haleem, Mahsana; Hall, David; Halladjian, Garabed; Hamacher, Klaus; Hamal, Petr; Hamano, Kenji; Hamer, Matthias; Hamilton, Andrew; Hamilton, Samuel; Han, Liang; Hanagaki, Kazunori; Hanawa, Keita; Hance, Michael; Hanke, Paul; Hansen, Jørgen Beck; Hansen, Jorn Dines; Hansen, Peter Henrik; Hansson, Per; Hara, Kazuhiko; Hard, Andrew; Harenberg, Torsten; Harkusha, Siarhei; Harper, Devin; Harrington, Robert; Harris, Orin; Harrison, Paul Fraser; Hartjes, Fred; Hasegawa, Satoshi; Hasegawa, Yoji; Hassani, Samira; Haug, Sigve; Hauschild, Michael; Hauser, Reiner; Havranek, Miroslav; Hawkes, Christopher; Hawkings, Richard John; Hawkins, Anthony David; Hayashi, Takayasu; Hayden, Daniel; Hays, Chris; Hayward, Helen; Haywood, Stephen; Head, Simon; Heck, Tobias; Hedberg, Vincent; Heelan, Louise; Heim, Sarah; Heim, Timon; Heinemann, Beate; Heinrich, Lukas; Heisterkamp, Simon; Hejbal, Jiri; Helary, Louis; Heller, Claudio; Heller, Matthieu; Hellman, Sten; Hellmich, Dennis; Helsens, Clement; Henderson, James; Henderson, Robert; Hengler, Christopher; Henrichs, Anna; Henriques Correia, Ana Maria; Henrot-Versille, Sophie; Hensel, Carsten; Herbert, Geoffrey Henry; Hernández Jiménez, Yesenia; Herrberg-Schubert, Ruth; Herten, Gregor; Hertenberger, Ralf; Hervas, Luis; Hesketh, Gavin Grant; Hessey, Nigel; Hickling, Robert; Higón-Rodriguez, Emilio; Higuchi, Kota; Hill, John; Hiller, Karl Heinz; Hillert, Sonja; Hillier, Stephen; Hinchliffe, Ian; Hines, Elizabeth; Hirose, Minoru; Hirschbuehl, Dominic; Hobbs, John; Hod, Noam; Hodgkinson, Mark; Hodgson, Paul; Hoecker, Andreas; Hoeferkamp, Martin; Hoffman, Julia; Hoffmann, Dirk; Hofmann, Julia Isabell; Hohlfeld, Marc; Holmes, Tova Ray; Hong, Tae Min; Hooft van Huysduynen, Loek; Hostachy, Jean-Yves; Hou, Suen; Hoummada, Abdeslam; Howard, Jacob; Howarth, James; Hrabovsky, Miroslav; Hristova, Ivana; Hrivnac, Julius; Hryn'ova, Tetiana; Hsu, Pai-hsien Jennifer; Hsu, Shih-Chieh; Hu, Diedi; Hu, Xueye; Huang, Yanping; Hubacek, Zdenek; Hubaut, Fabrice; Huegging, Fabian; Huettmann, Antje; Huffman, Todd Brian; Hughes, Emlyn; Hughes, Gareth; Huhtinen, Mika; Hülsing, Tobias Alexander; Hurwitz, Martina; Huseynov, Nazim; Huston, Joey; Huth, John; Iacobucci, Giuseppe; Iakovidis, Georgios; Ibragimov, Iskander; Iconomidou-Fayard, Lydia; Ideal, Emma; Iengo, Paolo; Igonkina, Olga; Iizawa, Tomoya; Ikegami, Yoichi; Ikematsu, Katsumasa; Ikeno, Masahiro; Ilchenko, Iurii; Iliadis, Dimitrios; Ilic, Nikolina; Inamaru, Yuki; Ince, Tayfun; Ioannou, Pavlos; Iodice, Mauro; Iordanidou, Kalliopi; Ippolito, Valerio; Irles Quiles, Adrian; Isaksson, Charlie; Ishino, Masaya; Ishitsuka, Masaki; Ishmukhametov, Renat; Issever, Cigdem; Istin, Serhat; Iturbe Ponce, Julia Mariana; Ivashin, Anton; Iwanski, Wieslaw; Iwasaki, Hiroyuki; Izen, Joseph; Izzo, Vincenzo; Jackson, Brett; Jackson, John; Jackson, Matthew; Jackson, Paul; Jaekel, Martin; Jain, Vivek; Jakobi, Katharina Bianca; Jakobs, Karl; Jakobsen, Sune; Jakoubek, Tomas; Jakubek, Jan; Jamin, David Olivier; Jana, Dilip; Jansen, Eric; Jansen, Hendrik; Janssen, Jens; Jansweijer, Peter Paul Maarten; Janus, Michel; Jarlskog, Göran; Jeanty, Laura; Jeng, Geng-yuan; Jennens, David; Jenni, Peter; Jentzsch, Jennifer; Jeske, Carl; Jézéquel, Stéphane; Jha, Manoj Kumar; Ji, Haoshuang; Ji, Weina; Jia, Jiangyong; Jiang, Yi; Jimenez Belenguer, Marcos; Jin, Shan; Jinaru, Adam; Jinnouchi, Osamu; Joergensen, Morten Dam; Johansson, Erik; Johansson, Per; Johns, Kenneth; Jon-And, Kerstin; Jones, Graham; Jones, Roger; Jones, Tim; Joos, Markus; Jorge, Pedro; Joshi, Kiran Daniel; Jovicevic, Jelena; Ju, Xiangyang; Jung, Christian; Jungst, Ralph Markus; Jussel, Patrick; Juste Rozas, Aurelio; Kaci, Mohammed; Kaczmarska, Anna; Kado, Marumi; Kagan, Harris; Kagan, Michael; Kahra, Christian; Kajomovitz, Enrique; Kaluza, Adam; Kama, Sami; Kanaya, Naoko; Kaneda, Michiru; Kaneti, Steven; Kanno, Takayuki; Kantserov, Vadim; Kanzaki, Junichi; Kaplan, Benjamin; Kapliy, Anton; Kar, Deepak; Karakostas, Konstantinos; Karastathis, Nikolaos; Karnevskiy, Mikhail; Karpov, Sergey; Karthik, Krishnaiyengar; Kartvelishvili, Vakhtang; Karyukhin, Andrey; Kashif, Lashkar; Kasieczka, Gregor; Kass, Richard; Kastanas, Alex; Kataoka, Yousuke; Katre, Akshay; Katzy, Judith; Kaushik, Venkatesh; Kawagoe, Kiyotomo; Kawamoto, Tatsuo; Kawamura, Gen; Kazama, Shingo; Kazanin, Vassili; Kazarinov, Makhail; Kazarov, Andrei; Keeler, Richard; Kehoe, Robert; Keil, Markus; Keller, John; Kempster, Jacob Julian; Keoshkerian, Houry; Kepka, Oldrich; Kerševan, Borut Paul; Kersten, Susanne; Kessoku, Kohei; Keung, Justin; Keyes, Robert; Khalil-zada, Farkhad; Khandanyan, Hovhannes; Khanov, Alexander; Kharchenko, Dmitri; Khodinov, Alexander; Khoo, Teng Jian; Khoriauli, Gia; Khoroshilov, Andrey; Khovanskiy, Valery; Khramov, Evgeniy; Khubua, Jemal; Kiese, Patric Karl; Kim, Hyeon Jin; Kim, Shinhong; Kimura, Naoki; Kind, Oliver; King, Barry; King, Matthew; King, Robert Steven Beaufoy; King, Samuel Burton; Kirk, Julie; Kiryunin, Andrey; Kishimoto, Tomoe; Kisielewska, Danuta; Kitamura, Takumi; Kiuchi, Kenji; Kladiva, Eduard; Klein, Max; Klein, Uta; Kleinknecht, Konrad; Klimek, Pawel; Klimentov, Alexei; Klimkovich, Tatsiana; Klingenberg, Reiner; Klinger, Joel Alexander; Klioutchnikova, Tatiana; Klok, Peter; Kluge, Eike-Erik; Kluit, Peter; Kluth, Stefan; Kneringer, Emmerich; Knoops, Edith; Knue, Andrea; Kobayashi, Dai; Kobayashi, Tomio; Kobel, Michael; Kocian, Martin; Kodys, Peter; Koevesarki, Peter; Koffas, Thomas; Koffeman, Els; Kogan, Lucy Anne; Kohlmann, Simon; Kohout, Zdenek; Kohriki, Takashi; Koi, Tatsumi; Kolanoski, Hermann; Koletsou, Iro; Koll, James; Kolos, Serguei; Komar, Aston; Komori, Yuto; Kondo, Takahiko; Köneke, Karsten; König, Adriaan; K{ö}nig, Sebastian; Kono, Takanori; Konoplich, Rostislav; Konstantinidis, Nikolaos; Kopeliansky, Revital; Koperny, Stefan; Köpke, Lutz; Kopp, Anna Katharina; Korcyl, Krzysztof; Kordas, Kostantinos; Korn, Andreas; Korol, Aleksandr; Korolkov, Ilya; Korolkova, Elena; Korotkov, Vladislav; Kortner, Oliver; Kortner, Sandra; Kostyukhin, Vadim; Kotov, Vladislav; Kotwal, Ashutosh; Kourkoumelis, Christine; Kouskoura, Vasiliki; Koutsman, Alex; Kowalewski, Robert Victor; Kowalski, Tadeusz; Kozanecki, Witold; Kozhin, Anatoly; Kral, Vlastimil; Kramarenko, Viktor; Kramberger, Gregor; Krasnopevtsev, Dimitriy; Krasny, Mieczyslaw Witold; Krasznahorkay, Attila; Kraus, Jana; Kravchenko, Anton; Kreiss, Sven; Kretzschmar, Jan; Kreutzfeldt, Kristof; Krieger, Nina; Krieger, Peter; Kroeninger, Kevin; Kroha, Hubert; Kroll, Joe; Kroseberg, Juergen; Krstic, Jelena; Kruchonak, Uladzimir; Krüger, Hans; Kruker, Tobias; Krumnack, Nils; Krumshteyn, Zinovii; Kruse, Amanda; Kruse, Mark; Kruskal, Michael; Kubota, Takashi; Kuday, Sinan; Kuehn, Susanne; Kugel, Andreas; Kuhl, Andrew; Kuhl, Thorsten; Kukhtin, Victor; Kulchitsky, Yuri; Kuleshov, Sergey; Kuna, Marine; Kunigo, Takuto; Kunkle, Joshua; Kupco, Alexander; Kurashige, Hisaya; Kurochkin, Yurii; Kurumida, Rie; Kus, Vlastimil; Kuwertz, Emma Sian; Kuze, Masahiro; Kvita, Jiri; La Rosa, Alessandro; La Rotonda, Laura; Lablak, Said; Lacasta, Carlos; Lacava, Francesco; Lacey, James; Lacker, Heiko; Lacour, Didier; Lacuesta, Vicente Ramón; Ladygin, Evgueni; Lafaye, Remi; Laforge, Bertrand; Lagouri, Theodota; Lai, Stanley; Laier, Heiko; Laisne, Emmanuel; Lambourne, Luke; Lampen, Caleb; Lampl, Walter; Lançon, Eric; Landgraf, Ulrich; Landon, Murrough; Lang, Valerie Susanne; Lange, Clemens; Lankford, Andrew; Lanni, Francesco; Lantzsch, Kerstin; Lanza, Agostino; Laplace, Sandrine; Lapoire, Cecile; Laporte, Jean-Francois; Lari, Tommaso; Larner, Aimee; Lassnig, Mario; Laurelli, Paolo; Laurens, Philippe; Lavorini, Vincenzo; Lavrijsen, Wim; Laycock, Paul; Le, Bao Tran; Le Dortz, Olivier; Le Guirriec, Emmanuel; Le Menedeu, Eve; LeCompte, Thomas; Ledroit-Guillon, Fabienne Agnes Marie; Lee, Claire, Alexandra; Lee, Hurng-Chun; Lee, Jason; Lee, Shih-Chang; Lee, Lawrence; Lefebvre, Guillaume; Lefebvre, Michel; Legger, Federica; Leggett, Charles; Lehan, Allan; Lehmacher, Marc; Lehmann Miotto, Giovanna; Lei, Xiaowen; Leister, Andrew Gerard; Leite, Marco Aurelio Lisboa; Leitner, Rupert; Lellouch, Daniel; Lemmer, Boris; Leney, Katharine; Lenz, Tatjana; Lenzi, Bruno; Leone, Robert; Leonhardt, Kathrin; Leonidopoulos, Christos; Leontsinis, Stefanos; Leroy, Claude; Lester, Christopher; Lester, Christopher Michael; Levêque, Jessica; Levin, Daniel; Levinson, Lorne; Lewis, Adrian; Lewis, George; Leyko, Agnieszka; Leyton, Michael; Li, Bing; Li, Bo; Li, Haifeng; Li, Ho Ling; Li, Shu; Li, Xuefei; Liang, Zhijun; Liao, Hongbo; Liberali, Valentino; Liberti, Barbara; Lie, Ki; Liebal, Jessica; Liebig, Wolfgang; Limbach, Christian; Limosani, Antonio; Lin, Simon; Linde, Frank; Lindquist, Brian Edward; Linnemann, James; Lipeles, Elliot; Lipniacka, Anna; Lisovyi, Mykhailo; Liss, Tony; Lister, Alison; Litke, Alan; Liu, Bo; Liu, Dong; Liu, Jianbei; Liu, Kun; Liu, Lulu; Liu, Miaoyuan; Liu, Minghui; Liu, Yanwen; Livan, Michele; Livermore, Sarah; Lleres, Annick; Llorente Merino, Javier; Lloyd, Stephen; Lo Sterzo, Francesco; Lobodzinska, Ewelina; Loch, Peter; Lockman, William; Loddenkoetter, Thomas; Loebinger, Fred; Loevschall-Jensen, Ask Emil; Loginov, Andrey; Loh, Chang Wei; Lohse, Thomas; Lohwasser, Kristin; Lokajicek, Milos; Lombardo, Vincenzo Paolo; Long, Brian Alexander; Long, Jonathan; Long, Robin Eamonn; Lopes, Lourenco; Lopez Mateos, David; Lopez Paredes, Brais; Lorenz, Jeanette; Lorenzo Martinez, Narei; Losada, Marta; Loscutoff, Peter; Lou, XinChou; Lounis, Abdenour; Love, Jeremy; Love, Peter; Lowe, Andrew; Lu, Feng; Lubatti, Henry; Luci, Claudio; Luciano, Pierluigi; Lucotte, Arnaud; Ludwig, Dörthe; Luehring, Frederick; Lukas, Wolfgang; Luminari, Lamberto; Lundberg, Johan; Lundberg, Olof; Lund-Jensen, Bengt; Lungwitz, Matthias; Luongo, Carmela; Lupu, Nachman; Lynn, David; Lysak, Roman; Lytken, Else; Ma, Hong; Ma, Lian Liang; Maccarrone, Giovanni; Macchiolo, Anna; Maček, Boštjan; Macey, Tom; Machado Miguens, Joana; Macina, Daniela; Madar, Romain; Maddocks, Harvey Jonathan; Mader, Wolfgang; Madsen, Alexander; Maeno, Mayuko; Maeno, Tadashi; Magnoni, Luca; Magradze, Erekle; Mahboubi, Kambiz; Mahlstedt, Joern; Mahmoud, Sara; Maiani, Camilla; Maidantchik, Carmen; Maio, Amélia; Majewski, Stephanie; Makida, Yasuhiro; Makovec, Nikola; Mal, Prolay; Malaescu, Bogdan; Maldaner, Stephan; Malecki, Pawel; Maleev, Victor; Malek, Fairouz; Mallik, Usha; Malon, David; Malone, Caitlin; Maltezos, Stavros; Malyshev, Vladimir; Malyukov, Sergei; Mamuzic, Judita; Mandelli, Beatrice; Mandelli, Luciano; Mandić, Igor; Mandrysch, Rocco; Maneira, José; Manfredini, Alessandro; Manhaes de Andrade Filho, Luciano; Manjarres Ramos, Joany Andreina; Mann, Alexander; Manning, Peter; Manousakis-Katsikakis, Arkadios; Mansoulie, Bruno; Mantifel, Rodger; Mapelli, Livio; March, Luis; Marchand, Jean-Francois; Marchese, Fabrizio; Marchiori, Giovanni; Marcisovsky, Michal; Marino, Christopher; Marroquim, Fernando; Marshall, Zach; Marti, Lukas Fritz; Marti-Garcia, Salvador; Martin, Brian; Martin, Brian; Martin, Tim; Martin, Victoria Jane; Martin dit Latour, Bertrand; Martinez, Homero; Martinez, Mario; Martin-Haugh, Stewart; Martyniuk, Alex; Marx, Marilyn; Marzin, Antoine; Masetti, Lucia; Mashimo, Tetsuro; Mashinistov, Ruslan; Masik, Jiri; Maslennikov, Alexey; Massa, Ignazio; Massa, Lorenzo; Massol, Nicolas; Mastrandrea, Paolo; Mastroberardino, Anna; Masubuchi, Tatsuya; Matsunaga, Hiroyuki; Matsushita, Takashi; Mättig, Peter; Mättig, Stefan; Mattmann, Johannes; Mattravers, Carly; Maurer, Julien; Maxfield, Stephen; Maximov, Dmitriy; Mazini, Rachid; Mazzaferro, Luca; Mc Goldrick, Garrin; Mc Kee, Shawn Patrick; McCarn, Allison; McCarthy, Robert; McCarthy, Tom; Mcfayden, Josh; Mchedlidze, Gvantsa; Mclaughlan, Tom; McMahon, Steve; McPherson, Robert; Meade, Andrew; Mechnich, Joerg; Medinnis, Michael; Meehan, Samuel; Meera-Lebbai, Razzak; Meessen, Christophe; Mehlhase, Sascha; Mehta, Andrew; Meier, Karlheinz; Meineck, Christian; Meirose, Bernhard; Melachrinos, Constantinos; Mellado Garcia, Bruce Rafael; Meloni, Federico; Mendoza Navas, Luis; Mengarelli, Alberto; Menke, Sven; Meoni, Evelin; Mercurio, Kevin Michael; Mergelmeyer, Sebastian; Meric, Nicolas; Mermod, Philippe; Merola, Leonardo; Meroni, Chiara; Merritt, Frank; Merritt, Hayes; Messina, Andrea; Metcalfe, Jessica; Mete, Alaettin Serhan; Meyer, Carsten; Meyer, Christopher; Meyer, Jean-Pierre; Meyer, Jochen; Meyer, Joerg; Middleton, Robin; Migas, Sylwia; Mijović, Liza; Mikenberg, Giora; Mikestikova, Marcela; Mikuž, Marko; Miller, David; Mills, Corrinne; Milov, Alexander; Milstead, David; Milstein, Dmitry; Minaenko, Andrey; Miñano Moya, Mercedes; Minashvili, Irakli; Mincer, Allen; Mindur, Bartosz; Mineev, Mikhail; Ming, Yao; Mir, Lluisa-Maria; Mitani, Takashi; Mitrevski, Jovan; Mitsou, Vasiliki A; Mitsui, Shingo; Miucci, Antonio; Miyagawa, Paul; Mjörnmark, Jan-Ulf; Moa, Torbjoern; Moeller, Victoria; Mohapatra, Soumya; Molander, Simon; Moles-Valls, Regina; Mönig, Klaus; Monini, Caterina; Monk, James; Monnier, Emmanuel; Montejo Berlingen, Javier; Monticelli, Fernando; Monzani, Simone; Moore, Roger; Mora Herrera, Clemencia; Moraes, Arthur; Morange, Nicolas; Morel, Julien; Moreno, Deywis; Moreno Llácer, María; Morettini, Paolo; Morgenstern, Marcus; Morii, Masahiro; Moritz, Sebastian; Morley, Anthony Keith; Mornacchi, Giuseppe; Morris, John; Morvaj, Ljiljana; Moser, Hans-Guenther; Mosidze, Maia; Moss, Josh; Mount, Richard; Mountricha, Eleni; Moyse, Edward; Muanza, Steve; Mudd, Richard; Mueller, Felix; Mueller, James; Mueller, Klemens; Mueller, Thibaut; Mueller, Timo; Muenstermann, Daniel; Munwes, Yonathan; Murillo Garcia, Raul; Murillo Quijada, Javier Alberto; Murray, Bill; Mussche, Ido; Musto, Elisa; Myagkov, Alexey; Myska, Miroslav; Nackenhorst, Olaf; Nadal, Jordi; Nagai, Koichi; Nagai, Ryo; Nagai, Yoshikazu; Nagano, Kunihiro; Nagarkar, Advait; Nagasaka, Yasushi; Nagel, Martin; Nairz, Armin Michael; Nakahama, Yu; Nakamura, Koji; Nakamura, Tomoaki; Nakano, Itsuo; Namasivayam, Harisankar; Nanava, Gizo; Napier, Austin; Narayan, Rohin; Nash, Michael; Nattermann, Till; Naumann, Thomas; Navarro, Gabriela; Nayyar, Ruchika; Neal, Homer; Nechaeva, Polina; Neep, Thomas James; Negri, Andrea; Negri, Guido; Negrini, Matteo; Nektarijevic, Snezana; Nelson, Andrew; Nelson, Timothy Knight; Nemecek, Stanislav; Nemethy, Peter; Nepomuceno, Andre Asevedo; Nessi, Marzio; Neubauer, Mark; Neumann, Manuel; Neusiedl, Andrea; Neves, Ricardo; Nevski, Pavel; Newman, Paul; Nguyen, Duong Hai; Nguyen Thi Hong, Van; Nickerson, Richard; Nicolaidou, Rosy; Nielsen, Jason; Nikiforou, Nikiforos; Nikiforov, Andriy; Nikolaenko, Vladimir; Nikolaidis, Spyridon; Nikolic-Audit, Irena; Nikolics, Katalin; Nikolopoulos, Konstantinos; Nilsson, Paul; Ninomiya, Yoichi; Nisati, Aleandro; Nisius, Richard; Nobe, Takuya; Nodulman, Lawrence; Nomachi, Masaharu; Nomidis, Ioannis; Norberg, Scarlet; Nordberg, Markus; Nozaki, Mitsuaki; Nozka, Libor; Ntekas, Konstantinos; Nuncio-Quiroz, Adriana-Elizabeth; Nunes Hanninger, Guilherme; Nunnemann, Thomas; Nurse, Emily; Nuti, Francesco; O'Brien, Brendan Joseph; O'grady, Fionnbarr; O'Neil, Dugan; O'Shea, Val; Oakes, Louise Beth; Oakham, Gerald; Oberlack, Horst; Ocariz, Jose; Ochi, Atsuhiko; Ochoa, Ines; Oda, Susumu; Odaka, Shigeru; Ogren, Harold; Oh, Alexander; Oh, Seog; Ohm, Christian; Ohshima, Takayoshi; Ohshita, Hidetoshi; Okamura, Wataru; Okawa, Hideki; Okumura, Yasuyuki; Okuyama, Toyonobu; Olariu, Albert; Olivares Pino, Sebastian Andres; Oliveira Damazio, Denis; Oliver Garcia, Elena; Olivito, Dominick; Olszewski, Andrzej; Olszowska, Jolanta; Onofre, António; Onyisi, Peter; Oreglia, Mark; Oren, Yona; Orestano, Domizia; Orlando, Nicola; Oropeza Barrera, Cristina; Orr, Robert; Osculati, Bianca; Ospanov, Rustem; Otero y Garzon, Gustavo; Otono, Hidetoshi; Ouchrif, Mohamed; Ouellette, Eric; Ould-Saada, Farid; Ouraou, Ahmimed; Oussoren, Koen Pieter; Ouyang, Qun; Ovcharova, Ana; Owen, Mark; Owen, Simon; Ozcan, Veysi Erkcan; Ozturk, Nurcan; Pachal, Katherine; Pacheco Pages, Andres; Padilla Aranda, Cristobal; Pagan Griso, Simone; Paganis, Efstathios; Pahl, Christoph; Paige, Frank; Pais, Preema; Pajchel, Katarina; Palacino, Gabriel; Palestini, Sandro; Pallin, Dominique; Palma, Alberto; Palmer, Jody; Pan, Yibin; Panagiotopoulou, Evgenia; Panduro Vazquez, William; Panes, Boris; Pani, Priscilla; Panikashvili, Natalia; Panitkin, Sergey; Pantea, Dan; Papadopoulou, Theodora; Papageorgiou, Konstantinos; Paramonov, Alexander; Paredes Hernandez, Daniela; Parker, Michael Andrew; Parodi, Fabrizio; Parsons, John; Parzefall, Ulrich; Pasqualucci, Enrico; Passaggio, Stefano; Pastore, Francesca; Pásztor, Gabriella; Pataraia, Sophio; Patel, Nikhul; Pater, Joleen; Patricelli, Sergio; Pauly, Thilo; Pearce, James; Pedersen, Maiken; Pedraza Lopez, Sebastian; Pedro, Rute; Peleganchuk, Sergey; Pelikan, Daniel; Peng, Haiping; Penning, Bjoern; Penwell, John; Perepelitsa, Dennis; Perez Codina, Estel; Pérez García-Estañ, María Teresa; Perini, Laura; Pernegger, Heinz; Perrella, Sabrina; Peschke, Richard; Peshekhonov, Vladimir; Peters, Krisztian; Peters, Yvonne; Petersen, Brian; Petersen, Troels; Petit, Elisabeth; Petridis, Andreas; Petridou, Chariclia; Petrolo, Emilio; Petrucci, Fabrizio; Petteni, Michele; Pezoa, Raquel; Phillips, Peter William; Piacquadio, Giacinto; Pianori, Elisabetta; Picazio, Attilio; Piccaro, Elisa; Piccinini, Maurizio; Piec, Sebastian Marcin; Piegaia, Ricardo; Piendibene, Marco; Pignotti, David; Pilcher, James; Pilkington, Andrew; Pinamonti, Michele; Pinder, Alex; Pinfold, James; Pingel, Almut; Pinto, Belmiro; Pizio, Caterina; Pleier, Marc-Andre; Pleskot, Vojtech; Plotnikova, Elena; Plucinski, Pawel; Poddar, Sahill; Podlyski, Fabrice; Poettgen, Ruth; Poggioli, Luc; Pohl, David-leon; Pohl, Martin; Polesello, Giacomo; Policicchio, Antonio; Polifka, Richard; Polini, Alessandro; Pollard, Christopher Samuel; Polychronakos, Venetios; Pomeroy, Daniel; Pommès, Kathy; Pontecorvo, Ludovico; Pope, Bernard; Popeneciu, Gabriel Alexandru; Popovic, Dragan; Portell Bueso, Xavier; Pospisil, Stanislav; Potamianos, Karolos; Potrap, Igor; Potter, Christina; Potter, Christopher; Poveda, Joaquin; Pozdnyakov, Valery; Pozo Astigarraga, Mikel Eukeni; Prabhu, Robindra; Pralavorio, Pascal; Pranko, Aliaksandr; Prasad, Srivas; Pravahan, Rishiraj; Prell, Soeren; Price, Darren; Price, Joe; Price, Lawrence; Primavera, Margherita; Proissl, Manuel; Prokofiev, Kirill; Prokoshin, Fedor; Protopapadaki, Eftychia-sofia; Protopopescu, Serban; Proudfoot, James; Prudent, Xavier; Przybycien, Mariusz; Przysiezniak, Helenka; Psoroulas, Serena; Ptacek, Elizabeth; Pueschel, Elisa; Puldon, David; Purohit, Milind; Puzo, Patrick; Pylypchenko, Yuriy; Qian, Jianming; Qian, Weiming; Quadt, Arnulf; Quarrie, David; Quayle, William; Quilty, Donnchadha; Quinonez, Fernando; Radescu, Voica; Radhakrishnan, Sooraj Krishnan; Radloff, Peter; Ragusa, Francesco; Rahal, Ghita; Rajagopalan, Srinivasan; Rammensee, Michael; Rammes, Marcus; Randle-Conde, Aidan Sean; Rangel-Smith, Camila; Rao, Kanury; Rauscher, Felix; Rave, Stefan; Rave, Tobias Christian; Ravenscroft, Thomas; Raymond, Michel; Read, Alexander Lincoln; Rebuzzi, Daniela; Redelbach, Andreas; Redlinger, George; Reece, Ryan; Reeves, Kendall; Rehnisch, Laura; Reinsch, Andreas; Reisin, Hernan; Reiss, Andreas; Relich, Matthew; Rembser, Christoph; Renaud, Adrien; Rescigno, Marco; Resconi, Silvia; Rezanova, Olga; Reznicek, Pavel; Rezvani, Reyhaneh; Richter, Robert; Ridel, Melissa; Rieck, Patrick; Rijssenbeek, Michael; Rimoldi, Adele; Rinaldi, Lorenzo; Ritsch, Elmar; Riu, Imma; Rizatdinova, Flera; Rizvi, Eram; Robertson, Steven; Robichaud-Veronneau, Andree; Robinson, Dave; Robinson, James; Robson, Aidan; Rocha de Lima, Jose Guilherme; Roda, Chiara; Roda Dos Santos, Denis; Rodrigues, Luis; Roe, Shaun; Røhne, Ole; Romaniouk, Anatoli; Romano, Marino; Romeo, Gaston; Romero Adam, Elena; Romero Maltrana, Diego; Rompotis, Nikolaos; Roos, Lydia; Ros, Eduardo; Rosati, Stefano; Rosbach, Kilian; Rose, Anthony; Rose, Matthew; Rosendahl, Peter Lundgaard; Rosenthal, Oliver; Rossetti, Valerio; Rossi, Elvira; Rossi, Leonardo Paolo; Rosten, Rachel; Rotaru, Marina; Roth, Itamar; Rothberg, Joseph; Rousseau, David; Royon, Christophe; Rozanov, Alexandre; Rozen, Yoram; Ruan, Xifeng; Rubbo, Francesco; Rubinskiy, Igor; Rud, Viacheslav; Rudolph, Christian; Rudolph, Matthew Scott; Rühr, Frederik; Ruiz-Martinez, Aranzazu; Rurikova, Zuzana; Rusakovich, Nikolai; Ruschke, Alexander; Rutherfoord, John; Ruthmann, Nils; Ruzicka, Pavel; Ryabov, Yury; Rybar, Martin; Rybkin, Grigori; Ryder, Nick; Saavedra, Aldo; Sacerdoti, Sabrina; Saddique, Asif; Sadeh, Iftach; Sadrozinski, Hartmut; Sadykov, Renat; Safai Tehrani, Francesco; Sakamoto, Hiroshi; Sakurai, Yuki; Salamanna, Giuseppe; Salamon, Andrea; Saleem, Muhammad; Salek, David; Sales De Bruin, Pedro Henrique; Salihagic, Denis; Salnikov, Andrei; Salt, José; Salvatore, Daniela; Salvatore, Pasquale Fabrizio; Salvucci, Antonio; Salzburger, Andreas; Sampsonidis, Dimitrios; Sanchez, Arturo; Sánchez, Javier; Sanchez Martinez, Victoria; Sandaker, Heidi; Sander, Heinz Georg; Sanders, Michiel; Sandhoff, Marisa; Sandoval, Tanya; Sandoval, Carlos; Sandstroem, Rikard; Sankey, Dave; Sansoni, Andrea; Santoni, Claudio; Santonico, Rinaldo; Santos, Helena; Santoyo Castillo, Itzebelt; Sapp, Kevin; Sapronov, Andrey; Saraiva, João; Sarkisyan-Grinbaum, Edward; Sarrazin, Bjorn; Sartisohn, Georg; Sasaki, Osamu; Sasaki, Yuichi; Sauvan, Emmanuel; Sauvan, Jean-Baptiste; Savage, Graham; Savard, Pierre; Savu, Dan Octavian; Sawyer, Craig; Sawyer, Lee; Saxon, James; Sbarra, Carla; Sbrizzi, Antonio; Scanlon, Tim; Scannicchio, Diana; Scarcella, Mark; Schaarschmidt, Jana; Schacht, Peter; Schaefer, Douglas; Schaeffer, Jan; Schaelicke, Andreas; Schaepe, Steffen; Schaetzel, Sebastian; Schäfer, Uli; Schaffer, Arthur; Schaile, Dorothee; Schamberger, R~Dean; Scharf, Veit; Schegelsky, Valery; Scheirich, Daniel; Schernau, Michael; Scherzer, Max; Schettino, Vinicius; Schiavi, Carlo; Schieck, Jochen; Schillo, Christian; Schioppa, Marco; Schlenker, Stefan; Schmidt, Evelyn; Schmieden, Kristof; Schmitt, Christian; Schmitt, Christopher; Schmitt, Klaus; Schmitt, Sebastian; Schneider, Basil; Schnellbach, Yan Jie; Schnoor, Ulrike; Schoeffel, Laurent; Schoening, Andre; Schoenrock, Bradley Daniel; Schorlemmer, Andre Lukas; Schott, Matthias; Schouten, Doug; Schovancova, Jaroslava; Schramm, Steven; Schreyer, Manuel; Schroeder, Christian; Schroer, Nicolai; Schuh, Natascha; Schultens, Martin Johannes; Schultz-Coulon, Hans-Christian; Schulz, Holger; Schumacher, Markus; Schumm, Bruce; Schune, Philippe; Schwartzman, Ariel; Schwegler, Philipp; Schwemling, Philippe; Schwienhorst, Reinhard; Schwindling, Jerome; Schwindt, Thomas; Schwoerer, Maud; Sciacca, Gianfranco; Scifo, Estelle; Sciolla, Gabriella; Scott, Bill; Scuri, Fabrizio; Scutti, Federico; Searcy, Jacob; Sedov, George; Sedykh, Evgeny; Seidel, Sally; Seiden, Abraham; Seifert, Frank; Seixas, José; Sekhniaidze, Givi; Sekula, Stephen; Selbach, Karoline Elfriede; Seliverstov, Dmitry; Sellers, Graham; Semprini-Cesari, Nicola; Serfon, Cedric; Serin, Laurent; Serkin, Leonid; Serre, Thomas; Seuster, Rolf; Severini, Horst; Sforza, Federico; Sfyrla, Anna; Shabalina, Elizaveta; Shamim, Mansoora; Shan, Lianyou; Shank, James; Shao, Qi Tao; Shapiro, Marjorie; Shatalov, Pavel; Shaw, Kate; Shehu, Ciwake Yusufu; Sherwood, Peter; Shimizu, Shima; Shimmin, Chase Owen; Shimojima, Makoto; Shiyakova, Mariya; Shmeleva, Alevtina; Shochet, Mel; Shooltz, Dean; Short, Daniel; Shrestha, Suyog; Shulga, Evgeny; Shupe, Michael; Shushkevich, Stanislav; Sicho, Petr; Sicoe, Alexandru Dan; Sidiropoulou, Ourania; Sidorov, Dmitri; Sidoti, Antonio; Siegert, Frank; Sijacki, Djordje; Silbert, Ohad; Silva, José; Silva Oliveira, Marcos Vinicius; Silver, Yiftah; Silverstein, Daniel; Silverstein, Samuel; Simak, Vladislav; Simard, Olivier; Simic, Ljiljana; Simion, Stefan; Simioni, Eduard; Simmons, Brinick; Simoniello, Rosa; Simonyan, Margar; Sinervo, Pekka; Sinev, Nikolai; Sipica, Valentin; Siragusa, Giovanni; Sircar, Anirvan; Sivoklokov, Serguei; Siyad, Mohamed Jimcaale; Sjölin, Jörgen; Sjursen, Therese; Skinnari, Louise Anastasia; Skottowe, Hugh Philip; Skovpen, Kirill; Skubic, Patrick; Slater, Mark; Slavicek, Tomas; Sliwa, Krzysztof; Smakhtin, Vladimir; Smart, Ben; Smestad, Lillian; Smirnov, Sergei; Smirnov, Yury; Smirnova, Lidia; Smirnova, Oxana; Smizanska, Maria; Smolek, Karel; Snesarev, Andrei; Snidero, Giacomo; Snow, Joel; Snyder, Scott; Sobie, Randall; Socher, Felix; Soffer, Abner; Soh, Dart-yin; Solans, Carlos; Solar, Michael; Solc, Jaroslav; Soldatov, Evgeny; Soldevila, Urmila; Solfaroli Camillocci, Elena; Solodkov, Alexander; Soloviev, Igor; Solovyanov, Oleg; Solovyev, Victor; Soni, Nitesh; Sood, Alexander; Sopko, Bruno; Sopko, Vit; Sorin, Veronica; Sosebee, Mark; Sotiropoulou, Calliope Louisa; Soualah, Rachik; Soueid, Paul; Soukharev, Andrey; South, David; Spagnolo, Stefania; Spanò, Francesco; Spearman, William Robert; Spighi, Roberto; Spigo, Giancarlo; Spiwoks, Ralf; Spousta, Martin; Spreitzer, Teresa; St Denis, Richard Dante; Stabile, Alberto; Stahlman, Jonathan; Staley, Richard; Stamen, Rainer; Stanecka, Ewa; Stanek, Robert; Stanescu, Cristian; Stanescu-Bellu, Madalina; Stanitzki, Marcel Michael; Stapnes, Steinar; Starchenko, Evgeny; Stark, Jan; Staroba, Pavel; Starovoitov, Pavel; Staszewski, Rafal; Steele, Genevieve; Steinbach, Peter; Steinberg, Peter; Stelzer, Bernd; Stelzer, Harald Joerg; Stelzer-Chilton, Oliver; Stenzel, Hasko; Stern, Sebastian; Stewart, Graeme; Stillings, Jan Andre; Stockton, Mark; Stoebe, Michael; Stoerig, Kathrin; Stoicea, Gabriel; Stonjek, Stefan; Stradling, Alden; Straessner, Arno; Strandberg, Jonas; Strandberg, Sara; Strauss, Emanuel; Strauss, Michael; Strizenec, Pavol; Ströhmer, Raimund; Strom, David; Stroynowski, Ryszard; Stucci, Stefania Antonia; Stugu, Bjarne; Stupak, John; Styles, Nicholas Adam; Su, Dong; Su, Jun; Subramania, Halasya Siva; Subramaniam, Rajivalochan; Succurro, Antonella; Sugaya, Yorihito; Suhr, Chad; Suk, Michal; Sulin, Vladimir; Sultansoy, Saleh; Sumida, Toshi; Sun, Xiaohu; Sundermann, Jan Erik; Suruliz, Kerim; Sutton, Mark; Suzuki, Yu; Svatos, Michal; Swedish, Stephen; Swiatlowski, Maximilian; Sykora, Ivan; Sykora, Tomas; Ta, Duc; Tackmann, Kerstin; Taenzer, Joe; Taffard, Anyes; Tafirout, Reda; Taghavirad, Saeed; Taiblum, Nimrod; Takahashi, Yuta; Takai, Helio; Takashima, Ryuichi; Takeda, Hiroshi; Takeshita, Tohru; Takubo, Yosuke; Talby, Mossadek; Talyshev, Alexey; Tam, Jason; Tamsett, Matthew; Tan, Kong Guan; Tanaka, Junichi; Tanaka, Reisaburo; Tanaka, Shuji; Tanasijczuk, Andres Jorge; Tani, Kazutoshi; Tannoury, Nancy; Tapprogge, Stefan; Tarem, Shlomit; Tarrade, Fabien; Tartarelli, Giuseppe Francesco; Tas, Petr; Tasevsky, Marek; Tashiro, Takuya; Tassi, Enrico; Tavares Delgado, Ademar; Tayalati, Yahya; Taylor, Christopher; Taylor, Frank; Taylor, Geoffrey; Taylor, Wendy; Teischinger, Florian Alfred; Teixeira Dias Castanheira, Matilde; Teixeira-Dias, Pedro; Temming, Kim Katrin; Ten Kate, Herman; Teng, Ping-Kun; Terada, Susumu; Terashi, Koji; Terron, Juan; Terzo, Stefano; Testa, Marianna; Teuscher, Richard; Therhaag, Jan; Theveneaux-Pelzer, Timothée; Thoma, Sascha; Thomas, Juergen; Thomas-Wilsker, Joshuha; Thompson, Emily; Thompson, Paul; Thompson, Peter; Thompson, Stan; Thomsen, Lotte Ansgaard; Thomson, Evelyn; Thomson, Mark; Thong, Wai Meng; Tian, Feng; Tibbetts, Mark James; Tikhomirov, Vladimir; Tikhonov, Yury; Timoshenko, Sergey; Tiouchichine, Elodie; Tipton, Paul; Tisserant, Sylvain; Todorov, Theodore; Todorova-Nova, Sharka; Toggerson, Brokk; Tojo, Junji; Tokár, Stanislav; Tokushuku, Katsuo; Tollefson, Kirsten; Tomlinson, Lee; Tomoto, Makoto; Tompkins, Lauren; Toms, Konstantin; Torrence, Eric; Torres, Heberth; Torró Pastor, Emma; Toth, Jozsef; Touchard, Francois; Tovey, Daniel; Tran, Huong Lan; Trefzger, Thomas; Tremblet, Louis; Tricoli, Alessandro; Trigger, Isabel Marian; Trincaz-Duvoid, Sophie; Tripiana, Martin; Triplett, Nathan; Trischuk, William; Trocmé, Benjamin; Troncon, Clara; Trottier-McDonald, Michel; Trovatelli, Monica; True, Patrick; Trzebinski, Maciej; Trzupek, Adam; Tsarouchas, Charilaos; Tseng, Jeffrey; Tsiareshka, Pavel; Tsionou, Dimitra; Tsipolitis, Georgios; Tsirintanis, Nikolaos; Tsiskaridze, Shota; Tsiskaridze, Vakhtang; Tskhadadze, Edisher; Tsukerman, Ilya; Tsulaia, Vakhtang; Tsung, Jieh-Wen; Tsuno, Soshi; Tsybychev, Dmitri; Tua, Alan; Tudorache, Alexandra; Tudorache, Valentina; Tuna, Alexander Naip; Tupputi, Salvatore; Turchikhin, Semen; Turecek, Daniel; Turra, Ruggero; Tuts, Michael; Twomey, Matthew Shaun; Tykhonov, Andrii; Tylmad, Maja; Uchida, Kirika; Ueda, Ikuo; Ueno, Ryuichi; Ughetto, Michael; Ugland, Maren; Uhlenbrock, Mathias; Ukegawa, Fumihiko; Unal, Guillaume; Undrus, Alexander; Unel, Gokhan; Ungaro, Francesca; Unno, Yoshinobu; Urbaniec, Dustin; Urquijo, Phillip; Urrejola, Pedro; Usai, Giulio; Usanova, Anna; Vacavant, Laurent; Vacek, Vaclav; Vachon, Brigitte; Valencic, Nika; Valentinetti, Sara; Valero, Alberto; Valery, Loic; Valkar, Stefan; Valladolid Gallego, Eva; Vallecorsa, Sofia; Valls Ferrer, Juan Antonio; Van Der Deijl, Pieter; van der Geer, Rogier; van der Graaf, Harry; Van Der Leeuw, Robin; van der Ster, Daniel; van Eldik, Niels; van Gemmeren, Peter; Van Nieuwkoop, Jacobus; van Vulpen, Ivo; van Woerden, Marius Cornelis; Vanadia, Marco; Vandelli, Wainer; Vaniachine, Alexandre; Vankov, Peter; Vannucci, Francois; Vardanyan, Gagik; Vari, Riccardo; Varnes, Erich; Varol, Tulin; Varouchas, Dimitris; Vartapetian, Armen; Varvell, Kevin; Vassilakopoulos, Vassilios; Vazeille, Francois; Vazquez Schroeder, Tamara; Veatch, Jason; Veloso, Filipe; Veneziano, Stefano; Ventura, Andrea; Ventura, Daniel; Venturi, Manuela; Venturi, Nicola; Venturini, Alessio; Vercesi, Valerio; Verducci, Monica; Verkerke, Wouter; Vermeulen, Jos; Vest, Anja; Vetterli, Michel; Viazlo, Oleksandr; Vichou, Irene; Vickey, Trevor; Vickey Boeriu, Oana Elena; Viehhauser, Georg; Vieira De Souza, Julio; Viel, Simon; Vigne, Ralph; Villa, Mauro; Villaplana Perez, Miguel; Vilucchi, Elisabetta; Vincter, Manuella; Vinogradov, Vladimir; Virzi, Joseph; Vitells, Ofer; Vivarelli, Iacopo; Vives Vaque, Francesc; Vlachos, Sotirios; Vladoiu, Dan; Vlasak, Michal; Vogel, Adrian; Vokac, Petr; Volpi, Guido; Volpi, Matteo; von der Schmitt, Hans; von Radziewski, Holger; von Toerne, Eckhard; Vorobel, Vit; Vos, Marcel; Voss, Rudiger; Vossebeld, Joost; Vranjes, Nenad; Vranjes Milosavljevic, Marija; Vrba, Vaclav; Vreeswijk, Marcel; Vu Anh, Tuan; Vuillermet, Raphael; Vukotic, Ilija; Vykydal, Zdenek; Wagner, Peter; Wagner, Wolfgang; Wahrmund, Sebastian; Wakabayashi, Jun; Walder, James; Walker, Rodney; Walkowiak, Wolfgang; Wall, Richard; Waller, Peter; Walsh, Brian; Wang, Chao; Wang, Chiho; Wang, Haichen; Wang, Hulin; Wang, Jike; Wang, Jin; Wang, Kuhan; Wang, Rui; Wang, Song-Ming; Wang, Tan; Wang, Xiaoxiao; Warburton, Andreas; Ward, Patricia; Wardrope, David Robert; Washbrook, Andrew; Wasicki, Christoph; Watanabe, Ippei; Watkins, Peter; Watson, Alan; Watson, Ian; Watson, Miriam; Watts, Gordon; Watts, Stephen; Waugh, Ben; Webb, Samuel; Weber, Michele; Weber, Stefan Wolf; Webster, Jordan S; Weidberg, Anthony; Weigell, Philipp; Weingarten, Jens; Weiser, Christian; Weits, Hartger; Wells, Phillippa; Wenaus, Torre; Wendland, Dennis; Weng, Zhili; Wengler, Thorsten; Wenig, Siegfried; Wenzel, Volker; Wermes, Norbert; Werner, Matthias; Werner, Per; Wessels, Martin; Wetter, Jeffrey; Whalen, Kathleen; White, Andrew; White, Martin; White, Ryan; Whiteson, Daniel; Whittington, Denver; Wickens, Fred; Wiedenmann, Werner; Wielers, Monika; Wienemann, Peter; Wiglesworth, Craig; Wiik-Fuchs, Liv Antje Mari; Wijeratne, Peter Alexander; Wildauer, Andreas; Wildt, Martin Andre; Wilkens, Henric George; Will, Jonas Zacharias; Williams, Hugh; Williams, Sarah; Willocq, Stephane; Wilson, Alan; Wilson, John; Wingerter-Seez, Isabelle; Winkelmann, Stefan; Winklmeier, Frank; Wittgen, Matthias; Wittig, Tobias; Wittkowski, Josephine; Wollstadt, Simon Jakob; Wolter, Marcin Wladyslaw; Wolters, Helmut; Wosiek, Barbara; Wotschack, Jorg; Woudstra, Martin; Wozniak, Krzysztof; Wraight, Kenneth; Wright, Michael; Wu, Sau Lan; Wu, Xin; Wu, Yusheng; Wulf, Evan; Wyatt, Terry Richard; Wynne, Benjamin; Xella, Stefania; Xiao, Meng; Xu, Da; Xu, Lailin; Yabsley, Bruce; Yacoob, Sahal; Yamada, Miho; Yamaguchi, Hiroshi; Yamaguchi, Yohei; Yamamoto, Akira; Yamamoto, Kyoko; Yamamoto, Shimpei; Yamamura, Taiki; Yamanaka, Takashi; Yamauchi, Katsuya; Yamazaki, Yuji; Yan, Zhen; Yang, Haijun; Yang, Hongtao; Yang, Un-Ki; Yang, Yi; Yanush, Serguei; Yao, Liwen; Yasu, Yoshiji; Yatsenko, Elena; Yau Wong, Kaven Henry; Ye, Jingbo; Ye, Shuwei; Yen, Andy L; Yildirim, Eda; Yilmaz, Metin; Yoosoofmiya, Reza; Yorita, Kohei; Yoshida, Rikutaro; Yoshihara, Keisuke; Young, Charles; Young, Christopher John; Youssef, Saul; Yu, David Ren-Hwa; Yu, Jaehoon; Yu, Jiaming; Yu, Jie; Yuan, Li; Yurkewicz, Adam; Zabinski, Bartlomiej; Zaidan, Remi; Zaitsev, Alexander; Zaman, Aungshuman; Zambito, Stefano; Zanello, Lucia; Zanzi, Daniele; Zaytsev, Alexander; Zeitnitz, Christian; Zeman, Martin; Zemla, Andrzej; Zengel, Keith; Zenin, Oleg; Ženiš, Tibor; Zerwas, Dirk; Zevi della Porta, Giovanni; Zhang, Dongliang; Zhang, Huaqiao; Zhang, Jinlong; Zhang, Lei; Zhang, Xueyao; Zhang, Zhiqing; Zhao, Zhengguo; Zhemchugov, Alexey; Zhong, Jiahang; Zhou, Bing; Zhou, Lei; Zhou, Ning; Zhu, Cheng Guang; Zhu, Hongbo; Zhu, Junjie; Zhu, Yingchun; Zhuang, Xuai; Zibell, Andre; Zieminska, Daria; Zimine, Nikolai; Zimmermann, Christoph; Zimmermann, Robert; Zimmermann, Simone; Zimmermann, Stephanie; Zinonos, Zinonas; Ziolkowski, Michael; Zobernig, Georg; Zoccoli, Antonio; zur Nedden, Martin; Zurzolo, Giovanni; Zutshi, Vishnu; Zwalinski, Lukasz; CERN. Geneva. The LHC experiments Committee; LHCC

    2013-01-01

    The Phase-I upgrade of the ATLAS Trigger and Data Acquisition (TDAQ) system is to allow the ATLAS experiment to efficiently trigger and record data at instantaneous luminosities that are up to three times that of the original LHC design while maintaining trigger thresholds close to those used in the initial run of the LHC.

  3. Measurement of Higgs boson production via vector boson fusion in decays into W bosons with the ATLAS detector

    International Nuclear Information System (INIS)

    Bronner, Johanna

    2014-01-01

    The vector boson fusion production rate of the Standard Model Higgs boson has been measured in decays into two W bosons, each subsequently decaying into an electron or muon and a neutrino, with the ATLAS detector at the Large Hadron Collider (LHC). The vector boson fusion production cross section in the Standard Model is about an order of magnitude smaller than the dominant Higgs boson production cross section from gluon fusion. Proton-proton collision data at a center-of-mass energy of 8 TeV delivered by the LHC recorded with the ATLAS detector corresponding to an integrated luminosity of 21 fb -1 have been analyzed. Motivated by the recent discovery of a Higgs-like boson with a mass of (125.5±0.6) GeV and (125.7±0.4) GeV by the ATLAS and CMS collaborations at the LHC, the analysis is optimized for this mass. An excess of events, compatible with the Standard Model expectation for a Higgs boson with m H =125 GeV, is observed with a significance of 2.8 standard deviations when compared to the background-only expectation. The corresponding signal strength, the observed event rate relative to the Standard Model prediction of m H =125 GeV is 2.1 -0.8 +1.0 . A Higgs boson produced via vector boson fusion is excluded with 95% confidence level in the mass range between 152 GeV and 185 GeV. When combined with measurements of other Higgs boson production and decay channels by ATLAS, evidence for vector boson fusion production with a significance of 3.3 standard deviations is observed. All measurements of Higgs boson couplings to Standard Model particles are in agreement with the predictions of the Standard Model.

  4. Recent ATLAS Measurements of the Diboson Production in pp Collisions at 13 TeV

    CERN Document Server

    Kordas, Konstantinos; The ATLAS collaboration

    2018-01-01

    Recent ATLAS measurements of the WW, WZ and ZZ production in pp collisions at 13 TeV will be reported: fiducial and total cross sections, differential fiducial, and extraction of limits on anomalous Triple Gauge Couplings. A brief mention on VBS processes will be made as well.

  5. The ATLAS Data Acquisition and High Level Trigger system

    International Nuclear Information System (INIS)

    2016-01-01

    This paper describes the data acquisition and high level trigger system of the ATLAS experiment at the Large Hadron Collider at CERN, as deployed during Run 1. Data flow as well as control, configuration and monitoring aspects are addressed. An overview of the functionality of the system and of its performance is presented and design choices are discussed.

  6. Heavy ion results from ATLAS

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00241915; The ATLAS collaboration

    2017-01-01

    These proceedings provide an overview of the new results obtained with the ATLAS Detector at the LHC, which are presented in the Quark Matter 2017 conference. These results are covered in twelve parallel talks, one flash talk and eleven posters, and they are grouped into five areas: initial state, jet quenching, quarkonium production, longitudinal flow dynamics, and collectivity in small systems.

  7. Study of the Z Boson Production at the ATLAS Experiment with First Data

    CERN Document Server

    Schott, Matthias

    The ATLAS detector, currently in its final installation phase at CERN, is designed to provide precise measurements of 14 TeV proton-proton collisions at the Large Hadron Collider. The measurements of the cross section and transverse momentum spectrum of the Z boson production at LHC provides first tests of the standard model in a new energy domain and may reveal exotic physics processes. Moreover, the properties of the Z boson resonance and its decay into two muons are known to very high precision from LEP experiments and hence can be used as a physics process for calibration and alignment. The Z boson production is also a common background process for many other physics analyses and must therefore be well understood. This thesis describes a measurement strategy of the cross section s for the process pp->Z->(mu)(mu) at the ATLAS experiment during its startup phase. As a result of this study a precision of d(sigma)/ (sigma) = 0.006(stat) + 0.008(sys) + 0.01 (pdf) is expected for an integrated luminosity of 50 ...

  8. Derived Physics Data Production in ATLAS: Experience with Run 1 and Looking Ahead (proceedings)

    CERN Document Server

    Laycock, P; The ATLAS collaboration; Beckingham, M; Henderson, R; Zhou, L

    2014-01-01

    While a significant fraction of ATLAS physicists directly analyse the AOD (Analysis Object Data) produced at the CERN Tier 0, a much larger fraction have opted to analyse data in a flat ROOT format. The large scale production of this Derived Physics Data (DPD) format must cater for both detailed performance studies of the ATLAS detector and object reconstruction, as well as higher level and generally lighter-content physics analysis. The delay between data-taking and DPD production allows for software improvements, while the ease of arbitrarily defined skimming/slimming of this format results in an optimally performant format for end-user analysis. Given the diversity of requirements, there are many flavours of DPDs, which can result in large peak computing resource demands. While the current model has proven to be very flexible for the individual groups and has successfully met the needs of the collaboration, the resource requirements at the end of Run 1 are much larger than planned. In the near future, ATLA...

  9. Mesure de la production de di-bosons WZ auprès du LHC avec l'experience ATLAS

    CERN Document Server

    Keoshkerian, Houry

    This thesis is performed in the frame of the ATLAS experiment at the LHC. A first part of the work presented in this document consists on the time calibration of the ATLAS Liquid Argon (LAr) calorimeter. The control of the time alignment of the calorimeter is important for the goodness of the quality of the energy reconstructed in the calorimeter. The results presented in this thesis have allowed an improvement of 30% of the global time resolution of the LAr calorimeter. The Standard model of particle physics predicts, during proton collisions, the production of the $W$ and $Z$ weak bosons as a pair due to the interaction of a quark with an anti-quark. The diboson production can be sensitive to the couplings between vector bosons. An anomalous deviation of these couplings from the prediction of the SM would point to the presence of new physics. The use of the full statistics of the 2012 ATLAS data allowed us to increase the precision of the measurement of these couplings compared to previous results based on...

  10. Integration Of PanDA Workload Management System With Supercomputers for ATLAS and Data Intensive Science

    Science.gov (United States)

    Klimentov, A.; De, K.; Jha, S.; Maeno, T.; Nilsson, P.; Oleynik, D.; Panitkin, S.; Wells, J.; Wenaus, T.

    2016-10-01

    The.LHC, operating at CERN, is leading Big Data driven scientific explorations. Experiments at the LHC explore the fundamental nature of matter and the basic forces that shape our universe. ATLAS, one of the largest collaborations ever assembled in the sciences, is at the forefront of research at the LHC. To address an unprecedented multi-petabyte data processing challenge, the ATLAS experiment is relying on a heterogeneous distributed computational infrastructure. The ATLAS experiment uses PanDA (Production and Data Analysis) Workload Management System for managing the workflow for all data processing on over 150 data centers. Through PanDA, ATLAS physicists see a single computing facility that enables rapid scientific breakthroughs for the experiment, even though the data centers are physically scattered all over the world. While PanDA currently uses more than 250,000 cores with a peak performance of 0.3 petaFLOPS, LHC data taking runs require more resources than grid can possibly provide. To alleviate these challenges, LHC experiments are engaged in an ambitious program to expand the current computing model to include additional resources such as the opportunistic use of supercomputers. We will describe a project aimed at integration of PanDA WMS with supercomputers in United States, in particular with Titan supercomputer at Oak Ridge Leadership Computing Facility. Current approach utilizes modified PanDA pilot framework for job submission to the supercomputers batch queues and local data management, with light-weight MPI wrappers to run single threaded workloads in parallel on LCFs multi-core worker nodes. This implementation was tested with a variety of Monte-Carlo workloads on several supercomputing platforms for ALICE and ATLAS experiments and it is in full pro duction for the ATLAS since September 2015. We will present our current accomplishments with running PanDA at supercomputers and demonstrate our ability to use PanDA as a portal independent of the

  11. Integration Of PanDA Workload Management System With Supercomputers for ATLAS and Data Intensive Science

    International Nuclear Information System (INIS)

    Klimentov, A; Maeno, T; Nilsson, P; Panitkin, S; Wenaus, T; De, K; Oleynik, D; Jha, S; Wells, J

    2016-01-01

    The.LHC, operating at CERN, is leading Big Data driven scientific explorations. Experiments at the LHC explore the fundamental nature of matter and the basic forces that shape our universe. ATLAS, one of the largest collaborations ever assembled in the sciences, is at the forefront of research at the LHC. To address an unprecedented multi-petabyte data processing challenge, the ATLAS experiment is relying on a heterogeneous distributed computational infrastructure. The ATLAS experiment uses PanDA (Production and Data Analysis) Workload Management System for managing the workflow for all data processing on over 150 data centers. Through PanDA, ATLAS physicists see a single computing facility that enables rapid scientific breakthroughs for the experiment, even though the data centers are physically scattered all over the world. While PanDA currently uses more than 250,000 cores with a peak performance of 0.3 petaFLOPS, LHC data taking runs require more resources than grid can possibly provide. To alleviate these challenges, LHC experiments are engaged in an ambitious program to expand the current computing model to include additional resources such as the opportunistic use of supercomputers. We will describe a project aimed at integration of PanDA WMS with supercomputers in United States, in particular with Titan supercomputer at Oak Ridge Leadership Computing Facility. Current approach utilizes modified PanDA pilot framework for job submission to the supercomputers batch queues and local data management, with light-weight MPI wrappers to run single threaded workloads in parallel on LCFs multi-core worker nodes. This implementation was tested with a variety of Monte-Carlo workloads on several supercomputing platforms for ALICE and ATLAS experiments and it is in full pro duction for the ATLAS since September 2015. We will present our current accomplishments with running PanDA at supercomputers and demonstrate our ability to use PanDA as a portal independent of the

  12. The ATLAS Trigger System: Ready for Run-2

    CERN Document Server

    Nakahama, Yu; The ATLAS collaboration

    2015-01-01

    The ATLAS trigger has been used very successfully for the online event selection during the first run of the LHC between 2009-2013 at a centre-of-mass energy between 900 GeV and 8 TeV. The trigger system consists of a hardware Level-1 (L1) and a software based high-level trigger (HLT) that reduces the event rate from the design bunch-crossing rate of 40 MHz to an average recording rate of a few hundred Hz. During the next data-taking period starting in early 2015 (Run-2) the LHC will operate at a centre-of-mass energy of about 13 TeV resulting in roughly five times higher trigger rates. We will review the upgrades to the ATLAS Trigger system that have been implemented during the shutdown and that will allow us to cope with these increased trigger rates while maintaining or even improving our efficiency to select relevant physics processes. This includes changes to the L1 calorimeter trigger, the introduction of a new L1 topological trigger module, improvements in the L1 muon system and the merging of the prev...

  13. Software framework developed for the slice test of the ATLAS endcap muon trigger system

    CERN Document Server

    Komatsu, S; Ishida, Y; Tanaka, K; Hasuko, K; Kano, H; Matsumoto, Y; Yakamura, Y; Sakamoto, H; Ikeno, M; Nakayoshi, K; Sasaki, O; Yasu, Y; Hasegawa, Y; Totsuka, M; Tsuji, S; Maeno, T; Ichimiya, R; Kurashige, H

    2002-01-01

    A sliced system test of the ATLAS end cap muon level 1 trigger system has been done in 2001 and 2002 separately. We have developed an own software framework for property and run controls for the slice test in 2001. The system is described in C++ throughout. The multi-PC control system is accomplished using the CORBA system. We have then restructured the software system on top of the ATLAS online software framework, and used this one for the slice test in 2002. In this report we discuss two systems in detail with emphasizing the module property configuration and run control. (8 refs).

  14. Cross-compilation of ATLAS online software to the power PC-Vx works system

    International Nuclear Information System (INIS)

    Tian Yuren; Li Jin; Ren Zhengyu; Zhu Kejun

    2005-01-01

    BES III, selected ATLAS online software as a framework of its run-control system. BES III applied Power PC-VxWorks system on its front-end readout system, so it is necessary to cross-compile this software to PowerPC-VxWorks system. The article demonstrates several aspects related to this project, such as the structure and organization of the ATLAS online software, the application of CMT tool while cross-compiling, the selection and configuration of the cross-compiler, methods to solve various problems due to the difference of compiler and operating system etc. The software, after cross-compiling, can normally run, and makes up a complete run-control system with the software running on Linux system. (authors)

  15. ATLAS experience with HEP software at the Argonne leadership computing facility

    International Nuclear Information System (INIS)

    Uram, Thomas D; LeCompte, Thomas J; Benjamin, D

    2014-01-01

    A number of HEP software packages used by the ATLAS experiment, including GEANT4, ROOT and ALPGEN, have been adapted to run on the IBM Blue Gene supercomputers at the Argonne Leadership Computing Facility. These computers use a non-x86 architecture and have a considerably less rich operating environment than in common use in HEP, but also represent a computing capacity an order of magnitude beyond what ATLAS is presently using via the LCG. The status and potential for making use of leadership-class computing, including the status of integration with the ATLAS production system, is discussed.

  16. ATLAS Experience with HEP Software at the Argonne Leadership Computing Facility

    CERN Document Server

    LeCompte, T; The ATLAS collaboration; Benjamin, D

    2014-01-01

    A number of HEP software packages used by the ATLAS experiment, including GEANT4, ROOT and ALPGEN, have been adapted to run on the IBM Blue Gene supercomputers at the Argonne Leadership Computing Facility. These computers use a non-x86 architecture and have a considerably less rich operating environment than in common use in HEP, but also represent a computing capacity an order of magnitude beyond what ATLAS is presently using via the LCG. The status and potential for making use of leadership-class computing, including the status of integration with the ATLAS production system, is discussed.

  17. Study of the Z boson production at the ATLAS experiment with first data

    International Nuclear Information System (INIS)

    Schott, M.

    2007-06-01

    The ATLAS detector, currently in its final installation phase at CERN, is designed to provide precise measurements of 14 TeV proton-proton collisions at the Large Hadron Collider. The measurements of the cross section and transverse momentum spectrum of the Z boson production at LHC provides first tests of the standard model in a new energy domain and may reveal exotic physics processes. Moreover, the properties of the Z boson resonance and its decay into two muons are known to very high precision from LEP experiments and hence can be used as a physics process for calibration and alignment. The Z boson production is also a common background process for many other physics analyses and must therefore be well understood. This thesis describes a measurement strategy of the cross section σ for the process pp→γ/Z→μ + μ - at the ATLAS experiment during its startup phase. As a result of this study a precision of (Δσ)/(σ)∼0.006(stat)±0.008(sys) +0.016 -0.008 (pdf) is expected for an integrated luminosity of 50 pb -1 , assuming a fully operational ATLAS detector, not including uncertainties in the luminosity measurements. A major goal of the approach presented was to minimize the dependence on Monte Carlo simulations. Hence, several methods for the determination of the detector response based on data have been studied. In addition, a strategy for the differential cross section measurement (dσ)/(dp Z T ) of the transverse momentum of the Z boson has been developed. In contrast to a measurement of the total cross section, it is expected that the statistical uncertainty dominates for the given integrated luminosity of 50 pb -1 . The predicted high p T resolution of the ATLAS Inner Detector and the Muon Spectrometer allow for the first observation of interesting parton distribution effects, i.e. the so-called x-broadening, even with the limited statistics expected during the first data taking period.

  18. Study of the Z boson production at the ATLAS experiment with first data

    Energy Technology Data Exchange (ETDEWEB)

    Schott, M

    2007-06-15

    The ATLAS detector, currently in its final installation phase at CERN, is designed to provide precise measurements of 14 TeV proton-proton collisions at the Large Hadron Collider. The measurements of the cross section and transverse momentum spectrum of the Z boson production at LHC provides first tests of the standard model in a new energy domain and may reveal exotic physics processes. Moreover, the properties of the Z boson resonance and its decay into two muons are known to very high precision from LEP experiments and hence can be used as a physics process for calibration and alignment. The Z boson production is also a common background process for many other physics analyses and must therefore be well understood. This thesis describes a measurement strategy of the cross section {sigma} for the process pp{yields}{gamma}/Z{yields}{mu}{sup +}{mu}{sup -} at the ATLAS experiment during its startup phase. As a result of this study a precision of ({delta}{sigma})/({sigma}){approx}0.006(stat){+-}0.008(sys){sup +0.016}{sub -0.008}(pdf) is expected for an integrated luminosity of 50 pb{sup -1}, assuming a fully operational ATLAS detector, not including uncertainties in the luminosity measurements. A major goal of the approach presented was to minimize the dependence on Monte Carlo simulations. Hence, several methods for the determination of the detector response based on data have been studied. In addition, a strategy for the differential cross section measurement (d{sigma})/(dp{sup Z}{sub T}) of the transverse momentum of the Z boson has been developed. In contrast to a measurement of the total cross section, it is expected that the statistical uncertainty dominates for the given integrated luminosity of 50 pb{sup -1}. The predicted high p{sub T} resolution of the ATLAS Inner Detector and the Muon Spectrometer allow for the first observation of interesting parton distribution effects, i.e. the so-called x-broadening, even with the limited statistics expected during the

  19. Thermo-dynamical measurements for ATLAS Inner Detector (evaporative cooling system)

    CERN Document Server

    Bitadze, Alexander; Buttar, Craig

    During the construction, installation and initial operation of the Evaporative Cooling System for the ATLAS Inner Detector SCT Barrel Sub-detector, some performance characteristics were observed to be inconsistent with the original design specifications, therefore the assumptions made in the ATLAS Inner Detector TDR were revisited. The main concern arose because of unexpected pressure drops in the piping system from the end of the detector structure to the distribution racks. The author of this theses made a series of measurements of these pressure drops and the thermal behavior of SCT-Barrel cooling Stave. Tests were performed on the installed detector in the pit, and using a specially assembled full scale replica in the SR1 laboratory at CERN. This test setup has been used to perform extensive tests of the cooling performance of the system including measurements of pressure drops in different parts of system, studies of the thermal profile along the stave pipe for different running conditions / parameters a...

  20. Direct isolated single and di-photon production at ATLAS and CMS

    International Nuclear Information System (INIS)

    Kolberg, T.R.

    2014-01-01

    Direct isolated single and di-photon production cross section measurements are a classic test of perturbative QCD and are used to constrain the gluon densities in the proton. The LHC general-purpose experiments ATLAS and CMS have made a number of differential cross section measurements for both the single and di-photon production processes using the 2011 dataset at 7 TeV center-of-mass energy. Overall, good agreement is found with the theoretical predictions, and the measurements are sufficiently precise to constrain the gluon PDF uncertainty for other production processes at the LHC. Some systematic differences between the di-photon data and the NLO (Next-to-Leading Order) theoretical predictions show a need to include higher-order effects in the predictions. (author)

  1. ATLAS Facility and Instrumentation Description Report

    International Nuclear Information System (INIS)

    Kang, Kyoung Ho; Moon, Sang Ki; Park, Hyun Sik

    2009-06-01

    A thermal-hydraulic integral effect test facility, ATLAS (Advanced Thermal-hydraulic Test Loop for Accident Simulation), has been constructed at KAERI (Korea Atomic Energy Research Institute). The ATLAS is a half-height and 1/288-volume scaled test facility with respect to the APR1400. The fluid system of the ATLAS consists of a primary system, a secondary system, a safety injection system, a break simulating system, a containment simulating system, and auxiliary systems. The primary system includes a reactor vessel, two hot legs, four cold legs, a pressurizer, four reactor coolant pumps, and two steam generators. The secondary system of the ATLAS is simplified to be of a circulating looptype. Most of the safety injection features of the APR1400 and the OPR1000 are incorporated into the safety injection system of the ATLAS. In the ATLAS test facility, about 1300 instrumentations are installed to precisely investigate the thermal-hydraulic behavior in simulation of the various test scenarios. This report describes the scaling methodology, the geometric data of the individual component, and the specification and the location of the instrumentations which are specific to the simulation of 50% DVI line break accident of the APR1400 for supporting the 50 th OECD/NEA International Standard Problem Exercise (ISP-50)

  2. Data analytics in the ATLAS Distributed Computing

    CERN Document Server

    Vukotic, Ilija; The ATLAS collaboration; Bryant, Lincoln

    2015-01-01

    The ATLAS Data analytics effort is focused on creating systems which provide the ATLAS ADC with new capabilities for understanding distributed systems and overall operational performance. These capabilities include: warehousing information from multiple systems (the production and distributed analysis system - PanDA, the distributed data management system - Rucio, the file transfer system, various monitoring services etc. ); providing a platform to execute arbitrary data mining and machine learning algorithms over aggregated data; satisfy a variety of use cases for different user roles; host new third party analytics services on a scalable compute platform. We describe the implemented system where: data sources are existing RDBMS (Oracle) and Flume collectors; a Hadoop cluster is used to store the data; native Hadoop and Apache Pig scripts are used for data aggregation; and R for in-depth analytics. Part of the data is indexed in ElasticSearch so both simpler investigations and complex dashboards can be made ...

  3. Data federation strategies for ATLAS using XRootD

    Science.gov (United States)

    Gardner, Robert; Campana, Simone; Duckeck, Guenter; Elmsheuser, Johannes; Hanushevsky, Andrew; Hönig, Friedrich G.; Iven, Jan; Legger, Federica; Vukotic, Ilija; Yang, Wei; Atlas Collaboration

    2014-06-01

    In the past year the ATLAS Collaboration accelerated its program to federate data storage resources using an architecture based on XRootD with its attendant redirection and storage integration services. The main goal of the federation is an improvement in the data access experience for the end user while allowing more efficient and intelligent use of computing resources. Along with these advances come integration with existing ATLAS production services (PanDA and its pilot services) and data management services (DQ2, and in the next generation, Rucio). Functional testing of the federation has been integrated into the standard ATLAS and WLCG monitoring frameworks and a dedicated set of tools provides high granularity information on its current and historical usage. We use a federation topology designed to search from the site's local storage outward to its region and to globally distributed storage resources. We describe programmatic testing of various federation access modes including direct access over the wide area network and staging of remote data files to local disk. To support job-brokering decisions, a time-dependent cost-of-data-access matrix is made taking into account network performance and key site performance factors. The system's response to production-scale physics analysis workloads, either from individual end-users or ATLAS analysis services, is discussed.

  4. Recent ATLAS Articles on WLAP

    CERN Multimedia

    Goldfarb, S

    2005-01-01

    As reported in the September 2004 ATLAS eNews, the Web Lecture Archive Project is a system for the archiving and publishing of multimedia presentations, using the Web as medium. We list here newly available WLAP items relating to ATLAS: Atlas Software Week Plenary 6-10 December 2004 North American ATLAS Physics Workshop (Tucson) 20-21 December 2004 (17 talks) Physics Analysis Tools Tutorial (Tucson) 19 December 2004 Full Chain Tutorial 21 September 2004 ATLAS Plenary Sessions, 17-18 February 2005 (17 talks) Coming soon: ATLAS Tutorial on Electroweak Physics, 14 Feb. 2005 Software Workshop, 21-22 February 2005 Click here to browse WLAP for all ATLAS lectures.

  5. The monitoring and calibration Web system of the ATLAS hadronic calorimeter

    Energy Technology Data Exchange (ETDEWEB)

    Maidantchik, Carmen; Gomes, Andressa Andrea Sivollela; Marroquim, Fernando [Universidade Federal do Rio de Janeiro (UFRJ), RJ (Brazil)

    2011-07-01

    Full text: The scintillator tiles hadronic calorimeter (TileCal) of the ATLAS detector measures the energy of resultant particles in a collision. The calorimetry system was designed to absorb the energy of the particles that crosses the detector and is composed by three barrels, each one equally divided into 64 modules. The ionizing particles that cross the tiles induce the production of light, which intensity is proportional to the energy deposited by the fragment. The produced light propagates through the tiles towards the edges, where it is absorbed and displaced until reaching the photomultiplier tubes (PMTs), also known as electronic reading channels. Each module combines till 45 PMTs. For each run, the reconstruction process starts with a data analysis that can comprises different levels of information granularity till arriving to the PMTs level. Following this phase, the Data Quality Monitoring Framework (DQMF) system automatically generates quality indicators associated to the channels. Depending on the configuration that is registered in the DQMF, the channel status can be automatically defined as good, affected or bad. The status of each module is defined by the percentage of existing good, affected or bad channels. At this point, the analysis of modules allows the identification of the ones that are problematic by the examination of graphics that are automatically generated during the data reconstruction stage. Then, an analysis of a module performance by a time period that encompasses different types of runs is performed. In this last step, the list of problematic channels can be modified through the insertion or exclusion of PTMs, as in the case when a channel is substituted. Additionally, during the whole calorimeter operation, it is fundamental to identify the electronic channels that are active, dead (nor working), noisy and the ones that presents saturation in the signal digitalisation process. The Monitoring and Calibration Web System (MCWS) was

  6. The monitoring and calibration Web system of the ATLAS hadronic calorimeter

    International Nuclear Information System (INIS)

    Maidantchik, Carmen; Gomes, Andressa Andrea Sivollela; Marroquim, Fernando

    2011-01-01

    Full text: The scintillator tiles hadronic calorimeter (TileCal) of the ATLAS detector measures the energy of resultant particles in a collision. The calorimetry system was designed to absorb the energy of the particles that crosses the detector and is composed by three barrels, each one equally divided into 64 modules. The ionizing particles that cross the tiles induce the production of light, which intensity is proportional to the energy deposited by the fragment. The produced light propagates through the tiles towards the edges, where it is absorbed and displaced until reaching the photomultiplier tubes (PMTs), also known as electronic reading channels. Each module combines till 45 PMTs. For each run, the reconstruction process starts with a data analysis that can comprises different levels of information granularity till arriving to the PMTs level. Following this phase, the Data Quality Monitoring Framework (DQMF) system automatically generates quality indicators associated to the channels. Depending on the configuration that is registered in the DQMF, the channel status can be automatically defined as good, affected or bad. The status of each module is defined by the percentage of existing good, affected or bad channels. At this point, the analysis of modules allows the identification of the ones that are problematic by the examination of graphics that are automatically generated during the data reconstruction stage. Then, an analysis of a module performance by a time period that encompasses different types of runs is performed. In this last step, the list of problematic channels can be modified through the insertion or exclusion of PTMs, as in the case when a channel is substituted. Additionally, during the whole calorimeter operation, it is fundamental to identify the electronic channels that are active, dead (nor working), noisy and the ones that presents saturation in the signal digitalisation process. The Monitoring and Calibration Web System (MCWS) was

  7. Development of a picosecond time-of-flight system in the ATLAS experiment

    International Nuclear Information System (INIS)

    Grabas, Herve

    2013-01-01

    In this thesis, we present a study of the sensitivity to Beyond Standard Model physics brought by the design and installation of picosecond time-of-flight detectors in the forward region of the ATLAS experiment at the LHC. The first part of the thesis present a study of the sensitivity to the quartic gauge anomalous coupling between the photon and the W boson, using exclusive WW pair production in ATLAS. The event selection is built considering the semi-leptonic decay of WW pair and the presence of the AFP detector in ATLAS. The second part gives a description of large area picosecond photo-detectors design and time reconstruction algorithms with a special care given to signal sampling and processing for precision timing. The third part presents the design of SamPic: a custom picosecond readout integrated circuit. At the end, its first results are reported, and in particular a world-class 5 ps timing precision in measuring the delay between two fast pulses. (author) [fr

  8. Supporting ATLAS

    CERN Multimedia

    2003-01-01

    Eighteen feet made of stainless steel will support the barrel ATLAS detector in the cavern at Point 1. In total, the ATLAS feet system will carry approximately 6000 tons, and will give the same inclination to the detector as the LHC accelerator. The installation of the feet is scheduled to finish during January 2004 with an installation precision at the 1 mm level despite their height of 5.3 metres. The manufacture was carried out in Russia (Company Izhorskiye Zavody in St. Petersburg), as part of a Russian and JINR Dubna in-kind contribution to ATLAS. Involved in the installation is a team from IHEP-Protvino (Russia), the ATLAS technical co-ordination team at CERN, and the CERN survey team. In all, about 15 people are involved. After the feet are in place, the barrel toroid magnet and the barrel calorimeters will be installed. This will keep the ATLAS team busy for the entire year 2004.

  9. ATLAS, an integrated structural analysis and design system. Volume 4: Random access file catalog

    Science.gov (United States)

    Gray, F. P., Jr. (Editor)

    1979-01-01

    A complete catalog is presented for the random access files used by the ATLAS integrated structural analysis and design system. ATLAS consists of several technical computation modules which output data matrices to corresponding random access file. A description of the matrices written on these files is contained herein.

  10. The ATLAS multi-user upgrade and potential applications

    Energy Technology Data Exchange (ETDEWEB)

    Mustapha, B.; Nolen, J. A.; Savard, G.; Ostroumov, P. N.

    2017-12-01

    With the recent integration of the CARIBU-EBIS charge breeder into the ATLAS accelerator system to provide for more pure and efficient charge breeding of radioactive beams, a multi-user upgrade of the ATLAS facility is being proposed to serve multiple users simultaneously. ATLAS was the first superconducting ion linac in the world and is the US DOE low-energy Nuclear Physics National User Facility. The proposed upgrade will take advantage of the continuous-wave nature of ATLAS and the pulsed nature of the EBIS charge breeder in order to simultaneously accelerate two beams with very close mass-to-charge ratios; one stable from the existing ECR ion source and one radioactive from the newly commissioned EBIS charge breeder. In addition to enhancing the nuclear physics program, beam extraction at different points along the linac will open up the opportunity for other potential applications; for instance, material irradiation studies at ~ 1 MeV/u and isotope production at ~ 6 MeV/u or at the full ATLAS energy of ~ 15 MeV/u. The concept and proposed implementation of the ATLAS multi-user upgrade will be presented. Future plans to enhance the flexibility of this upgrade will also be presented.

  11. Development and Test of the Cooling System for the ATLAS Hadron Tile Calorimeter

    CERN Document Server

    Schlager, Gerolf

    2002-01-01

    The ATLAS detector is a general-purpose experiment for proton-proton collisions designed to investigate the full range of physical processes at the Large Hadron Collider (LHC). The ATLAS Tile Hadron Calorimeter is designed to measure the energies of jets with a resolution of E/E = 50%/pE 3%, for j j<3. This thesis presents the detailed studies which were carried out with prototypes of the Tilecal cooling system during my year as technical student at CERN. The results will be used to validate and to determine the nal design of the cooling system of the ATLAS Tile calorimeter. The performance of the cooling unit built for the calibration of Tilecal modules was evaluated for various parameters like temperature stability and safety conditions during operation. Additionally I contributed to the analysis of the calorimeter response for di erent cooling temperatures. These results determined the constraints on the operation conditions of the cooling system in terms of temperature stability that will be needed d...

  12. Report to users of ATLAS [Argonne Tandem-Line Accelerator System

    International Nuclear Information System (INIS)

    Ahmad, I.; Glagola, B.

    1987-03-01

    The operation and development of ATLAS are reported, including accelerator improvements. Particularly noted is an upgrade to extend the mass range of projectiles up to uranium and to increase the beam intensity by at least two orders of magnitude for all ions. Meetings are discussed, particularly of the Program Advisory Committee and the User Group Executive Committee. Some basic information is provided for users planning to run experiments at ATLAS, including a table of beams available. The data acquisition system for ATLAS, DAPHNE, is discussed, as are the following experimental facilities: the Argonne-Notre Dame Gamma Ray Facility, a proposal submitted for constructing a large-acceptance Fragment Mass Analyzer. Brief summaries are provided of some recent experiments for which data analysis is complete. Experiments performed during the period from June 1, 1986 to January 31, 1987 are tabulated, providing the experiment number, scientists, institution, experiment name, number of days, beam, and energy

  13. Report to users of ATLAS (Argonne Tandem-Line Accelerator System)

    Energy Technology Data Exchange (ETDEWEB)

    Ahmad, I.; Glagola, B. (eds.)

    1987-03-01

    The operation and development of ATLAS are reported, including accelerator improvements. Particularly noted is an upgrade to extend the mass range of projectiles up to uranium and to increase the beam intensity by at least two orders of magnitude for all ions. Meetings are discussed, particularly of the Program Advisory Committee and the User Group Executive Committee. Some basic information is provided for users planning to run experiments at ATLAS, including a table of beams available. The data acquisition system for ATLAS, DAPHNE, is discussed, as are the following experimental facilities: the Argonne-Notre Dame Gamma Ray Facility, a proposal submitted for constructing a large-acceptance Fragment Mass Analyzer. Brief summaries are provided of some recent experiments for which data analysis is complete. Experiments performed during the period from June 1, 1986 to January 31, 1987 are tabulated, providing the experiment number, scientists, institution, experiment name, number of days, beam, and energy. (LEW)

  14. First Results from the Online Radiation Dose Monitoring System in ATLAS experiment

    CERN Document Server

    Mandić, I; The ATLAS collaboration; Deliyergiyev, M; Gorišek, A; Kramberger, G; Mikuž, M; Franz, S; Hartert, J; Dawson, I; Miyagawa, P; Nicolas, L

    2011-01-01

    High radiation doses which will accumulate in components of ATLAS experiment during data taking will causes damage to detectors and readout electronics. It is therefore important to continuously monitor the doses to estimate the level of degradation caused by radiation. Online radiation monitoring system measures ionizing dose in SiO2 , displacement damage in silicon in terms of 1-MeV(Si) equivalent neutron fluence and fluence of thermal neutrons at several locations in ATLAS detector. In this paper design of the system, results of measurements and comparison of measured integrated doses and fluences with predictions from FLUKA simulation will be shown.

  15. Improving the ATLAS physics potential with the Fast Track Trigger System

    CERN Document Server

    Cavaliere, Viviana; The ATLAS collaboration

    2015-01-01

    The ATLAS Fast TracKer (FTK) is a custom electronics system that will operate at the full Level-1 accept rate, 100 kHz, to provide high quality tracks as input to the High-Level Trigger. The event reconstruction is performed in hardware, thanks to the massive parallelism of associative memories (AM) and FPGAs. We present the advantages for the physics goals of the ATLAS experiment and the recent results on the design, technological advancements and testing of some of the core components used in the processor.

  16. Leveraging the checkpoint-restart technique for optimizing CPU efficiency of ATLAS production applications on opportunistic platforms

    CERN Document Server

    Cameron, David; The ATLAS collaboration

    2017-01-01

    Data processing applications of the ATLAS experiment, such as event simulation and reconstruction, spend considerable amount of time in the initialization phase. This phase includes loading a large number of shared libraries, reading detector geometry and condition data from external databases, building a transient representation of the detector geometry and initializing various algorithms and services. In some cases the initialization step can take as long as 10-15 minutes. Such slow initialization, being inherently serial, has a significant negative impact on overall CPU efficiency of the production job, especially when the job is executed on opportunistic, often short-lived, resources such as commercial clouds or volunteer computing. In order to improve this situation, we can take advantage of the fact that ATLAS runs large numbers of production jobs with similar configuration parameters (e.g. jobs within the same production task). This allows us to checkpoint one job at the end of its configuration step a...

  17. ATLAS DataFlow Infrastructure: Recent results from ATLAS cosmic and first-beam data-taking

    Energy Technology Data Exchange (ETDEWEB)

    Vandelli, Wainer, E-mail: wainer.vandelli@cern.c

    2010-04-01

    The ATLAS DataFlow infrastructure is responsible for the collection and conveyance of event data from the detector front-end electronics to the mass storage. Several optimized and multi-threaded applications fulfill this purpose operating over a multi-stage Gigabit Ethernet network which is the backbone of the ATLAS Trigger and Data Acquisition System. The system must be able to efficiently transport event-data with high reliability, while providing aggregated bandwidths larger than 5 GByte/s and coping with many thousands network connections. Nevertheless, routing and streaming capabilities and monitoring and data accounting functionalities are also fundamental requirements. During 2008, a few months of ATLAS cosmic data-taking and the first experience with the LHC beams provided an unprecedented test-bed for the evaluation of the performance of the ATLAS DataFlow, in terms of functionality, robustness and stability. Besides, operating the system far from its design specifications helped in exercising its flexibility and contributed in understanding its limitations. Moreover, the integration with the detector and the interfacing with the off-line data processing and management have been able to take advantage of this extended data taking-period as well. In this paper we report on the usage of the DataFlow infrastructure during the ATLAS data-taking. These results, backed-up by complementary performance tests, validate the architecture of the ATLAS DataFlow and prove that the system is robust, flexible and scalable enough to cope with the final requirements of the ATLAS experiment.

  18. ATLAS DataFlow Infrastructure: Recent results from ATLAS cosmic and first-beam data-taking

    International Nuclear Information System (INIS)

    Vandelli, Wainer

    2010-01-01

    The ATLAS DataFlow infrastructure is responsible for the collection and conveyance of event data from the detector front-end electronics to the mass storage. Several optimized and multi-threaded applications fulfill this purpose operating over a multi-stage Gigabit Ethernet network which is the backbone of the ATLAS Trigger and Data Acquisition System. The system must be able to efficiently transport event-data with high reliability, while providing aggregated bandwidths larger than 5 GByte/s and coping with many thousands network connections. Nevertheless, routing and streaming capabilities and monitoring and data accounting functionalities are also fundamental requirements. During 2008, a few months of ATLAS cosmic data-taking and the first experience with the LHC beams provided an unprecedented test-bed for the evaluation of the performance of the ATLAS DataFlow, in terms of functionality, robustness and stability. Besides, operating the system far from its design specifications helped in exercising its flexibility and contributed in understanding its limitations. Moreover, the integration with the detector and the interfacing with the off-line data processing and management have been able to take advantage of this extended data taking-period as well. In this paper we report on the usage of the DataFlow infrastructure during the ATLAS data-taking. These results, backed-up by complementary performance tests, validate the architecture of the ATLAS DataFlow and prove that the system is robust, flexible and scalable enough to cope with the final requirements of the ATLAS experiment.

  19. Thermal Performance of ATLAS Laser Thermal Control System Demonstration Unit

    Science.gov (United States)

    Ku, Jentung; Robinson, Franklin; Patel, Deepak; Ottenstein, Laura

    2013-01-01

    The second Ice, Cloud, and Land Elevation Satellite mission currently planned by National Aeronautics and Space Administration will measure global ice topography and canopy height using the Advanced Topographic Laser Altimeter System {ATLAS). The ATLAS comprises two lasers; but only one will be used at a time. Each laser will generate between 125 watts and 250 watts of heat, and each laser has its own optimal operating temperature that must be maintained within plus or minus 1 degree Centigrade accuracy by the Laser Thermal Control System (LTCS) consisting of a constant conductance heat pipe (CCHP), a loop heat pipe (LHP) and a radiator. The heat generated by the laser is acquired by the CCHP and transferred to the LHP, which delivers the heat to the radiator for ultimate rejection. The radiator can be exposed to temperatures between minus 71 degrees Centigrade and minus 93 degrees Centigrade. The two lasers can have different operating temperatures varying between plus 15 degrees Centigrade and plus 30 degrees Centigrade, and their operating temperatures are not known while the LTCS is being designed and built. Major challenges of the LTCS include: 1) A single thermal control system must maintain the ATLAS at 15 degrees Centigrade with 250 watts heat load and minus 71 degrees Centigrade radiator sink temperature, and maintain the ATLAS at plus 30 degrees Centigrade with 125 watts heat load and minus 93 degrees Centigrade radiator sink temperature. Furthermore, the LTCS must be qualification tested to maintain the ATLAS between plus 10 degrees Centigrade and plus 35 degrees Centigrade. 2) The LTCS must be shut down to ensure that the ATLAS can be maintained above its lowest desirable temperature of minus 2 degrees Centigrade during the survival mode. No software control algorithm for LTCS can be activated during survival and only thermostats can be used. 3) The radiator must be kept above minus 65 degrees Centigrade to prevent ammonia from freezing using no more

  20. Vector boson and quarkonia production in lead-lead collisions with ATLAS detector

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00015179; The ATLAS collaboration

    2017-01-01

    Photons and weak bosons do not interact strongly with the dense and hot medium formed in the nuclear collisions, thus should be sensitive to the nuclear modification of parton distribution functions (nPDFs). The in-medium modification of heavy quarkonium states plays an important role in studying the hot and dense medium formed in the larger collision systems. The ATLAS detector at the LHC, optimized for searching for new physics in proton-proton collisions, is especially well equipped to measure photons, Z, W bosons and quarkonia in the high occupancy environment produced in heavy-ion collisions. We present recent results on Z boson and charmonia yields as a functions of centrality, transverse momentum, and rapidity, from the ATLAS experiment.

  1. Benefits and performance of ATLAS approaches to utilizing opportunistic resources

    CERN Document Server

    Filip\\v{c}i\\v{c}, Andrej; The ATLAS collaboration

    2016-01-01

    ATLAS has been extensively exploring possibilities of using computing resources extending beyond conventional grid sites in the WLCG fabric to deliver as many computing cycles as possible and thereby enhance the significance of the Monte-Carlo samples to deliver better physics results. The difficulties of using such opportunistic resources come from architectural differences such as unavailability of grid services, the absence of network connectivity on worker nodes or inability to use standard authorization protocols. Nevertheless, ATLAS has been extremely successful in running production payloads on a variety of sites, thanks largely to the job execution workflow design in which the job assignment, input data provisioning and execution steps are clearly separated and can be offloaded to custom services. To transparently include the opportunistic sites in the ATLAS central production system, several models with supporting services have been developed to mimic the functionality of a full WLCG site. Some are e...

  2. Electroweak measurements with the ATLAS detector.

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00349845; The ATLAS collaboration

    2015-01-01

    ATLAS measurements of multi-boson production processes involving combinations of W, Z and isolated photons at 8 TeV are summarized. Standard Model cross section are measured with high precision by ATLAS and are compared to world averages. Production processes sensitive to vector-boson fusion and vector-boson scattering are also presented and used for the triple and quartic gauge boson couplings limits setting.

  3. Single top quark production cross-section and properties using the ATLAS detector at the LHC

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00358737; The ATLAS collaboration

    2015-01-01

    ATLAS measurements of single top-quark processes are summarized. Measurements using data at 7 and 8 TeV collisions from the LHC are presented. The measurements are performed using the semi-leptonic decay mode of the top-quark. Production cross-sections and the $\\left|V_{tb}\\right|$ CKM matrix element extraction are shown. All measurements are compared to theoretical calculations. In addition, the $s$-channel production mode is explored along with limits on exotic production modes.

  4. The NeuARt II system: a viewing tool for neuroanatomical data based on published neuroanatomical atlases

    Directory of Open Access Journals (Sweden)

    Cheng Wei-Cheng

    2006-12-01

    Full Text Available Abstract Background Anatomical studies of neural circuitry describing the basic wiring diagram of the brain produce intrinsically spatial, highly complex data of great value to the neuroscience community. Published neuroanatomical atlases provide a spatial framework for these studies. We have built an informatics framework based on these atlases for the representation of neuroanatomical knowledge. This framework not only captures current methods of anatomical data acquisition and analysis, it allows these studies to be collated, compared and synthesized within a single system. Results We have developed an atlas-viewing application ('NeuARt II' in the Java language with unique functional properties. These include the ability to use copyrighted atlases as templates within which users may view, save and retrieve data-maps and annotate them with volumetric delineations. NeuARt II also permits users to view multiple levels on multiple atlases at once. Each data-map in this system is simply a stack of vector images with one image per atlas level, so any set of accurate drawings made onto a supported atlas (in vector graphics format could be uploaded into NeuARt II. Presently the database is populated with a corpus of high-quality neuroanatomical data from the laboratory of Dr Larry Swanson (consisting 64 highly-detailed maps of PHAL tract-tracing experiments, made up of 1039 separate drawings that were published in 27 primary research publications over 17 years. Herein we take selective examples from these data to demonstrate the features of NeuArt II. Our informatics tool permits users to browse, query and compare these maps. The NeuARt II tool operates within a bioinformatics knowledge management platform (called 'NeuroScholar' either as a standalone or a plug-in application. Conclusion Anatomical localization is fundamental to neuroscientific work and atlases provide an easily-understood framework that is widely used by neuroanatomists and non

  5. Higgs Boson fermionic production and decay modes with the ATLAS detector

    CERN Document Server

    Piacquadio, Giacinto; The ATLAS collaboration

    2018-01-01

    The Higgs couplings to heavy quarks, top and bottom quarks, are difficult to directly access experimentally, nevertheless their measurement is crucial for a full characterization of the Higgs sector and to probe a large variety of physics scenarios beyond the Standard Model. In this talk I will discuss the results recently obtained by the ATLAS Collaboration based on Run-2 2015 and 2016 data, which led to evidence for Higgs boson decays to b-quarks and for the ttH production mode.

  6. Detector Control System for the ATLAS Forward Proton detector

    CERN Document Server

    Czekierda, Sabina; The ATLAS collaboration

    2017-01-01

    The ATLAS Forward Proton (AFP) is a forward detector using a Roman Pot technique, recently installed in the LHC tunnel. It is aiming at registering protons that were diffractively or electromagnetically scattered in soft and hard processes. Infrastructure of the detector consists of hardware placed both in the tunnel and in the control room USA15 (about 330 meters from the Roman Pots). AFP detector, like the other detectors of the ATLAS experiment, uses the Detector Control System (DCS) to supervise the detector and to ensure its safe and coherent operation, since the incorrect detector performance may influence the physics results. The DCS continuously monitors the detector parameters, subset of which is stored in data bases. Crucial parameters are guarded by alarm system. A detector representation as a hierarchical tree-like structure of well-defined subsystems built with the use of the Finite State Machine (FSM) toolkit allows for overall detector operation and visualization. Every node in the hierarchy is...

  7. Recent advances in precision laser cutting for the ATLAS hadron calorimeter absorbers production

    International Nuclear Information System (INIS)

    Alikov, B.; Budagov, Yu.

    1995-01-01

    The optimised precision laser cutting technology for high tolerances ATLAS hadron calorimeter absorber plates production is described. Some recent results of laser cut absorber plates dimension measurements are presented. The plates have been manufactured by 'Universalmash' (S.-Petersburg) and RCTL RAS (Shatura). It has been shown that the proved accuracy of the laser machines is not worse than 45 microns. 9 figs

  8. Production and qualification of 40 km of Al-stabilized NbTi cable for the ATLAS experiment at CERN

    CERN Document Server

    Baccaglioni, G; Cartegni, G C; Horváth, I L; Neuenschwander, J; Pedrini, D; Rossi, L; Volpini, G

    2002-01-01

    The production of the conductor for the superconducting toroids of the ATLAS experiment at LHC (CERN) is now in progress. The toroid system, composed of one barrel toroid (BT) and two end cap toroids (ECTs), exploits aluminum-clad Rutherford-type NbTi conductors of large size (57 * 12 mm for BT, 42 * 12 mm for ECTs) and high critical current (Ic) (58 kA for BT and 60 kA for ECTs @ 4.2 K, 5 T). Some 55 km of conductor are required for the BT and 26 km for the ECTs, respectively. An Italian-Swiss (ETH Zurich and INFN) consortium is in charge of the delivery of half of the whole amount. This paper describes the results of this production with particular emphasis to the quality control system developed to monitor the production with both on-line controls and the post-production quality assessment protocols. The main result is the confirmation that the technologies selected and the whole process are reliable and reproducible over large production quantities. The overall degradation due to the cabling and the co-ex...

  9. ATCA-based ATLAS FTK input interface system

    Energy Technology Data Exchange (ETDEWEB)

    Okumura, Yasuyuki [Chicago U., EFI; Liu, Tiehui Ted [Fermilab; Olsen, Jamieson [Fermilab; Iizawa, Tomoya [Waseda U.; Mitani, Takashi [Waseda U.; Korikawa, Tomohiro [Waseda U.; Yorita, Kohei [Waseda U.; Annovi, Alberto [Frascati; Beretta, Matteo [Frascati; Gatta, Maurizio [Frascati; Sotiropoulou, C-L. [Aristotle U., Thessaloniki; Gkaitatzis, Stamatios [Aristotle U., Thessaloniki; Kordas, Konstantinos [Aristotle U., Thessaloniki; Kimura, Naoki [Aristotle U., Thessaloniki; Cremonesi, Matteo [Chicago U., EFI; Yin, Hang [Fermilab; Xu, Zijun [Peking U.

    2015-04-27

    The first stage of the ATLAS Fast TracKer (FTK) is an ATCA-based input interface system, where hits from the entire silicon tracker are clustered and organized into overlapping eta-phi trigger towers before being sent to the tracking engines. First, FTK Input Mezzanine cards receive hit data and perform clustering to reduce data volume. Then, the ATCA-based Data Formatter system will organize the trigger tower data, sharing data among boards over full mesh backplanes and optic fibers. The board and system level design concepts and implementation details, as well as the operation experiences from the FTK full-chain testing, will be presented.

  10. ATCA-based ATLAS FTK input interface system

    CERN Document Server

    Okumura, Y; The ATLAS collaboration; Olsen, J; Iizawa, T; Mitani, T; Korikawa, T; Yorita, K; Annovi, A; Beretta, M; Gatta, M; Sotiropoulou, C; Gkaitatzis, S; Kordas, K; Kimura, N; Cremonesi, M; Yin, H; Xu, Z

    2014-01-01

    The first stage of the ATLAS Fast TracKer (FTK) is an ATCA-based input interface system, where hits from the entire silicon tracker must be clustered and organized into overlapping eta-phi trigger towers before being sent to the tracking processors. First, FTK Input Mezzanine cards receive hit data and perform clustering to reduce data volume. Then, the ATCA-based Data Formatter system will organize the trigger tower data, sharing data among boards over a full-mesh backplane. The board and system level performance studies and implementation details, as well as the operation experiences from the FTK full-chain testing, will be presented.

  11. Top quark production cross-section measurements with the ATLAS detector at EPS-HEP 2017

    CERN Document Server

    Pollard, Christopher Samuel; The ATLAS collaboration

    2017-01-01

    Measurements of the inclusive and differential top-quark pair and single-top production cross sections in proton-proton collisions with the ATLAS detector at the Large Hadron Collider at center-of-mass energies of 8 TeV and 13 TeV are presented. The inclusive measurements reach high precision and are compared to the best available theoretical calculations. Differential measurements of the kinematic properties of the top-quark production are also discussed. These measurements, including results using boosted tops, probe our understanding of top-quark pair production in the TeV regime.

  12. FELIX - the new detector readout system for the ATLAS experiment

    CERN Document Server

    AUTHOR|(SzGeCERN)754725; The ATLAS collaboration; Anderson, John Thomas; Borga, Andrea; Boterenbrood, Hendrik; Chen, Hucheng; Chen, Kai; Drake, Gary; Donszelmann, Mark; Francis, David; Gorini, Benedetto; Guest, Daniel; Lanni, Francesco; Lehmann Miotto, Giovanna; Levinson, Lorne; Roich, Alexander; Schreuder, Frans Philip; Schumacher, J\\"orn; Vandelli, Wainer; Vermeulen, Jos; Wu, Weihao; Zhang, Jinlong

    2016-01-01

    From the ATLAS Phase-I upgrade and onward, new or upgraded detectors and trigger systems will be interfaced to the data acquisition, detector control and timing (TTC) systems by the Front-End Link eXchange (FELIX). FELIX is the core of the new ATLAS Trigger/DAQ architecture. Functioning as a router between custom serial links and a commodity network, FELIX is implemented by server PCs with commodity network interfaces and PCIe cards with large FPGAs and many high speed serial fiber transceivers. By separating data transport from data manipulation, the latter can be done by software in commodity servers attached to the network. Replacing traditional point-to-point links between Front-end components and the DAQ system by a switched network, FELIX provides scaling, flexibility uniformity and upgradability. Different Front-end data types or different data sources can be routed to different network endpoints that handle that data type or source: e.g. event data, configuration, calibration, detector control, monito...

  13. The Resource Manager the ATLAS Trigger and Data Acquisition System

    CERN Document Server

    Aleksandrov, Igor; The ATLAS collaboration; Lehmann Miotto, Giovanna; Soloviev, Igor

    2016-01-01

    The Resource Manager of the ATLAS Trigger and Data Acquisition system The Resource Manager is one of the core components of the Data Acquisition system of the ATLAS experiment at the LHC. The Resource Manager marshals the right for applications to access resources which may exist in multiple but limited copies, in order to avoid conflicts due to program faults or operator errors. The access to resources is managed in a manner similar to what a lock manager would do in other software systems. All the available resources and their association to software processes are described in the Data Acquisition configuration database. The Resource Manager is queried about the availability of resources every time an application needs to be started. The Resource Manager’s design is based on a client-server model, hence it consists of two components: the Resource Manager "server" application and the "client" shared library. The Resource Manager server implements all the needed functionalities, while the Resource Manager c...

  14. The Data Acquisition and Calibration System for the ATLAS Semiconductor Tracker

    CERN Document Server

    Abdesselam, A; Barr, A J; Bell, P; Bernabeu, J; Butterworth, J M; Carter, J R; Carter, A A; Charles, E; Clark, A; Colijn, A P; Costa, M J; Dalmau, J M; Demirkoz, B; Dervan, P J; Donega, M; D'Onifrio, M; Escobar, C; Fasching, D; Ferguson, D P S; Ferrari, P; Ferrère, D; Fuster, J; Gallop, B; García, C; González, S; González-Sevilla, S; Goodrick, M J; Gorisek, A; Greenall, A; Grillo, A A; Hessey, N P; Hill, J C; Jackson, J N; Jared, R C; Johannson, P D C; de Jong, P; Joseph, J; Lacasta, C; Lane, J B; Lester, C G; Limper, M; Lindsay, S W; McKay, R L; Magrath, C A; Mangin-Brinet, M; Martí i García, S; Mellado, B; Meyer, W T; Mikulec, B; Minano, M; Mitsou, V A; Moorhead, G; Morrissey, M; Paganis, E; Palmer, M J; Parker, M A; Pernegger, H; Phillips, A; Phillips, P W; Postranecky, M; Robichaud-Véronneau, A; Robinson, D; Roe, S; Sandaker, H; Sciacca, F; Sfyrla, A; Stanecka, E; Stapnes, S; Stradling, A; Tyndel, M; Tricoli, A; Vickey, T; Vossebeld, J H; Warren, M R M; Weidberg, A R; Wells, P S; Wu, S L

    2008-01-01

    The SemiConductor Tracker (SCT) data acquisition (DAQ) system will calibrate, configure, and control the approximately six million front-end channels of the ATLAS silicon strip detector. It will provide a synchronized bunch-crossing clock to the front-end modules, communicate first-level triggers to the front-end chips, and transfer information about hit strips to the ATLAS high-level trigger system. The system has been used extensively for calibration and quality assurance during SCT barrel and endcap assembly and for performance confirmation tests after transport of the barrels and endcaps to CERN. Operating in data-taking mode, the DAQ has recorded nearly twenty million synchronously-triggered events during commissioning tests including almost a million cosmic ray triggered events. In this paper we describe the components of the data acquisition system, discuss its operation in calibration and data-taking modes and present some detector performance results from these tests.

  15. The data acquisition and calibration system for the ATLAS Semiconductor Tracker

    International Nuclear Information System (INIS)

    Abdesselam, A; Barr, A J; Demirkoez, B; Barber, T; Carter, J R; Bell, P; Bernabeu, J; Costa, M J; Escobar, C; Butterworth, J M; Carter, A A; Dalmau, J M; Charles, E; Fasching, D; Ferguson, D P S; Clark, A; Donega, M; D'Onifrio, M; Colijn, A-P; Dervan, P J

    2008-01-01

    The SemiConductor Tracker (SCT) data acquisition (DAQ) system will calibrate, configure, and control the approximately six million front-end channels of the ATLAS silicon strip detector. It will provide a synchronized bunch-crossing clock to the front-end modules, communicate first-level triggers to the front-end chips, and transfer information about hit strips to the ATLAS high-level trigger system. The system has been used extensively for calibration and quality assurance during SCT barrel and endcap assembly and for performance confirmation tests after transport of the barrels and endcaps to CERN. Operating in data-taking mode, the DAQ has recorded nearly twenty million synchronously-triggered events during commissioning tests including almost a million cosmic ray triggered events. In this paper we describe the components of the data acquisition system, discuss its operation in calibration and data-taking modes and present some detector performance results from these tests

  16. Integrated System for Performance Monitoring of ATLAS TDAQ Network

    CERN Document Server

    Savu, D; The ATLAS collaboration; Martin, B; Sjoen, R; Batraneanu, S; Stancu, S

    2010-01-01

    The ATLAS TDAQ Network consists of three separate networks spanning four levels of the experimental building. Over 200 edge switches and 5 multi-blade chassis routers are used to interconnect 2000 processors, adding up to more than 7000 high speed interfaces. In order to substantially speed-up ad-hoc and post mortem analysis, a scalable, yet flexible, integrated system for monitoring both network statistics and environmental conditions, processor parameters and data taking characteristics was required. For successful up-to-the-minute monitoring, information from many SNMP compliant devices, independent databases and custom APIs was gathered, stored and displayed in an optimal way. Easy navigation and compact aggregation of multiple data sources were the main requirements; characteristics not found in any of the tested products, either open-source or commercial. This paper describes how performance, scalability and display issues were addressed and what challenges the project faced during development and deplo...

  17. Recent ATLAS Articles on WLAP

    CERN Multimedia

    Goldfarb, S.

    As reported in the September 2004 ATLAS eNews, the Web Lecture Archive Project is a system for the archiving and publishing of multimedia presentations, using the Web as medium. We list here newly available WLAP items relating to ATLAS: June ATLAS Plenary Meeting Tutorial on Physics EDM and Tools (June) Freiburg Overview Week Ketevi Assamagan's Tutorial on Analysis Tools Click here to browse WLAP for all ATLAS lectures.

  18. Control in the ATLAS TDAQ System

    CERN Document Server

    Liko, D; Flammer, J; Dobson, M; Jones, R; Mapelli, L; Alexandrov, I; Korobov, S; Kotov, V; Mineev, M; Amorim, A; Fiuza de Barros, N; Klose, D; Pedro, L; Badescu, E; Caprini, M; Kolos, S; Kazarov, A; Ryabov, Yu; Soloviev, I; Computing In High Energy Physics

    2005-01-01

    TDAQ system requires a comprehensive and flexible control system. Its role ranges from the so-called run-control, e.g. starting and stopping the data taking, to error handling and fault tolerance. It also includes initialization and verification of the overall system. Following the traditional approach a hierarchical system of customizable controllers has been proposed. For the final system all functionality will be therefore available in a distributed manner, with the possibility of local customization. After a technology survey the open source expert system CLIPS has been chosen as a basis for the implementation of the supervision and the verification system. The CLIPS interpreter has been extended to provide a general control framework. Other ATLAS Online software components have been integrated as plug-ins and provide the mechanism for configuration and communication. Several components have been implemented sharing this technology. The dynamic behavior of the individual component is fully described by th...

  19. Fine grained event processing on HPCs with the ATLAS Yoda system

    CERN Document Server

    Calafiura, Paolo; The ATLAS collaboration; Guan, Wen; Maeno, Tadashi; Nilsson, Paul; Oleynik, Danila; Panitkin, Sergey; Tsulaia, Vakhtang; van Gemmeren, Peter; Wenaus, Torre

    2015-01-01

    High performance computing facilities present unique challenges and opportunities for HENP event processing. The massive scale of many HPC systems means that fractionally small utilizations can yield large returns in processing throughput. Parallel applications which can dynamically and efficiently fill any scheduling opportunities the resource presents benefit both the facility (maximal utilization) and the (compute-limited) science. The ATLAS Yoda system provides this capability to HENP-like event processing applications by implementing event-level processing in an MPI-based master-client model that integrates seamlessly with the more broadly scoped ATLAS Event Service. Fine grained, event level work assignments are intelligently dispatched to parallel workers to sustain full utilization on all cores, with outputs streamed off to destination object stores in near real time with similarly fine granularity, such that processing can proceed until termination with full utilization. The system offers the efficie...

  20. The liquid helium system of ATLAS

    International Nuclear Information System (INIS)

    Nixon, J.M.; Bollinger, L.M.

    1989-01-01

    Starting in 1978 with one small refrigerator and distribution line, the LHe system of ATLAS has gradually grown into a complex network, as required by several enlargements of the superconducting linac. The cryogenic system now comprises 3 refrigerators, 11 helium compressors, /approximately/340 ft. of coaxial LHe transfer line, 3 1000-l dewars, and /approximately/76 LHe valves that deliver steady-state flowing LHe to 16 beam-line cryostats. In normal operation, the 3 refrigerators are linked so as to provide cooling where needed. LHe heat exchangers in distribution lines play an important role. This paper discusses design features of the system, including the logic of the controls that permit the coupled refrigerators to operate stably in the presence of large and sudden changes in heat load. 8 refs., 3 figs

  1. The ATLAS Liquid Argon Calorimeter: Construction, Integration, Commissioning

    International Nuclear Information System (INIS)

    Aleksa, Martin

    2006-01-01

    The ATLAS liquid argon (LAr) calorimeter system consists of an electromagnetic barrel calorimeter and two end caps with electromagnetic, hadronic and forward calorimeters. The liquid argon sampling technique, with an accordion geometry was chosen for the barrel electromagnetic calorimeter (EMB) and adapted to the end cap (EMEC). The hadronic end cap calorimeter (HEC) uses a copper-liquid argon sampling technique with flat plate geometry and is subdivided in depth in two wheels per end-cap. Finally, the forward calorimeter (FCAL) is composed of three modules employing cylindrical electrodes with thin liquid argon gaps.The construction of the full calorimeter system is complete since mid-2004. Production modules constructed in the home institutes were integrated into wheels at CERN in 2003-2004, and inserted into the three cryostats. They passed their first complete cold test before the lowering into the ATLAS cavern. Results of quality checks (e.g. electrical, mechanical, ...) performed on all the 190304 read-out channels after cool down will be reported. End 2004 the ATLAS barrel electromagnetic (EM) calorimeter was installed in the ATLAS cavern and since summer 2005 the front-end electronics are being connected and tested. Results of this first commissioning phase will be shown to demonstrate the high standards of quality control for our detectors

  2. ATLAS Fact Sheet : To raise awareness of the ATLAS detector and collaboration on the LHC

    CERN Multimedia

    ATLAS Outreach

    2010-01-01

    Facts on the Detector, Calorimeters, Muon System, Inner Detector, Pixel Detector, Semiconductor Tracker, Transition Radiation Tracker,, Surface hall, Cavern, Detector, Magnet system, Solenoid, Toroid, Event rates, Physics processes, Supersymmetric particles, Comparing LHC with Cosmic rays, Heavy ion collisions, Trigger and Data Acquisition TDAQ, Computing, the LHC and the ATLAS collaboration. This fact sheet also contains images of ATLAS and the collaboration as well as a short list of videos on ATLAS available for viewing.

  3. MBAT: A scalable informatics system for unifying digital atlasing workflows

    Directory of Open Access Journals (Sweden)

    Sane Nikhil

    2010-12-01

    Full Text Available Abstract Background Digital atlases provide a common semantic and spatial coordinate system that can be leveraged to compare, contrast, and correlate data from disparate sources. As the quality and amount of biological data continues to advance and grow, searching, referencing, and comparing this data with a researcher's own data is essential. However, the integration process is cumbersome and time-consuming due to misaligned data, implicitly defined associations, and incompatible data sources. This work addressing these challenges by providing a unified and adaptable environment to accelerate the workflow to gather, align, and analyze the data. Results The MouseBIRN Atlasing Toolkit (MBAT project was developed as a cross-platform, free open-source application that unifies and accelerates the digital atlas workflow. A tiered, plug-in architecture was designed for the neuroinformatics and genomics goals of the project to provide a modular and extensible design. MBAT provides the ability to use a single query to search and retrieve data from multiple data sources, align image data using the user's preferred registration method, composite data from multiple sources in a common space, and link relevant informatics information to the current view of the data or atlas. The workspaces leverage tool plug-ins to extend and allow future extensions of the basic workspace functionality. A wide variety of tool plug-ins were developed that integrate pre-existing as well as newly created technology into each workspace. Novel atlasing features were also developed, such as supporting multiple label sets, dynamic selection and grouping of labels, and synchronized, context-driven display of ontological data. Conclusions MBAT empowers researchers to discover correlations among disparate data by providing a unified environment for bringing together distributed reference resources, a user's image data, and biological atlases into the same spatial or semantic context

  4. Rucio - The next generation large scale distributed system for ATLAS Data Management

    CERN Document Server

    Beermann, T; The ATLAS collaboration; Lassnig, M; Barisits, M; Vigne, R; Serfon, C; Stewart, G A; Goossens, L; Nairz, A; Molfetas, A

    2014-01-01

    Rucio is the next-generation Distributed Data Management (DDM) system benefiting from recent advances in cloud and "Big Data" computing to address the ATLAS experiment scaling requirements. Rucio is an evolution of the ATLAS DDM system Don Quijote 2 (DQ2), which has demonstrated very large scale data management capabilities with more than 150 petabytes spread worldwide across 130 sites, and accesses from 1,000 active users. However, DQ2 is reaching its limits in terms of scalability, requiring a large number of support staff to operate and being hard to extend with new technologies. Rucio will deal with these issues by relying on new technologies to ensure system scalability, address new user requirements and employ a new automation framework to reduce operational overheads.

  5. Deployment of job priority mechanisms in the Italian Cloud of the ATLAS experiment

    International Nuclear Information System (INIS)

    Doria, Alessandra; Carlino, Gianpaolo; Salvo, Alessandro De; Musto, Elisa; Barchiesi, Alex; Campana, Simone; Ciocca, Claudia; Italiano, Alessandro; Rinaldi, Lorenzo; Salomoni, Davide; Perini, Laura; Pistolese, Massimo; Vaccarossa, Luca; Vilucchi, Elisabetta

    2010-01-01

    An optimized use of the Grid computing resources in the ATLAS experiment requires the enforcement of a mechanism of job priorities and of resource sharing among the different activities inside the ATLAS VO. This mechanism has been implemented through the VOViews publication in the information system and the fair share implementation per UNIX group in the batch system. The VOView concept consists of publishing resource information, such as running and waiting jobs, as a function of VO groups and roles. The ATLAS Italian Cloud is composed of the CNAF Tier1 and Roma Tier2, with farms based on the LSF batch system, and the Tier2s of Frascati, Milano and Napoli based on PBS/Torque. In this paper we describe how test and deployment of the job priorities has been performed in the cloud, where the VOMS-based regional group /atlas/it has been created. We show that the VOViews are published and correctly managed by the WMS and that the resources allocated to generic VO users, users with production role and users of the /atlas/it group correspond to the defined share.

  6. Input Mezzanine Card for the Fast Tracker at ATLAS

    CERN Document Server

    Iizawa, Tomoya; The ATLAS collaboration

    2016-01-01

    The Fast Tracker (FTK) is an integral part of trigger upgrade program for the ATLAS experiment. At LHC Run 2, which started operations in June 2015 at a center-of-mass energy of 13 TeV, the luminosity could reach up to 2*1034 cm-2s-1 and an average of 40-50 simultaneous proton collisions per beam crossing will be expected. The higher luminosity demands a more sophisticated trigger system with increased use of tracking information. The Fast Tracker is a highly-parallel hardware system that rapidly finds and reconstructs tracks in the ATLAS inner-detector at the triggering stage. This paper focuses on the FTK Input Mezzanine Board that is input module of entire system. The functions of this board are to receive the insertable b-layer, pixel and micro-strip data from the ATLAS Silicon read-out drivers, perform clustering, and forward the data to its mother board. Mass production and quality control tests of Mezzanine Boards were completed, and staged installation and commissioning are ongoing. Details of its fun...

  7. ATLAS Calorimeter system: Run-2 performance, Phase-1 and Phase-2 upgrades

    CERN Document Server

    Starz, Steffen; The ATLAS collaboration

    2018-01-01

    The ATLAS detector was designed and built to study proton-proton collisions produced at the LHC at centre-of-mass energies up to 14 TeV and instantaneous luminosities up to 10^{34} cm^{−2} s^{−1}. A liquid argon-lead sampling calorimeter (LAr) is employed as electromagnetic calorimeter and hadronic calorimeter, except in the barrel region, where a scintillator-steel sampling calorimeter (TileCal) is used as hadronic calorimeter. ATLAS recorded 87 fb^{-1} of data at a center-of-mass energy of 13 TeV between 2015 and 2017. In order to achieve the level-1 acceptance rate of 100 kHz, certain adjustments have been performed. The calorimetry system performed accordingly to its design values and have played a crucial role in the ATLAS physics programme. This contribution will give an overview of the detector operation, monitoring and data quality, as well as the achieved performance, including the calibration and stability of the energy scale, noise level, response uniformity and time resolution of the ATLAS cal...

  8. Data federation strategies for ATLAS using XRootD

    International Nuclear Information System (INIS)

    Gardner, Robert; Vukotic, Ilija; Campana, Simone; Iven, Jan; Duckeck, Guenter; Elmsheuser, Johannes; Hönig, Friedrich G; Legger, Federica; Hanushevsky, Andrew; Yang, Wei

    2014-01-01

    In the past year the ATLAS Collaboration accelerated its program to federate data storage resources using an architecture based on XRootD with its attendant redirection and storage integration services. The main goal of the federation is an improvement in the data access experience for the end user while allowing more efficient and intelligent use of computing resources. Along with these advances come integration with existing ATLAS production services (PanDA and its pilot services) and data management services (DQ2, and in the next generation, Rucio). Functional testing of the federation has been integrated into the standard ATLAS and WLCG monitoring frameworks and a dedicated set of tools provides high granularity information on its current and historical usage. We use a federation topology designed to search from the site's local storage outward to its region and to globally distributed storage resources. We describe programmatic testing of various federation access modes including direct access over the wide area network and staging of remote data files to local disk. To support job-brokering decisions, a time-dependent cost-of-data-access matrix is made taking into account network performance and key site performance factors. The system's response to production-scale physics analysis workloads, either from individual end-users or ATLAS analysis services, is discussed.

  9. WWγ and WZγ production at the ATLAS experiment

    Energy Technology Data Exchange (ETDEWEB)

    Baas, Alessandra Edda

    2017-07-26

    The study of three electroweak gauge bosons is not only a precise test of the electroweak structure of the Standard Model, but it is also sensitive to new physics. In this thesis, the measurement of WWγ and WZγ production in proton-proton collisions is presented, in which events are analysed where the W boson decays leptonically and the other heavy gauge boson decays hadronically. Two decay channels differing in the flavour of the final state lepton are studied: the electron and the muon channel. The analysed proton-proton collisions are recorded in 2012 with the ATLAS experiment at a centre-of-mass energy of √(s) = 8 TeV and correspond to an integrated luminosity of 20.2 fb{sup -1}. Exclusion limits on the production cross section are derived at 95 % confidence level (CL) in two different phase space regions: one is optimised for the Standard Model production and one for the search of new physics beyond the Standard Model. The best upper limits on the WWγ and WZγ production, determined in this analysis, exclude cross sections above 2.5 times the Standard Model expectation. In addition, Frequentist limits at 95% CL are derived on fourteen different anomalous quartic gauge couplings of mass dimension eight, using the framework of an effective field theory. The exclusion limits are presented with and without unitarisation.

  10. The ATLAS/TILECAL Detector Control System

    CERN Document Server

    Santos, H; The ATLAS collaboration

    2010-01-01

    Tilecal, the barrel hadronic calorimeter of ATLAS, is a sampling calorimeter where scintillating tiles are embedded in an iron matrix. The tiles are optically coupled to wavelength shifting fibers that carry the optical signal to photo-multipliers. It has a cylindrical shape and is made out of 3 cylinders, the Long Barrel with the LBA and LBC partitions, and the two Extended Barrel with the EBA and EBC partitions. The main task of the Tile calorimeter Detector Control System (DCS) is to enable the coherent and safe operation of the calorimeter. All actions initiated by the operator, as well as all errors, warnings and alarms concerning the hardware of the detector are handled by DCS. The Tile calorimeter DCS controls and monitors mainly the low voltage and high voltage power supply systems, but it is also interfaced with the infrastructure (cooling system and racks), the laser and cesium calibration systems, the data acquisition system, configuration and conditions databases and the detector safety system. In...

  11. ATLAS Magnet System Nearing Completion

    CERN Document Server

    ten Kate, H H J

    2008-01-01

    The ATLAS Detector at the Large Hadron Collider at CERN is equipped with a superconducting magnet system that consists of a Barrel Toroid, two End-Cap Toroids and a Central Solenoid. The four magnets generate the magnetic field for the muon- and inner tracking detectors, respectively. After 10 years of construction in industry, integration and on-surface tests at CERN, the magnets are now in the underground cavern where they undergo the ultimate test before data taking in the detector can start during the course of next year. The system with outer dimensions of 25 m length and 22 m diameter is based on using conduction cooled aluminum stabilized NbTi conductors operating at 4.6 K and 20.5 kA maximum coil current with peak magnetic fields in the windings of 4.1 T and a system stored magnetic energy of 1.6 GJ. The Barrel Toroid and Central Solenoid were already successfully charged after installation to full current in autumn 2006. This year the system is completed with two End Cap Toroids. The ultimate test of...

  12. Measurement of Single-top t-channel Production Using ATLAS Data

    CERN Document Server

    Holzbauer, Jenny; Schwienhorst, R; Repko, W; Lapidus, L; Pratt, S

    2012-01-01

    This document reports the measurement of the single-top t-channel cross-section using data from the ATLAS detector, located at the Large Hadron Collider on the border of France and Switzerland. The data used were collected during the first half of 2011, from proton-proton collisions with a 7 TeV center-of-mass collision energy. Single-top is electroweak top-quark production and t-channel is one of the standard model production modes. To isolate this production, selections are applied to find events with a similar final state. A cut-based analysis is used to further isolate the signal using a series of selections in several orthogonal kinematic regions. Finally, a statistical analysis is performed to determine the measured cross-section and the CKM matrix element |Vtb|. The cross-section for top and anti-top production is considered separately and the resulting cross-sections are σ t+ = 59 +18 -16 pb for the positive charge channel and σ t- = 33 +13 -12 pb for the negative charge chan...

  13. ATLAS silicon microstrip Semiconductor Tracker (SCT)

    International Nuclear Information System (INIS)

    Unno, Y.

    2000-01-01

    Silicon microstrip semiconductor tracking system (SCT) will be in operation in the ATLAS detector in the Large Hadron Collider (LHC) at CERN. Challenging issues in the SCT are the radiation tolerance to the fluence of 2x10 14 1-MeV-neutron-equivalent particles/cm 2 at the designed luminosity of 1x10 34 cm -2 /s of the proton-proton collisions and the speed of the electronics to identify the crossing bunches at 25 ns. The developments and the status of the SCT are presented from the point of view of these issues. Series production of the SCT will start in the year 2001 and the SCT will be installed into the ATLAS detector during 2003-2004

  14. ATLAS Tier-2 monitoring system for the German cloud

    Energy Technology Data Exchange (ETDEWEB)

    Meyer, Joerg; Quadt, Arnulf; Weber, Pavel [II. Physikalisches Institut, Georg-August-Universitaet, Goettingen (Germany)

    2011-07-01

    The ATLAS tier centers in Germany provide their computing resources for the ATLAS experiment. The stable and sustainable operation of this so-called DE-cloud heavily relies on effective monitoring of the Tier-1 center GridKa and its associated Tier-2 centers. Central and local grid information services constantly collect and publish the status information from many computing resources and sites. The cloud monitoring system discussed in this presentation evaluates the information related to different cloud resources and provides a coherent and comprehensive view of the cloud. The main monitoring areas covered by the tool are data transfers, cloud software installation, site batch systems, Service Availability Monitoring (SAM). The cloud monitoring system consists of an Apache-based Python application, which retrieves the information and publishes it on the generated HTML web page. This results in an easy-to-use web interface for the limited number of sites in the cloud with fast and efficient access to the required information starting from a high level summary for the whole cloud to detailed diagnostics for the single site services. This approach provides the efficient identification of correlated site problems and simplifies the administration on both cloud and site level.

  15. Application of the ATLAS DAQ and Monitoring System for MDT and RPC Commissioning

    CERN Document Server

    Pasqualucci, E

    2007-01-01

    The ATLAS DAQ and monitoring software are currently commonly used to test detectors during the commissioning phase. In this paper, their usage in MDT and RPC commissioning is described, both at the surface pre-commissioning and commissioning stations and in the ATLAS pit. Two main components are heavily used for detector tests. The ROD Crate DAQ software is based on the ATLAS Readout application. Based on the plug-in mechanism, it provides a complete environment to interface any kind of detector or trigger electronics to the ATLAS DAQ system. All the possible flavours of this application are used to test and run the MDT and RPC detectors at the pre-commissioning and commissioning sites. Ad-hoc plug-ins have been developed to implement data readout via VME, both with ROD prototypes and emulating final electronics to read out data with temporary solutions, and to provide trigger distribution and busy management in a multi-crate environment. Data driven event building functionality is also used to combine data f...

  16. The LUCID detector ATLAS luminosity monitor and its electronic system

    CERN Document Server

    AUTHOR|(INSPIRE)INSPIRE-00378808; The ATLAS collaboration

    2016-01-01

    Starting from 2015 LHC is performing a new run, at higher center of mass energy (13 TeV) and with 25 ns bunch-spacing. The ATLAS luminosity monitor LUCID has been completely renewed, both on detector design and in the electronics, in order to cope with the new running conditions. The new detector electronics is presented, featuring a new read-out board (LUCROD), for signal acquisition and digitization, PMT-charge integration and single-side luminosity measurements, and the revisited LUMAT board for side-A-side-C combination. The contribution covers the new boards design, the firmware and software developments, the implementation of luminosity algorithms, the optical communication between boards and the integration into the ATLAS TDAQ system.

  17. NODC Standard Product: International ocean atlas Volume 11 - Climatic atlas of the Sea of Azov 2008 (1 disc set) (NODC Accession 0098574)

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — This Atlas, Climatic Atlas of the Sea of Azov 2008 on CD-ROM, is an update to Volume 10, Climatic Atlas of the Sea of Azov 2006 on CD-ROM (NODC Accession 0098572),...

  18. NASA Technical Interchange Meeting (TIM): Advanced Technology Lifecycle Analysis System (ATLAS) Technology Tool Box

    Science.gov (United States)

    ONeil, D. A.; Craig, D. A.; Christensen, C. B.; Gresham, E. C.

    2005-01-01

    The objective of this Technical Interchange Meeting was to increase the quantity and quality of technical, cost, and programmatic data used to model the impact of investing in different technologies. The focus of this meeting was the Technology Tool Box (TTB), a database of performance, operations, and programmatic parameters provided by technologists and used by systems engineers. The TTB is the data repository used by a system of models known as the Advanced Technology Lifecycle Analysis System (ATLAS). This report describes the result of the November meeting, and also provides background information on ATLAS and the TTB.

  19. Further studies on simulation of top-quark production for the ATLAS experiment at √s = 13 TeV

    CERN Document Server

    The ATLAS collaboration

    2016-01-01

    This note summarizes recent studies of modern Monte Carlo generator setups for the pair production of top quarks at the LHC. Samples at a center of mass energy of 7 or 13 TeV have been generated using various setup such as MadGraph5_aMC@NLO+Pythia8 at leading order, MadGraph5_aMC@NLO+Pythia8 at next-to-leading order using the FxFx merging scheme prescription, Sherpa 2.2 and Powheg+Herwig7. The performance of these samples for several physical observables are compared to the ATLAS baseline generator configuration, Powheg+Pythia6, and to √s = 7 TeV ATLAS measurements. New studies of tt production in association with bb pair are also presented.

  20. Measurement of WW production with the ATLAS detector and constraints on associated triple gauge couplings

    CERN Document Server

    Liu, JB; The ATLAS collaboration

    2013-01-01

    The measurement of the WW diboson production cross section in proton proton interactions with the ATLAS experiment is presented. The measurement is performed using W boson fully leptonical decays including ee, emu and mumu final states associated with large missing energy. Measured total and differential cross sections of WW production are compared to NLO predictions of the standard model. Leading lepton pT distribution is used to place constraints on anomalous triple gauge boson couplings.

  1. The Tilecal/ATLAS detector control system

    CERN Document Server

    Tomasio Pina, João Antonio

    2004-01-01

    Tilecal is the barrel hadronic calorimeter of the ATLAS detector that is presently being built at CERN to operate at the LHC accelerator. The main task of the Tilecal detector control system (DCS) is to enable the coherent and safe operation of the detector. All actions initiated by the operator and all errors, warnings, and alarms concerning the hardware of the detector are handled by DCS. The DCS has to continuously monitor all operational parameters, give warnings and alarms concerning the hardware of the detector. The DCS architecture consists of a distributed back-end (BE) system running on PC's and different front-end (FE) systems. The implementation of the BE will he achieved with a commercial supervisory control and data acquisition system (SCADA) and the FE instrumentation will consist on a wide variety of equipment. The connection between the FE and BE is provided by fieldbus or L

  2. The Evolution of the Trigger and Data Acquisition System in the ATLAS Experiment

    CERN Document Server

    Garelli, N; The ATLAS collaboration

    2014-01-01

    The ATLAS experiment, aimed at recording the results of LHC proton-proton collisions, is upgrading its Trigger and Data Acquisition (TDAQ) system during the current LHC first long shutdown. The purpose of such upgrade is to add robustness and flexibility to the selection and the conveyance of the physics data, simplify the maintenance of the infrastructure, exploit new technologies and, overall, make ATLAS data-taking capable of dealing with increasing event rates. \

  3. Radioactive sources for ATLAS hadron tile calorimeter calibration

    International Nuclear Information System (INIS)

    Budagov, Yu.; Cavalli-Sforza, M.; Ivanyushenkov, Yu.

    1997-01-01

    The main requirements for radioactive sources applied in the TileCal calibration systems are formulated; technology of the sources production developed in the Laboratory of Nuclear Problems, JINR is described. Design and characteristics of the prototype sources manufactured in Dubna and tested on ATLAS TileCal module 0 are presented

  4. Studies for the detector control system of the ATLAS pixel at the HL-LHC

    International Nuclear Information System (INIS)

    Püllen, L; Becker, K; Boek, J; Kersten, S; Kind, P; Mättig, P; Zeitnitz, C

    2012-01-01

    In the context of the LHC upgrade to the HL-LHC the inner detector of the ATLAS experiment will be replaced completely. As part of this redesign there will also be a new pixel detector. This new pixel detector requires a control system which meets the strict space requirements for electronics in the ATLAS experiment. To accomplish this goal we propose a DCS (Detector Control System) network with the smallest form factor currently available. This network consists of a DCS chip located in close proximity to the interaction point and a DCS controller located in the outer regions of the ATLAS detector. These two types of chips form a star shaped network with several DCS chips being controlled by one DCS controller. Both chips are manufactured in deep sub-micron technology. We present prototypes with emphasis on studies concerning single event upsets.

  5. The ATLAS Data Management Software Engineering Process

    CERN Document Server

    Lassnig, M; The ATLAS collaboration; Stewart, G A; Barisits, M; Beermann, T; Vigne, R; Serfon, C; Goossens, L; Nairz, A; Molfetas, A

    2014-01-01

    Rucio is the next-generation data management system of the ATLAS experiment. The software engineering process to develop Rucio is fundamentally different to existing software development approaches in the ATLAS distributed computing community. Based on a conceptual design document, development takes place using peer-reviewed code in a test-driven environment. The main objectives are to ensure that every engineer understands the details of the full project, even components usually not touched by them, that the design and architecture are coherent, that temporary contributors can be productive without delay, that programming mistakes are prevented before being committed to the source code, and that the source is always in a fully functioning state. This contribution will illustrate the workflows and products used, and demonstrate the typical development cycle of a component from inception to deployment within this software engineering process. Next to the technological advantages, this contribution will also hi...

  6. The ATLAS Data Management Software Engineering Process

    CERN Document Server

    Lassnig, M; The ATLAS collaboration; Stewart, G A; Barisits, M; Beermann, T; Vigne, R; Serfon, C; Goossens, L; Nairz, A

    2013-01-01

    Rucio is the next-generation data management system of the ATLAS experiment. The software engineering process to develop Rucio is fundamentally different to existing software development approaches in the ATLAS distributed computing community. Based on a conceptual design document, development takes place using peer-reviewed code in a test-driven environment. The main objectives are to ensure that every engineer understands the details of the full project, even components usually not touched by them, that the design and architecture are coherent, that temporary contributors can be productive without delay, that programming mistakes are prevented before being committed to the source code, and that the source is always in a fully functioning state. This contribution will illustrate the workflows and products used, and demonstrate the typical development cycle of a component from inception to deployment within this software engineering process. Next to the technological advantages, this contribution will also hi...

  7. PanDA for ATLAS Distributed Computing in the Next Decade

    CERN Document Server

    Barreiro Megino, Fernando Harald; The ATLAS collaboration

    2016-01-01

    The Production and Distributed Analysis (PanDA) system has been developed to meet ATLAS production and analysis requirements for a data-driven workload management system capable of operating at the Large Hadron Collider (LHC) data processing scale. Heterogeneous resources used by the ATLAS experiment are distributed worldwide at hundreds of sites, thousands of physicists analyse the data remotely, the volume of processed data is beyond the exabyte scale, dozens of scientific applications are supported, while data processing requires more than a few billion hours of computing usage per year. PanDA performed very well over the last decade including the LHC Run 1 data taking period. However, it was decided to upgrade the whole system concurrently with the LHC’s first long shutdown in order to cope with rapidly changing computing infrastructure. After two years of reengineering efforts, PanDA has embedded capabilities for fully dynamic and flexible workload management. The static batch job paradigm was discarde...

  8. PanDA for ATLAS distributed computing in the next decade

    CERN Document Server

    AUTHOR|(SzGeCERN)643806; The ATLAS collaboration; De, Kaushik; Klimentov, Alexei; Maeno, Tadashi; Nilsson, Paul; Oleynik, Danila; Padolski, Siarhei; Panitkin, Sergey; Wenaus, Torre

    2017-01-01

    The Production and Distributed Analysis (PanDA) system has been developed to meet ATLAS production and analysis requirements for a data-driven workload management system capable of operating at the Large Hadron Collider (LHC) data processing scale. Heterogeneous resources used by the ATLAS experiment are distributed worldwide at hundreds of sites, thousands of physicists analyse the data remotely, the volume of processed data is beyond the exabyte scale, dozens of scientific applications are supported, while data processing requires more than a few billion hours of computing usage per year. PanDA performed very well over the last decade including the LHC Run 1 data taking period. However, it was decided to upgrade the whole system concurrently with the LHC’s first long shutdown in order to cope with rapidly changing computing infrastructure. After two years of reengineering efforts, PanDA has embedded capabilities for fully dynamic and flexible workload management. The static batch job paradigm was discarde...

  9. PanDA Beyond ATLAS: Workload Management for Data Intensive Science

    CERN Document Server

    Schovancova, J; The ATLAS collaboration; Klimentov, A; Maeno, T; Nilsson, P; Oleynik, D; Panitkin, S; Petrosyan, A; Vaniachine, A; Wenaus, T; Yu, D

    2013-01-01

    The PanDA Production ANd Distributed Analysis system has been developed by ATLAS to meet the experiment's requirements for a data-driven workload management system for production and distributed analysis processing capable of operating at LHC data processing scale. After 7 years of impressively successful PanDA operation in ATLAS there are also other experiments which can benefit from PanDA in the Big Data challenge, with several at various stages of evaluation and adoption. The new project "Next Generation Workload Management and Analysis System for Big Data" is extending PanDA to meet the needs of other data intensive scientific applications in HEP, astro-particle and astrophysics communities, bio-informatics and other fields as a general solution to large scale workload management. PanDA can utilize dedicated or opportunistic computing resources such as grids, clouds, and High Performance Computing facilities, and is being extended to leverage next generation intelligent networks in automated workflow mana...

  10. Searches for direct pair production of third generation squarks with the ATLAS detector

    CERN Document Server

    Gonski, Julia Lynne; The ATLAS collaboration

    2018-01-01

    Searches for supersymmetry are a key focus of the LHC experimental program. In particular, natural SUSY models motivate third generation squarks with masses light enough to be produced at the LHC. As a result, the ATLAS experiment has a variety of analyses devoted to stop/sbottom direct production. In this talk, recent results from these searches for are discussed, and the current stop/sbottom mass exclusion limits are presented.

  11. The High-Resolution IRAS Galaxy Atlas

    Science.gov (United States)

    Cao, Yu; Terebey, Susan; Prince, Thomas A.; Beichman, Charles A.; Oliversen, R. (Technical Monitor)

    1997-01-01

    An atlas of the Galactic plane (-4.7 deg is less than b is less than 4.7 deg), along with the molecular clouds in Orion, rho Oph, and Taurus-Auriga, has been produced at 60 and 100 microns from IRAS data. The atlas consists of resolution-enhanced co-added images with 1 min - 2 min resolution and co-added images at the native IRAS resolution. The IRAS Galaxy Atlas, together with the Dominion Radio Astrophysical Observatory H(sub I) line/21 cm continuum and FCRAO CO (1-0) Galactic plane surveys, which both have similar (approx. 1 min) resolution to the IRAS atlas, provides a powerful tool for studying the interstellar medium, star formation, and large-scale structure in our Galaxy. This paper documents the production and characteristics of the atlas.

  12. ATLAS diamond Beam Condition Monitor

    Energy Technology Data Exchange (ETDEWEB)

    Gorisek, A. [CERN (Switzerland)]. E-mail: andrej.gorisek@cern.ch; Cindro, V. [J. Stefan Institute (Slovenia); Dolenc, I. [J. Stefan Institute (Slovenia); Frais-Koelbl, H. [Fotec (Austria); Griesmayer, E. [Fotec (Austria); Kagan, H. [Ohio State University, OH (United States); Korpar, S. [J. Stefan Institute (Slovenia); Kramberger, G. [J. Stefan Institute (Slovenia); Mandic, I. [J. Stefan Institute (Slovenia); Meyer, M. [CERN (Switzerland); Mikuz, M. [J. Stefan Institute (Slovenia); Pernegger, H. [CERN (Switzerland); Smith, S. [Ohio State University, OH (United States); Trischuk, W. [University of Toronto (Canada); Weilhammer, P. [CERN (Switzerland); Zavrtanik, M. [J. Stefan Institute (Slovenia)

    2007-03-01

    The ATLAS experiment has chosen to use diamond for its Beam Condition Monitor (BCM) given its radiation hardness, low capacitance and short charge collection time. In addition, due to low leakage current diamonds do not require cooling. The ATLAS Beam Condition Monitoring system is based on single beam bunch crossing measurements rather than integrating the accumulated particle flux. Its fast electronics will allow separation of LHC collisions from background events such as beam gas interactions or beam accidents. There will be two stations placed symmetrically about the interaction point along the beam axis at z=+/-183.8cm. Timing of signals from the two stations will provide almost ideal separation of beam-beam interactions and background events. The ATLAS BCM module consists of diamond pad detectors of 1cm{sup 2} area and 500{mu}m thickness coupled to a two-stage RF current amplifier. The production of the final detector modules is almost done. A S/N ratio of 10:1 has been achieved with minimum ionizing particles (MIPs) in the test beam setup at KEK. Results from the test beams and bench measurements are presented.

  13. ATLAS diamond Beam Condition Monitor

    International Nuclear Information System (INIS)

    Gorisek, A.; Cindro, V.; Dolenc, I.; Frais-Koelbl, H.; Griesmayer, E.; Kagan, H.; Korpar, S.; Kramberger, G.; Mandic, I.; Meyer, M.; Mikuz, M.; Pernegger, H.; Smith, S.; Trischuk, W.; Weilhammer, P.; Zavrtanik, M.

    2007-01-01

    The ATLAS experiment has chosen to use diamond for its Beam Condition Monitor (BCM) given its radiation hardness, low capacitance and short charge collection time. In addition, due to low leakage current diamonds do not require cooling. The ATLAS Beam Condition Monitoring system is based on single beam bunch crossing measurements rather than integrating the accumulated particle flux. Its fast electronics will allow separation of LHC collisions from background events such as beam gas interactions or beam accidents. There will be two stations placed symmetrically about the interaction point along the beam axis at z=+/-183.8cm. Timing of signals from the two stations will provide almost ideal separation of beam-beam interactions and background events. The ATLAS BCM module consists of diamond pad detectors of 1cm 2 area and 500μm thickness coupled to a two-stage RF current amplifier. The production of the final detector modules is almost done. A S/N ratio of 10:1 has been achieved with minimum ionizing particles (MIPs) in the test beam setup at KEK. Results from the test beams and bench measurements are presented

  14. ATLAS-AWS

    International Nuclear Information System (INIS)

    Gehrcke, Jan-Philip; Stonjek, Stefan; Kluth, Stefan

    2010-01-01

    We show how the ATLAS offline software is ported on the Amazon Elastic Compute Cloud (EC2). We prepare an Amazon Machine Image (AMI) on the basis of the standard ATLAS platform Scientific Linux 4 (SL4). Then an instance of the SLC4 AMI is started on EC2 and we install and validate a recent release of the ATLAS offline software distribution kit. The installed software is archived as an image on the Amazon Simple Storage Service (S3) and can be quickly retrieved and connected to new SL4 AMI instances using the Amazon Elastic Block Store (EBS). ATLAS jobs can then configure against the release kit using the ATLAS configuration management tool (cmt) in the standard way. The output of jobs is exported to S3 before the SL4 AMI is terminated. Job status information is transferred to the Amazon SimpleDB service. The whole process of launching instances of our AMI, starting, monitoring and stopping jobs and retrieving job output from S3 is controlled from a client machine using python scripts implementing the Amazon EC2/S3 API via the boto library working together with small scripts embedded in the SL4 AMI. We report our experience with setting up and operating the system using standard ATLAS job transforms.

  15. The ATLAS Data Flow system in Run2: Design and Performance

    CERN Document Server

    Rifki, Othmane; The ATLAS collaboration

    2016-01-01

    The ATLAS detector uses a real time selective triggering system to reduce the high interaction rate from 40 MHz to its data storage capacity of 1 kHz. A hardware first level trigger limits the rate to 100 kHz and a software high level trigger selects events for offline analysis. By building on the experience gained during the successful first run of the LHC, the ATLAS Trigger and Data Acquisition system has been simplified and upgraded to take advantage of state of the art technologies. The Dataflow element of the system is composed of distributed hardware and software responsible for buffering and transporting event data from the Readout system to the High Level Trigger and to the event storage. This system has been reshaped in order to maximize the flexibility and efficiency of the data selection process. The updated dataflow is different from the previous implementation both in terms of architecture and performance. The biggest difference is within the high level trigger, where the merger of region-of-inte...

  16. Module and electronics developments for the ATLAS ITK pixel system

    CERN Document Server

    Nellist, Clara; The ATLAS collaboration

    2016-01-01

    Summary ATLAS is preparing for an extensive modification of its detector in the course of the planned HL‐ LHC accelerator upgrade around 2025 which includes a replacement of the entire tracking system by an all‐silicon detector (Inner Tracker, ITk). A revised trigger and data taking system is foreseen with triggers expected at lowest level at an average rate of 1 MHz. The five innermost layers of ITk will comprise of a pixel detector built of new sensor and readout electronics technologies to improve the tracking performance and cope with the severe HL‐LHC environment in terms of occupancy and radiation. The total area of the new pixel system could measure up to 14 m2, depending on the final layout choice that is expected to take place in early 2017. A new on‐detector readout chip is designed in the context of the RD53 collaboration in 65 nm CMOS technology. This paper will present the on‐going R&D within the ATLAS ITK project towards the new pixel modules and the off‐detector electronics. Pla...

  17. Strong Production SUSY Searches at ATLAS and CMS

    CERN Document Server

    Marshall, Z L

    2015-01-01

    The results of searches for strongly-produced supersymmetry at the Large Hadron Collider by the ATLAS and CMS collaborations are presented. Several of the historically strongest zero-and one-lepton final state searches have been updated to include multi-bin fits and combinations. In addition, new two-lepton final state search results are shown from CMS and ATLAS, which show 2.6 and 3.0 standard deviation excesses, respectively, above the standard model expectation, albeit in different regions of phase space. Both collaborations have also shown new searches that cover areas uncovered by previous searches, in both searches for light stops and searches for stealth supersymmetry.

  18. Simulation of top quark production for the ATLAS experiment at sqrt(s) = 13 TeV

    CERN Document Server

    The ATLAS collaboration

    2016-01-01

    This note summarises the Monte Carlo simulation setup for the pair and single production of top quarks for the ATLAS experiment at the LHC for sqrt(s)=13 TeV. In addition to the settings available and recommended for analyses using the 2015 dataset, the anticipated setup for 2016 analysis is also discussed.

  19. Advanced Alignment of the ATLAS Tracking System

    CERN Document Server

    Butti, P; The ATLAS collaboration

    2014-01-01

    In order to reconstruct the trajectories of charged particles, the ATLAS experiment exploits a tracking system built using different technologies, planar silicon modules or microstrips (PIX and SCT detectors) and gaseous drift tubes (TRT), all embedded in a 2T solenoidal magnetic field. Misalignments and deformations of the active detector elements deteriorate the track reconstruction resolution and lead to systematic biases on the measured track parameters. The alignment procedures exploits various advanced tools and techniques in order to determine for module positions and correct for deformations. For the LHC Run II, the system is being upgraded with the installation of a new pixel layer, the Insertable B-layer (IBL).

  20. Taus at ATLAS

    Energy Technology Data Exchange (ETDEWEB)

    Demers, Sarah M. [Yale Univ., New Haven, CT (United States). Dept. of Physics

    2017-12-06

    The grant "Taus at ATLAS" supported the group of Sarah Demers at Yale University over a period of 8.5 months, bridging the time between her Early Career Award and her inclusion on Yale's grant cycle within the Department of Energy's Office of Science. The work supported the functioning of the ATLAS Experiment at CERN's Large Hadron Collider and the analysis of ATLAS data. The work included searching for the Higgs Boson in a particular mode of its production (with a W or Z boson) and decay (to a pair of tau leptons.) This was part of a broad program of characterizing the Higgs boson as we try to understand this recently discovered particle, and whether or not it matches our expectations within the current standard model of particle physics. In addition, group members worked with simulation to understand the physics reach of planned upgrades to the ATLAS experiment. Supported group members include postdoctoral researcher Lotte Thomsen and graduate student Mariel Pettee.