Graeme Stewart
(University of Glasgow (GB))
4/13/15, 11:00 AM
Christian Nieke
(Brunswick Technical University (DE))
4/13/15, 2:00 PM
Optimising a computing infrastructure on the scale of LHC requires a quantitative understanding of a complex network of many different resources and services. For this purpose the CERN IT department and the LHC experiments are collecting a large multitude of logs and performance probes, which are already successfully used for short-term analysis (e.g. operational dashboards) within each...
Andrea Formica
(CEA/IRFU,Centre d'etude de Saclay Gif-sur-Yvette (FR))
4/13/15, 2:00 PM
The ATLAS and CMS Conditions Database infrastructures have served each of the respective experiments well through LHC Run 1, providing efficient access to a wide variety of conditions information needed in online data taking and offline processing and analysis. During the long shutdown between Run 1 and Run 2, we have taken various measures to improve our systems for Run 2. In some cases, a...
Stephen Gowdy
(Fermi National Accelerator Lab. (US))
4/13/15, 2:00 PM
The global distributed computing system (WLCG) used by the Large Hadron
Collider (LHC) is evolving. The treatment of wide-area-networking (WAN) as
a scarce resource that needs to be strictly managed is far less
necessary that originally foreseen. Static data placement and replication,
intended to limit interdependencies among computing centers, is giving way
to global data federations...
Federico Stagni
(CERN)
4/13/15, 2:00 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
In the last few years, new types of computing infrastructures, such as IAAS (Infrastructure as a Service) and IAAC (Infrastructure as a Client), gained popularity. New resource may come as part of pledged resources, while others are in the form of opportunistic ones. Most of these new infrastructures are based on virtualization techniques, others don't. Meanwhile, some concepts, such as...
Frederic Bruno Magniette
(Ecole Polytechnique (FR))
4/13/15, 2:00 PM
High-energy physics experiments produce huge amounts of data that need to be processed and stored for further analysis and eventually treated in real time for triggering and monitoring purposes. In addition, more and more often these requirements are also being found on other fields such as on-line video processing, proteomics and astronomical facilities.
The complexity of such experiments...
Ian Fisk
(Fermi National Accelerator Lab. (US))
4/13/15, 2:15 PM
Beginning in 2015 CMS will collected and produce data and simulation adding to 10B new events a year. In order to realize the physics potential of the experiment these events need to be stored, processed, and delivered to analysis users on a global scale. CMS has 150k processor cores and 80PB of disk storage and there is constant pressure to reduce the resources needed and increase the...
David Lange
(Lawrence Livermore Nat. Laboratory (US))
4/13/15, 2:15 PM
Over the past several years, the CMS experiment has made significant changes to its detector simulation and reconstruction applications motivated by the planned program of detector upgrades over the next decade. These upgrades include both completely new tracker and calorimetry systems and changes to essentially all major detector components to meet the requirements of very high pileup...
Dr
paolo branchini
(INFN Roma Tre)
4/13/15, 2:15 PM
The Data Acquisition System (DAQ) and the Front-End electronics for an array of Kinetic Inductance Detectors (KIDs) are described. KIDs are superconductive detectors, in which electrons are organized in Cooper pairs. Any incident radiation could break such pairs generating quasi-particles, whose effect is increasing the inductance of the detector. Electrically, any KID is equivalent to a...
Tadashi Maeno
(Brookhaven National Laboratory (US))
4/13/15, 2:15 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
Experiments at the Large Hadron Collider (LHC) face unprecedented computing challenges. Heterogeneous resources are distributed worldwide at hundreds of sites, thousands of physicists analyze the data remotely, the volume of processed data is beyond the exabyte scale, while data processing requires more than a few billion hours of computing usage per year. The PanDA (Production and Distributed...
Birgit Lewendel
(Deutsches Elektronen-Synchrotron (DE))
4/13/15, 2:30 PM
DESY operates a multi-VO Grid site for 20 HEP and non-HEP
collaborations and is one of the world-wide largest Tier-2 sites for
ATLAS, CMS, LHCb, and BELLE2. In one common Grid infrastructure
computing resources are shared by all VOs according to MoUs and
agreements, applying an opportunistic usage model allows to distribute
free resources among the VOs. Currently, the Grid site...
Dr
Antonio Perez-Calero Yzquierdo
(Centro de Investigaciones Energ. Medioambientales y Tecn. - (ES)
4/13/15, 2:30 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
The successful exploitation of the multicore processor architectures available at the computing sites is a key element of the LHC distributed computing system in the coming era of the LHC Run 2. High-pileup complex-collision events represent a challenge for the traditional sequential programming in terms of memory and processing time budget. The CMS data production and processing framework has...
Roland Sipos
(Eotvos Lorand University (HU))
4/13/15, 2:30 PM
With the restart of the LHC in 2015, the growth of the CMS Conditions dataset will continue, therefore the need of consistent and highly available access to the Conditions makes a great cause to revisit different aspects of the current data storage solutions.
We present a study of alternative data storage backends for the Conditions Databases, by evaluating some of the most popular NoSQL...
Nathalie Rauschmayr
(CERN)
4/13/15, 2:45 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
The main goal of a Workload Management System (WMS) is to find and allocate resources for the jobs it is handling. The more and more accurate information the WMS receives about the jobs, the easier it will be to accomplish its task, which will directly translate into a better utilization of resources. Traditionally, the information associated with each job, like expected runtime or memory...
Barbara Storaci
(Universitaet Zuerich (CH))
4/13/15, 2:45 PM
The LHCb track reconstruction uses sophisticated pattern recognition algorithms to reconstruct trajectories of charged particles. Their main feature is the use of a Hough-transform like approach to connect track segments from different subdetectors, allowing for having no tracking stations in the magnet of LHCb. While yielding a high efficiency, the track reconstruction is a major contributor...
Eduard Ebron Simioni
(Johannes-Gutenberg-Universitaet Mainz (DE))
4/13/15, 2:45 PM
The Large Hadron Collider (LHC) in 2015 will collide proton beams with
increased luminosity from $10^{34}$ up to $3 \times 10^{34}$ cm$^{−2}$ s$^{−1}$. ATLAS
is an LHC experiment designed to measure decay properties of highly
energetic particles produced in these proton-collisions. The high
luminosity places stringent physical and operational requirements on
the ATLAS Trigger in order to...
Dr
Peter Elmer
(Princeton University (US))
4/13/15, 3:00 PM
Deploying the Worldwide LHC Computing Grid (WLCG) was greatly
facilitated by the convergence, around the year 2000, on Linux and
commodity x86 processors as a standard scientific computing platform.
This homogeneity enabled a relatively simple "build once, run
anywhere" model for applications. A number of factors are now driving
interest in alternative platforms. Power limitations at the...
Jiri Chudoba
(Acad. of Sciences of the Czech Rep. (CZ))
4/13/15, 3:00 PM
Pierre Auger Observatory operates the largest system of detectors for ultra-high energy cosmic ray measurements. Comparison of theoretical models of interactions with recorded data requires thousands of computing cores for Monte Carlo simulations. Since 2007 distributed resources connected via EGI grid are succesfully used. The first and the second versions of production system based on bash...
Dominick Rocco
(urn:Google)
4/13/15, 3:00 PM
The NOvA experiment is a long baseline neutrino oscillation experiment utilizing the NuMI beam generated at Fermilab. The experiment will measure the oscillations within a muon neutrino beam in a 300 ton Near Detector located underground at Fermilab and a functionally-identical 14 kiloton Far Detector placed 810 km away. The detectors are liquid scintillator tracking calorimeters with a...
Helio Takai
(Brookhaven National Laboratory (US))
4/13/15, 3:00 PM
The global feature extractor (gFEX) is a component of the Level-1
Calorimeter trigger Phase-I upgrade for the ATLAS experiment. It is
intended to identify patterns of energy associated with the hadronic
decays of high momentum Higgs, W, & Z bosons, topquarks, and exotic
particles in real time at the LHC crossing rate. The single processor
board will be implemented as a fast reconfigurable...
Ms
Marina Golosova
(National Research Centre "Kurchatov Institute")
4/13/15, 3:00 PM
In recent years the concepts of Big Data became well established in IT-technologies. Most systems (for example Distributed Data Management or Workload Management systems) produce metadata that describes actions performed on jobs, stored data or other entities and its volume takes one to the realms of Big Data on many occasions. This metadata can be used to obtain information about the current...
James Letts
(Univ. of California San Diego (US))
4/13/15, 3:00 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
CMS will require access to more than 125k processor cores for the beginning of Run2 in 2015 to carry out its ambitious physics program with more and higher complexity events. During Run1 these resources were predominantly provided by a mix of grid sites and local batch resources. During the long shut down cloud infrastructures, diverse opportunistic resources and HPC supercomputing centers...
Michael Boehler
(Albert-Ludwigs-Universitaet Freiburg (DE))
4/13/15, 3:15 PM
The ATLAS detector consists of several sub-detector systems. Both data taking and Monte Carlo (MC) simulation rely on an accurate description of the detector conditions from every sub system, such as calibration constants, different scenarios of pile-up and noise conditions, size and position of the beam spot, etc. In order to guarantee database availability for critical online applications...
Shawn Mc Kee
(University of Michigan (US))
4/13/15, 3:15 PM
The Worldwide LHC Computing Grid relies on the network as a critical part of its infrastructure and therefore needs to guarantee effective network usage and prompt detection and resolution of any network issues, including connection failures, congestion, traffic routing, etc. The WLCG Network and Transfer Metrics project aims to integrate and combine all network-related monitoring data...
Giuseppe Cerati
(Univ. of California San Diego (US))
4/13/15, 3:15 PM
Power density constraints are limiting the performance improvements of modern CPUs. To address this we have seen the introduction of lower-power, multi-core processors, but the future will be even more exciting. In order to stay within the power density limits but still obtain Moore's Law performance/price gains, it will be necessary to parallelize algorithms to exploit larger numbers of...
Alec Habig
(Univ. of Minnesota Duluth)
4/13/15, 3:15 PM
The NOvA experiment at Fermilab is a long-baseline neutrino experiment designed to study nu-e appearance in a nu-mu beam. Over the last few years there has been intense work to streamline the computing infrastructure in preparation for data, which started to flow in from the far detector in Fall 2013. Major accomplishments for this effort include migration to the use of offsite resources...
Vincent Garonne
(CERN)
4/13/15, 3:15 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
For more than 8 years, the Distributed Data Management (DDM) system of ATLAS called DQ2 has been able to demonstrate very large scale data management capabilities with more than 600M files, 160 petabytes spread worldwide across 130 sites, and accesses from 1,000 active users. However, the system does not scale for LHC run2 and a new DDM system called Rucio has been developed to be DQ2's...
Bruno Heinrich Hoeft
(KIT - Karlsruhe Institute of Technology (DE))
4/13/15, 3:30 PM
The Steinbuch Center for Computing (SCC) at Karlsruhe Institute of Technology (KIT) was involved quite early in 100G network technology. In 2010 already a first 100G wide area network testbed over a distance of approx. 450 km was deployed between the national research organizations KIT and FZ-Jülich - initiated by DFN (the German NREN). Only three years later 2013, KIT joined the Caltech SC13...
Dr
Baosong Shan
(Beihang University (CN))
4/13/15, 3:30 PM
The Alpha Magnetic Spectrometer (AMS) is a high energy physics experiment installed and operating on board of the International Space Station (ISS) from May 2011 and expected to last through Year 2024 and beyond. The computing strategy of the AMS experiment is discussed in the paper, including software design, data processing and modelling details, simulation of the detector performance and...
Ludovico Bianchi
(Forschungszentrum Jülich)
4/13/15, 3:30 PM
The PANDA experiment is a next generation particle detector planned for operation at the FAIR facility, currently under construction in Darmstadt, Germany. PANDA will detect events generated by colliding an antiproton beam on a fixed proton target, allowing studies in hadron spectroscopy, hypernuclei production, open charm and nucleon structure.
The nature of hadronic collisions means that...
Martin Barisits
(CERN)
4/13/15, 3:30 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
The ATLAS Distributed Data Management system stores more than 160PB of physics data across more than 130 sites globally. Rucio, the next-generation data management system of ATLAS has been introduced to cope with the anticipated workload of the coming decade. The previous data management system DQ2 pursued a rather simplistic approach for resource management, but with the increased data volume...
Dr
Dario Barberis
(Università e INFN Genova (IT))
4/13/15, 3:30 PM
The EventIndex is the complete catalogue of all ATLAS events, keeping the references to all files that contain a given event in any processing stage. It replaces the TAG database, which had been in use during LHC Run 1. For each event it contains its identifiers, the trigger pattern and the GUIDs of the files containing it. Major use cases are event picking, feeding the Event Service used on...
Dr
Ivan Kisel
(Johann-Wolfgang-Goethe Univ. (DE))
4/13/15, 3:45 PM
The future heavy-ion experiment CBM (FAIR/GSI, Darmstadt, Germany) will focus on the measurement of very rare probes at interaction rates up to 10 MHz with data flow of up to 1 TB/s. The beam will provide free stream of beam particles without bunch structure. That requires full online event reconstruction and selection not only in space, but also in time, so-called 4D event building and...
Adam Jedrzej Otto
(Ministere des affaires etrangeres et europeennes (FR))
4/13/15, 3:45 PM
The LHCb experiment is preparing a major upgrade of both the detector and the data acquisition system. A system capable of transporting up to 50 Tbps of data will be required. This can only be achieved in a manageable way using 100 Gbps links. Such links recently became available also in the servers, while they have been available between switches already for a while.
We present first...
Dr
Tony Wildish
(Princeton University (US))
4/13/15, 3:45 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
AsyncStageOut (ASO) is a new component of the distributed data analysis system of CMS, CRAB, designed for managing users' data. It addresses a major weakness of the previous model, namely that data movement was part of the job execution resulting in inefficient use of job slots and an unacceptable failure rate at the end of the jobs.
ASO foresees the management of up to 400k files per day...
Dr
Takashi SUGIMOTO
(Japan Synchrotron Radiation Research Institute)
4/13/15, 3:45 PM
An X-ray free electron laser (XFEL) facility, SACLA, is generating ultra-short, high peak brightness, and full-spatial-coherent X-ray pulses [1]. The unique characteristics of the X-ray pulses, which have never been obtained with conventional synchrotron orbital radiation, are now opening new opportunities in a wide range of scientific fields such as atom, molecular and optical physics,...
Javier Sanchez
(Instituto de Fisica Corpuscular (ES))
4/13/15, 3:45 PM
The ATLAS EventIndex contains records of all events processed by ATLAS, in all processing stages. These records include the references to the files containing each event (the GUID of the file) and the internal “pointer” to each event in the file. This information is collected by all jobs that run at Tier-0 or on the Grid and process ATLAS events. Each job produces a snippet of information for...
Christoph Paus
(Massachusetts Inst. of Technology (US))
4/13/15, 4:30 PM
The Dynamic Data Management (DDM) framework is designed to manage the majority of the CMS data in an automated fashion. At the moment 51 CMS Tier-2 data centers have the ability to host about 20 PB of data. Tier-1 centers will also be included adding substantially more space. The goal of DDM is to facilitate the management of the data distribution and optimize the accessibility of data for the...
Dr
Junichi Kanzaki
(KEK)
4/13/15, 4:30 PM
Track8: Performance increase and optimization exploiting hardware features
oral presentation
Fast event generation system of physics processes is developed using graphics processing unit (GPU).
The system is based on the Monte Carlo integration and event generation programs, BASES/SPRING, which were originally developed in FORTRAN.
They were rewritten on the CUDA platform provided by NVIDIA in order for the implementation of these programs to GPUs.
Since the Monte Carlo integration...
David Schultz
(University of Wisconsin-Madison)
4/13/15, 4:30 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
We describe the overall structure and new features of the second generation of IceProd, a data processing and management framework. IceProd was developed by the IceCube Neutrino Observatory for processing of Monte Carlo simulations and detector data, and has been a key component of the IceCube offline computing infrastructure since it was first deployed in 2006. It runs fully in user space as...
Marko Bracko
(Jozef Stefan Institute (SI))
4/13/15, 4:30 PM
The Belle II experiment, a next-generation B factory experiment at the KEK laboratory, Tsukuba, Japan, is expected to collect an experimental data sample fifty times larger than its predecessor, the Belle experiment. The data taking and processing rates are expected to be at least one order of magnitude larger as well.
In order to cope with these large data processing rates and huge data...
Thomas Beermann
(Bergische Universitaet Wuppertal (DE))
4/13/15, 4:45 PM
This contribution presents a study on the applicability and usefulness of dynamic data placement methods for data-intensive systems, such as ATLAS distributed data management (DDM). In this system the jobs are sent to the data, therefore having a good distribution of data is significant. Ways of forecasting workload patterns are examined which then are used to redistribute data to achieve a...
Hideki Miyake
(KEK)
4/13/15, 4:45 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
In Belle II experiment a large amount of physics data will be continuously taken and the production rate is equivalent to LHC experiments.
Considerable resources of computing, storage, and network, are necessary to handle not only the taken data but also substantial simulated data.
Therefore Belle II exploits distributed computing system based on DIRAC interware.
DIRAC is a general...
Dr
Ulrich Schwickerath
(CERN)
4/13/15, 4:45 PM
As part of CERN's Agile Infrastructure project, large parts of the CERN batch farm have been moved to virtual machines running on CERNs private IaaS (link is external) cloud. During this process a large fraction of the resources, which had previously been used as physical batch worker nodes, were converted into hypervisors. Due to the large spread of the per-core performance (rated in HS06) in...
Scott Snyder
(Brookhaven National Laboratory (US))
4/13/15, 4:45 PM
During the 2013-2014 shutdown of the Large Hadron Collider, ATLAS switched to a new event data model for analysis, called the xAOD. A key feature of this model is the separation of the object data from the objects themselves (the `auxiliary store'). Rather being stored as member variables of the analysis classes, all object data are stored separately, as vectors of simple values. Thus, the...
Dr
Sami Kama
(Southern Methodist University (US))
4/13/15, 4:45 PM
Track8: Performance increase and optimization exploiting hardware features
oral presentation
The growing size and complexity of events produced at the high luminosities expected in 2015 at the Large Hadron Collider demands much more computing power for the online event selection and for the offline data reconstruction than in the previous data taking period. In recent years, the explosive performance growth of low-cost, massively parallel processors like Graphical Processing Units...
Dr
Peter Van Gemmeren
(Argonne National Laboratory (US))
4/13/15, 5:00 PM
ATLAS developed and employed for Run 1 of the Large Hadron Collider a sophisticated infrastructure for metadata handling in event processing jobs. This infrastructure profits from a rich feature set provided by the ATLAS execution control framework, including standardized interfaces and invocation mechanisms for tools and services, segregation of transient data stores with concomitant object...
Ruben Domingo Gaspar Aparicio
(CERN)
4/13/15, 5:00 PM
CERN IT-DB group is migrating its storage platform, mainly NetApp NAS’s running on 7-mode but also SAN arrays, to a set of NetApp C-mode clusters. The largest one is made of 14 controllers and it will hold a range of critical databases from administration to accelerators control or experiment control databases. This talk shows our setup: network, monitoring, use of features like transparent...
Prof.
Daniele Bonacorsi
(University of Bologna)
4/13/15, 5:00 PM
During the LHC Run-1 data taking, all experiments collected large data volumes from proton-proton and heavy-ion collisions. The collisions data, together with massive volumes of simulated data, were replicated in multiple copies, transferred among various Tier levels, transformed/slimmed in format/content. These data were then accessed (both locally and remotely) by large groups of distributed...
Richard Calland
4/13/15, 5:00 PM
Track8: Performance increase and optimization exploiting hardware features
oral presentation
The Tokai-to-Kamioka (T2K) experiment is a second generation long baseline neutrino experiment, which uses a near detector to constrain systematic uncertainties for oscillation measurements with its far detector. Event-by-event reweighting of Monte Carlo (MC) events is applied to model systematic effects and construct PDFs describing predicted event distributions. However when analysing...
Federico Stagni
(CERN)
4/13/15, 5:00 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
The DIRAC workload management system used by LHCb Distributed Computing is based on Computing Resource reservation and late binding (also known as pilot job in the case of batch resources) that allows the serial execution of several jobs obtained from a central task queue. CPU resources can usually be reserved for limited duration only (e.g. batch queue time limit) and in order to optimize...
Andrew McNab
(University of Manchester (GB))
4/13/15, 5:00 PM
We compare the Vac and Vcycle virtual machine lifecycle managers and our experiences in providing production job execution services for ATLAS, LHCb, and the GridPP VO at sites in the UK and at CERN. In both the Vac and Vcycle systems, the virtual machines are created outside of the experiment's job submission and pilot framework. In the case of Vac, a daemon runs on each physical host which...
Elizabeth Sexton-Kennedy
(Fermi National Accelerator Lab. (US))
4/13/15, 5:15 PM
Today there are many different experimental event processing frameworks in use by running or about to be running experiments. This talk will compare and contrast the different components of these frameworks and highlight the different solutions chosen by different groups. In the past there have been attempts at shared framework projects for example the collaborations on the BaBar framework...
Andrew John Washbrook
(University of Edinburgh (GB))
4/13/15, 5:15 PM
Cloud computing enables ubiquitous, convenient and on-demand access to a shared pool of configurable computing resources that can be rapidly provisioned with minimal management effort. The flexible and scalable nature of the cloud computing model is attractive to both industry and academia. In HEP, the use of the “cloud” has become more prevalent with LHC experiments making use of standard...
David Michael Rohr
(Johann-Wolfgang-Goethe Univ. (DE))
4/13/15, 5:15 PM
Track8: Performance increase and optimization exploiting hardware features
oral presentation
ALICE (A Large Heavy Ion Experiment) is one of the four major experiments at the Large Hadron Collider (LHC) at CERN, which is today the most powerful particle accelerator worldwide. The High Level Trigger (HLT) is an online compute farm of about 200 nodes, which reconstructs events measured by the ALICE detector in real-time. The HLT uses a custom online data-transport framework to distribute...
Jeffrey Michael Dost
(Univ. of California San Diego (US))
4/13/15, 5:15 PM
In April of 2014, the UCSD T2 Center deployed hdfs-xrootd-fallback, a UCSD-developed software system that interfaces Hadoop with XRootD to increase reliability of the Hadoop file system. The hdfs-xrootd-fallback system allows a site to depend less on local file replication and more on global replication provided by the XRootD federation to ensure data redundancy. Deploying the software has...
Dr
Torre Wenaus
(Brookhaven National Laboratory (US))
4/13/15, 5:15 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
The ATLAS Event Service (ES) implements a new fine grained approach to HEP event processing, designed to be agile and efficient in exploiting transient, short-lived resources such as HPC hole-filling, spot market commercial clouds, and volunteer computing. Input and output control and data flows, bookkeeping, monitoring, and data storage are all managed at the event level in an implementation...
Marco Mascheroni
(Universita & INFN, Milano-Bicocca (IT))
4/13/15, 5:30 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
The CMS Remote Analysis Builder (CRAB) provides the service for managing analysis tasks isolating users from the technical details of the distributed Grid infrastructure. Throughout the LHC Run 1, CRAB has been successfully employed by an average 350 distinct users every week executing about 200,000 jobs per day.
In order to face the new challenges posed by the LHC Run 2, CRAB has been...
Philippe Canal
(Fermi National Accelerator Lab. (US))
4/13/15, 5:30 PM
Track8: Performance increase and optimization exploiting hardware features
oral presentation
The recent prevalence of hardware architectures of many-core or accelerated
processors opens opportunities for concurrent programming models taking
advantages of both SIMD and SIMT architectures. The Geant Vector Prototype
has been designed both to exploit the vector capability of main stream
CPUs and to take advantage of Coprocessors including NVidia’s GPU and Intel
Xeon Phi. The...
Dr
Carl Vuosalo
(University of Wisconsin (US))
4/13/15, 5:30 PM
The CMS experiment has developed a new analysis object format (the "mini-AOD") targeted to be less than 10% of the size of the Run 1 AOD format. The motivation for the Mini-AOD format is to have a small and quickly derived data format from which the majority of CMS analysis users can perform their analysis work. This format is targeted at having sufficient information to serve about 80% of CMS...
Dr
Bodhitha Jayatilaka
(Fermilab)
4/13/15, 5:30 PM
The Open Science Grid (OSG) ties together individual experiments' computing power, connecting their resources to create a large, robust computing grid; this computing infrastructure started primarily as a collection of sites associated with large HEP experiments such as ATLAS, CDF, CMS, and DZero. OSG has been funded by the Department of Energy Office of Science and National Science Foundation...
Sebastian Neubert
(CERN)
4/13/15, 5:45 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
Reproducibility of results is a fundamental quality of scientific research. However, as data analyses become more and more complex and research is increasingly carried out by larger and larger teams, it becomes a challenge to keep up this standard. The decomposition of complex problems into tasks that can be effectively distributed over a team in a reproducible manner becomes...
Federica Legger
(Ludwig-Maximilians-Univ. Muenchen (DE))
4/13/15, 5:45 PM
The ATLAS experiment accumulated more than 140 PB of data during the first run of the Large Hadron Collider (LHC) at CERN. The analysis of such an amount of data for the distributed physics community is a challenging task. The Distributed Analysis (DA) system of the ATLAS experiment is an established and stable component of the ATLAS distributed computing operations. About half a million user...
Michele Martinelli
(INFN Rome)
4/13/15, 5:45 PM
Track8: Performance increase and optimization exploiting hardware features
oral presentation
The computing nodes of modern hybrid HPC systems are built using the CPU+GPU paradigm.
When this class of systems is scaled to large size, the efficiency of the network connecting GPUs mesh and supporting the internode traffic is a critical factor. The adoption of a low latency, high performance dedicated network architecture, exploiting peculiar characteristics of CPU and GPU hardware,...
Sara Vallero
(Universita e INFN (IT))
4/13/15, 5:45 PM
The INFN computing centre in Torino hosts a private Cloud, which is managed with the OpenNebula cloud controller. The infrastructure offers IaaS services to different scientific computing applications. The main stakeholders of the facility are a grid Tier-2 site for the ALICE collaboration at LHC, an interactive analysis facility for the same experiment and a separate grid Tier-2 site for the...
Janusz Martyniak
(Imperial College London)
4/13/15, 5:45 PM
The Muon Ionization Cooling Experiment (MICE) has developed the MICE Analysis User Software (MAUS) to simulate and analyse experimental data. It serves as the primary codebase for the experiment, providing for offline batch simulation and reconstruction as well as online data quality checks . The software provides both traditional particle physics functionalities such as track reconstruction...
Johannes Elmsheuser
(Ludwig-Maximilians-Univ. Muenchen (DE))
4/13/15, 5:45 PM
With the exponential growth of LHC (Large Hadron Collider) data in the years 2010-2012, distributed computing has become the established way to analyze collider data. The ATLAS experiment Grid infrastructure includes more than 130 sites worldwide, ranging from large national computing centres to smaller university clusters. So far the storage technologies and access protocols to the clusters...
Thomas Maier
(Ludwig-Maximilians-Univ. Muenchen (DE))
4/13/15, 6:00 PM
I/O is a fundamental determinant in the overall performance of physics analysis and other data-intensive scientific computing. It is, further, crucial to effective resource delivery by the facilities and infrastructure that support data-intensive science. To understand I/O performance, clean measurements in controlled environments are essential, but effective optimization requires as well an...
Andrew David Lahiff
(STFC - Rutherford Appleton Lab. (GB))
4/13/15, 6:00 PM
The recently introduced vacuum model offers an alternative to the traditional methods that virtual organisations (VOs) use to run computing tasks at sites, where they either submit jobs using grid middleware or create virtual machines (VMs) using cloud APIs. In the vacuum model VMs are created and contextualized by the site itself, and start the appropriate pilot job framework which fetches...
Dr
Tian Yan
(Institution of High Energy Physics, Chinese Academy of Science)
4/13/15, 6:00 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
For Beijing Spectrometer III (BESIII) experiment located at the Institute of High Energy Physics (IHEP), China, the distributed computing environment (DCE) has been setup and been in production status since 2012. The basic framework or middleware is DIRAC (Distributed Infrastructure with Remote Agent Control) with BES-DIRAC extensions. About 2000 CPU cores and 400 TB storage contributed by...
Mr
Steffen Baehr
(Karlsruhe Institute of Technology)
4/13/15, 6:00 PM
Track8: Performance increase and optimization exploiting hardware features
oral presentation
The impending Upgrade of the Belle experiment is expected to increase the generated data set by a factor of 50.
This means that for the planned pixeldetector, which is the closest to the interaction point, the data rates are going to increase to over 20 GB/s.
Combined with data generated by the other detectors, this rate is too big to be efficiently send out to offline processing.
This is...
Dirk Hufnagel
(Fermi National Accelerator Lab. (US))
4/13/15, 6:00 PM
In 2015, CMS will embark on a new era of collecting LHC collisions at unprecedented rates and complexity. This will put a tremendous stress on our computing systems. Prompt Processing of the raw data by the Tier-0 infrastructure will no longer be constrained to CERN alone due to the significantly increased resource requirements. In LHC Run 2, we will need to operate it as a distributed system...
Adam Aurisano
(University of Cincinnati)
4/13/15, 6:00 PM
The NOvA experiment is a two-detector, long-baseline neutrino experiment operating in the recently upgraded NuMI muon neutrino beam. Simulating neutrino interactions and backgrounds requires many steps including: the simulation of the neutrino beam flux using FLUKA and the FLUGG interface; cosmic ray generation using CRY; neutrino interaction modeling using GENIE; and a simulation of the...
Srikanth Sridharan
(CERN)
4/13/15, 6:15 PM
Track8: Performance increase and optimization exploiting hardware features
oral presentation
The proposed upgrade for the Large Hadron Collider LHCb experiment at CERN envisages a system of 500 Data sources each generating data at 100 Gbps, the acquisition and processing of which is a challenge even for state of the art FPGAs. This challenge splits into two, the Data Acquisition (DAQ) part and the Algorithm acceleration part, the later not necessarily immediately following the former....
Janusz Martyniak
(Imperial College London)
4/13/15, 6:15 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
The GridPP consortium in the UK is currently testing a multi-VO DIRAC service aimed at non-LHC VOs. These VOs are typically small (fewer than two hundred members) and generally do not have a dedicated computing support post. The majority of these represent particle physics experiments (e.g. T2K, NA62 and COMET), although the scope of the DIRAC service is not limited to this field. A few VOs...
Norman Anthony Graf
(SLAC National Accelerator Laboratory (US))
4/13/15, 6:15 PM
The Heavy Photon Search (HPS) is an experiment at the Thomas Jefferson National Accelerator Facility (JLab) designed to search for a hidden sector photon (A’) in fixed target electroproduction. It uses a silicon microstrip tracking and vertexing detector inside a dipole magnet to measure charged particle trajectories and a fast electromagnetic calorimeter just downstream of the magnet to...
Dr
Jonathan Dorfan
(OIST)
4/14/15, 9:00 AM
Robert Group
(University of Virginia)
4/14/15, 9:15 AM
4/14/15, 10:00 AM
Amber Boehnlein
4/14/15, 11:00 AM
Edgar Fajardo Hernandez
(Univ. of California San Diego (US))
4/14/15, 2:00 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
The HTCondor-CE is the next-generation gateway software for the Open Science Grid (OSG). This is responsible for providing a network service which authorizes remote users and provides a resource provisioning service (other well-known gatekeepers include Globus GRAM, CREAM, Arc-CE, and Openstack’s Nova). Based on the venerable HTCondor software, this new CE is simply a highly-specialized...
Jeremy Coles
(University of Cambridge (GB))
4/14/15, 2:00 PM
This first section of this paper elaborates upon the operational status and directions within the UK Computing for Particle Physics (GridPP) project as we approach LHC Run-2. It details the pressures that have been gradually reshaping the deployed hardware and middleware environments at GridPP sites – from the increasing adoption of larger multicore nodes to the move towards alternative batch...
Vakho Tsulaia
(Lawrence Berkeley National Lab. (US))
4/14/15, 2:00 PM
AthenaMP is a multi-process version of the ATLAS reconstruction and data analysis framework Athena. By leveraging Linux fork and copy-on-write, it allows the sharing of memory pages between event processors running on the same compute node with little to no change in the application code. Originally targeted to optimize the memory footprint of reconstruction jobs, AthenaMP has demonstrated...
James Catmore
(University of Oslo (NO))
4/14/15, 2:15 PM
During the Long shutdown of the LHC, the ATLAS collaboration overhauled its analysis model based on experience gained during Run 1. A significant component of the model is a "Derivation Framework" that takes the Petabyte-scale AOD output from ATLAS reconstruction and produces samples, typically Terabytes in size, targeted at specific analyses. The framework incorporates all of the...
Christopher Jon Lee
(University of Johannesburg (ZA))
4/14/15, 2:15 PM
The ATLAS Trigger and Data Acquisition (TDAQ) system is responsible for the online processing of live data, streaming from the ATLAS experiment at the Large Hadron Collider (LHC) at CERN. The online farm is composed of ~3000 servers, processing the data readout from ~100 million detector channels through multiple trigger levels.
During the two years of the first Long Shutdown (LS1) there has...
David Cameron
(University of Oslo (NO))
4/14/15, 2:15 PM
A recent common theme among HEP computing is exploitation of opportunistic resources in order to provide the maximum statistics possible for Monte-Carlo simulation. Volunteer computing has been used over the last few years in many other scientific fields and by CERN itself to run simulations of the LHC beams. The ATLAS@Home project was started to allow volunteers to run simulations of...
Andrej Filipcic
(Jozef Stefan Institute (SI))
4/14/15, 2:15 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
Distributed computing resources available for high-energy physics research are becoming less dedicated to one type of workflow and researchers’ workloads are increasingly exploiting modern computing technologies such as parallelism. The current pilot job management model used by many experiments relies on static dedicated resources and cannot easily adapt to these changes. The model used for...
Jose Flix Molina
(Centro de Investigaciones Energ. Medioambientales y Tecn. - (ES)
4/14/15, 2:15 PM
The LHC experiments will collect unprecedented data volumes in the next Physics run, with high pile-up collisions resulting in events which require a more complex processing. The collaborations have been asked to update their Computing Models to optimize the use of the available resources in order to cope with the Run2 conditions, in the midst of widespread funding restrictions. The changes in...
Martin Urban
(Rheinisch-Westfaelische Tech. Hoch. (DE))
4/14/15, 2:15 PM
VISPA provides a graphical front-end to computing infrastructures giving its users all functionality needed for working conditions comparable to a personal computer. It is a framework that can be extended with custom applications to support individual needs, e.g. graphical interfaces for experiment-specific software. By design, VISPA serves as a multi-purpose platform for many disciplines and...
Jon Kerr Nilsen
(University of Oslo (NO))
4/14/15, 2:30 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
While current grid middlewares are quite advanced in terms of connecting jobs to resources, their client tools are generally quite minimal and features for managing large sets of jobs are left to the user to implement. The ARC Control Tower (aCT) is a very flexible job management framework that can be run on anything from a single user’s laptop to a multi-server distributed setup. aCT was...
Dr
Bodhitha Jayatilaka
(Fermilab)
4/14/15, 2:30 PM
The Fermilab Tevatron collider's data-taking run ended in September 2011, yielding a dataset with rich scientific potential. The CDF and D0 experiments each have nearly 9 PB of collider and simulated data stored on tape. A large computing infrastructure consisting of tape storage, disk cache, and distributed grid computing for physics analysis with the Tevatron data is present at...
Dr
Sami Kama
(Southern Methodist University (US))
4/14/15, 2:30 PM
The challenge faced by HEP experiments from the current and expected architectural evolution of CPUs and co-processors is how to successfully exploit concurrency and keep memory consumption within reasonable limits. This is a major change from frameworks which were designed for serial event processing on single core processors in the 2000s. ATLAS has recently considered this problem in some...
Jeff Templon
(NIKHEF (NL))
4/14/15, 2:30 PM
With the advent of workloads containing explicit requests for multiple
cores in a single grid job, grid sites faced a new set of challenges
in workload scheduling. The most common batch schedulers deployed at
HEP computing sites do a poor job at multicore scheduling when using
only the native capabilities of those schedulers. This talk describes
how efficient multicore scheduling was...
Reiner Hauser
(Michigan State University (US))
4/14/15, 2:30 PM
After its first shutdown, LHC will provide pp collisions with increased luminosity and energy. In the ATLAS experiment the Trigger and Data Acquisition (TDAQ) system has been upgraded to deal with the increased event rates. The Data Flow (DF) element of the TDAQ is a distributed hardware and software system responsible for buffering and transporting event data from the Readout system to the...
Manuel Giffels
(KIT - Karlsruhe Institute of Technology (DE))
4/14/15, 2:45 PM
Recent developments in high energy physics (HEP) including multi-core jobs and multi-core pilots require data centres to gain a deep understanding of the system to correctly design and upgrade computing clusters. Especially networking is a critical component as the increased usage of data federations relies on WAN connectivity and availability as a fallback to access data. The specific...
Roger Jones
(Lancaster University (GB))
4/14/15, 2:45 PM
Complementary to parallel open access and analysis preservation initiatives, ATLAS is taking steps to ensure that the data taken by the experiment during run-1 remain accessible and available for future analysis by the collaboration. An evaluation of what is required to achieve this is underway, examining the ATLAS data production chain to establish the effort required and potential problems....
Charles Leggett
(Lawrence Berkeley National Lab. (US))
4/14/15, 2:45 PM
The ATLAS experiment has successfully used its Gaudi/Athena software framework for data taking and analysis during the first LHC run, with billions of events successfully processed. However, the design of Gaudi/Athena dates from early 2000 and the software and the physics code has been written using a single threaded, serial design. This programming model has increasing difficulty in...
Mr
Thomas Hauth
(KIT - Karlsruhe Institute of Technology (DE))
4/14/15, 2:45 PM
Modern high-energy physics experiments rely on the extensive usage of computing resources,
both for the reconstruction of measured events as well as for Monte Carlo simulation. The
Institut für Experimentelle Kernphysik (EKP) at KIT is participating in both the CMS and Belle
experiments with computing and storage resources. In the upcoming years, these requirements
are expected to...
Jorn Schumacher
(University of Paderborn (DE))
4/14/15, 2:45 PM
The ATLAS experiment at CERN is planning the full deployment of a new, unified link technology for connecting detector front-end electronics on the timescale of the LHC Run 4 (2025). It is estimated that roughly 8000 Gigabit Transceiver links (GBT), with transfer rates probably up to 9.6 Gbps, will replace existing links used for readout, detector control and distribution of timing and trigger...
Andres Gomez Ramirez
(Johann-Wolfgang-Goethe Univ. (DE))
4/14/15, 2:45 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
Grid infrastructures allow users flexible on-demand usage of computing resources using an Internet connection. A remarkable example of a Grid in High Energy Physics (HEP) research is used by the ALICE experiment at European Organization for Nuclear Research CERN. Physicists can submit jobs used to process the huge amount of particle collision data produced by the Large Hadron Collider (LHC) at...
Remi Mommsen
(Fermi National Accelerator Lab. (US))
4/14/15, 3:00 PM
The data acquisition system (DAQ) of the CMS experiment at the CERN Large Hadron Collider (LHC) assembles events at a rate of 100 kHz, transporting event data at an aggregate throughput of 100 GB/s to the high-level trigger (HLT) farm. The DAQ system has been redesigned during the LHC shutdown in 2013/14. The new DAQ architecture is based on state-of-the-art network technologies for the event...
Jetendr Shamdasani
(University of the West of England (GB))
4/14/15, 3:00 PM
In complex data analyses it is increasingly important to capture information about the usage of data sets in addition to their preservation over time in order to ensure reproducibility of results, to verify the work of others and to ensure appropriate conditions data have been used for specific analyses. This so-called provenance data in the computer science world is defined as the history or...
Dr
Christopher Jones
(Fermi National Accelerator Lab. (US))
4/14/15, 3:00 PM
During 2014, the CMS Offline and Computing Organization completed the necessary changes to use the CMS threaded framework in the full production environment. Running reconstruction workflows using the multi-threaded framework is a crucial element of CMS' 2015 and beyond production plan. We will briefly discuss the design of the CMS Threaded Framework, in particular how the design affects...
Dr
Tony Wildish
(Princeton)
4/14/15, 3:00 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
The ANSE project has been working with the CMS and ATLAS experiments to bring network awareness into their middleware stacks. For CMS, this means enabling control of virtual network circuits in PhEDEx, the CMS data-transfer management system. PhEDEx orchestrates the transfer of data around the CMS experiment to the tune of 1 PB per week spread over about 70 sites.
The goal of ANSE is to...
Dr
Alexei Klimentov
(Brookhaven National Laboratory (US))
4/14/15, 3:15 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
A crucial contributor to the success of the massively scaled global computing system that delivers the analysis needs of the LHC experiments is the networking infrastructure upon which the system is built. The experiments have been able to exploit excellent high-bandwidth networking in adapting their computing models for the most efficient utilization of resources.
New advanced networking...
Dr
Andrew Norman
(Fermilab)
4/14/15, 3:15 PM
The ability of modern HEP experiments to acquire and process unprecedented amounts of data and simulation have led to an explosion in the volume of information that individual scientists deal with on a daily basis. This explosion has resulted in a need for individuals to generate and keep large “personal analysis” data sets which represent the skimmed portions of official data collections...
Paul Millar
(Deutsches Elektronen-Synchrotron (DE))
4/14/15, 3:15 PM
Traditionally storage systems have had well understood
responsibilities and behaviour, codified by the POSIX standards. More
sophisticated systems (such as dCache) support additional
functionality, such as storing data on media with different latencies
(SSDs, HDDs, tapes). From a user's perspective, this forms a
relatively simple adjunct to POSIX: providing optional
quality-of-service...
Dr
Florian Uhlig
(GSI Darmstadt)
4/14/15, 3:15 PM
The FairRoot framework is the standard framework for simulation, reconstruction and data analysis developed at GSI for the future experiments at the FAIR facility.
The framework delivers base functionality for simulation, i.e.: Infrastructure to easily implement a set of detectors, fields, and event generators. Moreover, the framework decouples the user code (e.g.: Geometry description,...
Renaud Vernet
(CC-IN2P3 - Centre de Calcul (FR))
4/14/15, 3:30 PM
The computing needs in the HEP community are increasing steadily, but the current funding situation in many countries is tight. As a consequence experiments, data centres, and funding agencies have to rationalize resource usage and expenditures.
CC-IN2P3 (Lyon, France) provides computing resources to many experiments including LHC, and is a major partner for astroparticle projects like...
Dr
Mohammad Al-Turany
(CERN)
4/14/15, 3:30 PM
The commonalities between the ALICE and FAIR experiments and their computing requirements lead to the development of large parts of a common software framework in an experiment independent way. The FairRoot project has already shown the feasibility of such an approach for the FAIR experiments and extending it beyond FAIR to experiments at other facilities. The ALFA framework is a joint...
Dirk Hufnagel
(Fermi National Accelerator Lab. (US))
4/14/15, 3:30 PM
With the increased pressure on computing brought by the higher energy and luminosity from the LHC in Run 2, CMS Computing Operations expects to require the ability to utilize “opportunistic” resources — resources not owned by, or a priori configured for CMS — to meet peak demands. In addition to our dedicated resources we look to add computing resources from non CMS grids, cloud resources, and...
Alessandra Forti
(University of Manchester (GB))
4/14/15, 3:30 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
After the successful first run of the LHC, data taking will restart in early 2015 with unprecedented experimental conditions leading to increased data volumes and event complexity. In order to process the data generated in such scenario and exploit the multicore architectures of current CPUs, the LHC experiments have developed parallelized software for data reconstruction and simulation. A...
Georgiana Lavinia Darlea
(Massachusetts Inst. of Technology (US))
4/14/15, 3:30 PM
The CMS experiment at CERN is one of the two general-purpose detectors on the Large Hadron Collider (LHC) in the Geneva area, Switzerland. Its infrastructure has undergone massive upgrades during 2013 and 2014, which lead to major changes in the philosophy of its DAQ (Data AcQuisition) system. One of the major components of this system is the Storage Manager, which is responsible for buffering...
Dr
Wenji Wu
(Fermi National Accelerator Laboratory)
4/14/15, 3:45 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
Multicore and manycore have become the norm for scientific computing environments. Multicore/manycore platform architectures provide advanced capabilities and features that can be exploited to enhance data movement performance for large-scale distributed computing environments, such as LHC. However, existing data movement tools do not take full advantage of these capabilities and features....
marko staric
(J. Stefan Institute, Ljubljana, Slovenia)
4/14/15, 3:45 PM
We present software framework being developed for physics analyses using
the data collected by the Belle II experiment. The analysis workflow is
organized in a modular way integrated within the Belle II software framework
(BASF2). A set
of physics analysis modules that perform simple and well defined tasks
and are common to almost all physics analyses are provided. The physics
modules do...
Matthias Richter
(University of Oslo (NO))
4/14/15, 4:30 PM
An upgrade of the ALICE detector is currently prepared for the Run 3 period of the Large Hadron
Collider (LHC) at CERN starting in 2020. The physics topics under study by ALICE during this
period will require the inspection of all collisions at a rate of 50 kHz for minimum bias Pb-Pb and 200
kHz for pp and p-Pb collisions in order to extract physics signals embedded into a large...
Dr
Makoto Asai
(SLAC National Accelerator Laboratory (US))
4/14/15, 4:30 PM
The Geant4 Collaboration released a new generation of the Geant4
simulation toolkit (version 10.0) in December 2013, and continues
to improve its physics, computing performance and usability. This
presentation will cover the major improvements made since version
10.0. The physics evolutions include improvement of the Fritiof
hadronics model, extension of the INCL++ model to higher...
Ian Gable
(University of Victoria (CA))
4/14/15, 4:30 PM
The use of distributed IaaS clouds with the CloudScheduler/HTCondor architecture has been in production for HEP and astronomy applications for a number of years. The design has proven to be robust and reliable for batch production using HEP clouds, academic non-HEP (opportunistic) clouds and commercial clouds. Further, the system is seamlessly integrated into the existing WLCG...
2.
Mean PB to Failure -- Initial results from a long-term study of disk storage patterns at the RACF
Christopher Hollowell
(Brookhaven National Laboratory)
4/14/15, 4:30 PM
The RACF (RHIC-ATLAS Computing Facility) has operated a large, multi-purpose dedicated computing facility since the mid-1990's, serving a worldwide, geographically diverse scientific community that is a major contributor to various HEPN projects. A central component of the RACF is the Linux-based worker node cluster that is used for both computing and data storage purposes. It currently has...
Rainer Schwemmer
(CERN)
4/14/15, 4:30 PM
Track8: Performance increase and optimization exploiting hardware features
oral presentation
For Run 2 of the LHC, LHCb is exchanging a significant part of its event filter farm with new compute nodes. For the evaluation of the best performing solution, we have developed a method to convert our high level trigger application into a stand-alone, bootable benchmark image. With additional instrumentation we turned it into a self-optimising benchmark which explores techniques such as late...
Mr
Jason Alexander Smith
(Brookhaven National Laboratory)
4/14/15, 4:30 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
Using centralized configuration management, including automation tools such as
Puppet, can greatly increase provisioning speed and efficiency when configuring
new systems or making changes to existing systems, reduce duplication of work,
and improve automated processes. However, centralized management also brings
with it a level of inherent risk: a single change in just one file can...
Alexey Rybalchenko
(GSI - Helmholtzzentrum fur Schwerionenforschung GmbH (DE))
4/14/15, 4:45 PM
After Long Shutdown 2, the upgraded ALICE detector at the LHC will produce more than a terabyte of data per second. The data, constituted from a continuous un-triggered stream data, have to be distributed from about 250 First Level Processor nodes (FLPs) to O(1000) Event Processing Nodes (EPNs). Each FLP receives a small subset of the detector data that is chopped in sub-timeframes. One EPN...
Ivana Hrivnacova
(IPNO, Université Paris-Sud, CNRS/IN2P3)
4/14/15, 4:45 PM
Virtual Monte Carlo (VMC) provides an abstract interface into Monte Carlo transport codes. A user VMC based application, independent from the specific Monte Carlo codes, can be then run with any of the supported simulation programs. Developed by the ALICE Offline Project and further included in ROOT, the interface and implementations have reached stability during the last decade and have...
Daniel Hugo Campora Perez
(CERN)
4/14/15, 4:45 PM
Track8: Performance increase and optimization exploiting hardware features
oral presentation
During the data taking process in the LHC at CERN, millions of collisions are recorded every second by the LHCb Detector. The LHCb "Online" computing farm, counting around 15000 cores, is dedicated to the recontruction of the events in real-time, in order to filter those with interesting Physics. The ones kept are later analysed "Offline" in a more precise fashion on the Grid. This imposes...
Alessandro De Salvo
(Universita e INFN, Roma I (IT))
4/14/15, 4:45 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
The ATLAS Installation System v2 is the evolution of the original system, used since 2003. The original tool has been completely re-designed in terms of database backend and components, adding support for submission to multiple backends, including the original WMS and the new Panda modules. The database engine has been changed from plain MySQL to Galera/Percona and the table structure has been...
Dr
David Colling
(Imperial College Sci., Tech. & Med. (GB))
4/14/15, 4:45 PM
The resources CMS is using are increasingly being offered as clouds. In Run 2 of the LHC the majority of CMS CERN resources, both in Meyrin and at the Wigner Computing Centre, will be presented as cloud resources on which CMS will have to build its own infrastructure. This infrastructure will need to run all of the CMS workflows including: Tier 0, production and user analysis. In addition, the...
Ms
Bowen Kan
(Institute of High Physics Chinese Academy of Sciences)
4/14/15, 5:00 PM
Track8: Performance increase and optimization exploiting hardware features
oral presentation
Scheduler is one of the most important components of high performance cluster. This paper introduces a self-adaptive dispatching system (SAPS) based on torque/maui which increases the resources utilization of cluster effectively and guarantees the high reliability of the computing platform. It provides great convenience for users to run various tasks on the computing platform. First of all,...
Dr
Giuseppe Avolio
(CERN)
4/14/15, 5:00 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
Complex Event Processing (CEP) is a methodology that combines data from different sources in order to identify events or patterns that need particular attention. It has gained a lot of momentum in the computing world in the past few years and is used in ATLAS to continuously monitor the behaviour of the data acquisition system, to trigger corrective actions and to guide the experiment’s...
Dr
Andrew Norman
(Fermilab)
4/14/15, 5:00 PM
Many experiments in the HEP and Astrophysics communities generate large extremely valuable datasets, which need to be efficiently cataloged and recorded to archival storage. These datasets, both new and legacy, are often structured in a manner that is not conducive to storage and cataloging with modern data handling systems and large file archive facilities. In this paper we discuss in...
Ryan Taylor
(University of Victoria (CA))
4/14/15, 5:00 PM
The ATLAS experiment has successfully incorporated cloud computing technology and cloud resources into its primarily grid-based model of distributed computing. Cloud R&D activities continue to mature and transition into stable production systems, while ongoing evolutionary changes are still needed to adapt and refine the approaches used, in response to changes in prevailing cloud technology....
Josef Novy
(Czech Technical University (CZ))
4/14/15, 5:00 PM
This contribution focuses on the deployment and first results of the new data acquisition system (DAQ) of the COMPASS experiment utilizing FPGA-based event builder. The new DAQ system is developed under name RCCARS (run control, configuration, and readout system).
COMPASS is a high energy physics experiment situated at the SPS particle accelerator at CERN laboratory in Geneva, Switzerland....
Norman Anthony Graf
(SLAC National Accelerator Laboratory (US))
4/14/15, 5:00 PM
As the complexity and resolution of particle detectors increases, the need for detailed simulation of the experimental setup also increases. We have developed efficient and flexible tools for detailed physics and detector response simulations which build on the power of the Geant4 toolkit but free the end user from any C++ coding. Geant4 is the de facto high-energy physics standard for...
Mr
Giulio Eulisse
(Fermi National Accelerator Lab. (US))
4/14/15, 5:15 PM
Track8: Performance increase and optimization exploiting hardware features
oral presentation
Power consumption will be a key constraint on the future growth of Distributed High Throughput Computing (DHTC) as used by High Energy Physics (HEP). This makes performance-per-watt a crucial metric for selecting cost-efficient computing solutions. For this paper, we have done a wide survey of current and emerging architectures becoming available on the market including x86-64 variants, ARMv7...
Katsuki Hiraide
(the University of Tokyo)
4/14/15, 5:15 PM
XMASS is a multi-purpose low-background experiment with a large volume of liquid xenon scintillator at Kamioka in Japan. The first phase of the experiment aiming at direct detection of dark matter was commissioned in 2010 and is currently taking data.
The detector uses ~830 kg of liquid xenon viewed by 642 photomultiplier tubes (PMTs). Signals from 642 PMTs are amplified and read out by 1...
Mr
Federico Carminati
(CERN)
4/14/15, 5:15 PM
Detector simulation is consuming at least half of the HEP computing cycles, and even so, experiments have to take hard decisions on what to simulate, as their needs greatly surpass the availability of computing resources. New experiments still in the design phase such as FCC, CLIC and ILC as well as upgraded versions of the existing LHC detectors will push further the simulation requirements....
Dr
Randy Sobie
(University of Victoria (CA))
4/14/15, 5:15 PM
The BelleII experiment is developing a global computing system for the simulation of MC data prior its collecting real collision data in the next few years. The system utilizes the grid middleware used in the WLCG and uses the DIRAC workload manager. We describe how IaaS cloud resources are being integrated into the BelleII production computing system in Australia and Canada. The IaaS...
Mr
Tigran Mkrtchyan
(DESY)
4/14/15, 5:15 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
Over the previous years, storage providers in scientific infrastructures were facing a significant change in the usage profile of their resources. While in the past, a small number of experiment frameworks were accessing those resources in a coherent manner, now, a large amount of small groups or even individuals request access in a completely chaotic way. Moreover, scientific laboratories...
David Lange
(Lawrence Livermore Nat. Laboratory (US))
4/14/15, 5:30 PM
This presentation will discuss new features of the CMS simulation for Run 2, where we have made considerable improvements during LHC shutdown to deal with the increased event complexity and rate for Run 2. For physics improvements migration from Geant4 9.4p03 to Geant4 10.0p02 has been performed. CPU performance was improved by introduction of the Russian roulette method inside CMS...
Asato Orii
(urn:Facebook)
4/14/15, 5:30 PM
Super-Kamiokande (SK), a 50-kiloton water Cherenkov detector, is one of
the most sensitive neutrino detectors. SK is continuously collecting
data as the neutrino observatory and can be used also for supernova
observations by detecting supernova burst neutrinos.
It is reported that Betelgeuse (640ly) is shrinking 15% in 15 years
(C.H.townes et al. 2009) and this may be an...
Peter Onyisi
(University of Texas (US))
4/14/15, 5:30 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
During LHC Run 1, the information flow through the offline data quality monitoring in ATLAS relied heavily on chains of processes polling each other's outputs for handshaking purposes. This resulted in a fragile architecture with many possible points of failure and an inability to monitor the overall state of the distributed system. We report on the status of a project undertaken during the...
Jakob Blomer
(CERN)
4/14/15, 5:30 PM
Track8: Performance increase and optimization exploiting hardware features
oral presentation
Most high-energy physics analysis jobs are embarrassingly parallel except for the final merging of the output objects, which are typically histograms. Currently, the merging of output histograms scales badly. The running time for distributed merging depends not only on the overall number of bins but
also on the number partial histogram output files. That means, while the time to analyze data...
Andrew McNab
(University of Manchester (GB))
4/14/15, 5:30 PM
The LHCb experiment has been running production jobs in virtual machines since 2013 as part of its DIRAC-based infrastructure. We describe the architecture of these virtual machines and the steps taken to replicate the WLCG worker node environment expected by user and production jobs. This relies on the CernVM 3 system for providing root images for virtual machines. We use the cvmfs...
Bruno Lange Ramos
(Univ. Federal do Rio de Janeiro (BR)),
Bruno Lange Ramos
(Univ. Federal do Rio de Janeiro (BR))
4/14/15, 5:45 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
In order to manage a heterogeneous and worldwide collaboration, the ATLAS experiment developed web systems that range from supporting the process of publishing scientific papers to monitoring equipment radiation levels. These systems are vastly supported by Glance, a technology that was set forward in 2004 to create an abstraction layer on top of different databases; it automatically...
Ms
Bowen Kan
(Institute of High Physics Chinese Academy of Sciences)
4/14/15, 5:45 PM
Mass data processing and analysis contribute much to the development and discoveries of a new generation of High Energy Physics. The BESIII experiment of IHEP(Institute of High Energy Physics, Beijing, China) studies particles in the tau-charm energy region ranges from 2 GeV to 4.6 GeV, and requires massive storage and computing resources, which is a typical kind of data intensive...
Gaelle Boudoul
(Universite Claude Bernard-Lyon I (FR))
4/14/15, 5:45 PM
CMS Detector Description (DD) is an integral part of the CMSSW software multithreaded framework. CMS software has evolved to be more flexible and to take advantage of new techniques, but many of the original concepts remain and are in active use. In this presentation we will discuss the limitations of the Run I DD model and changes implemented for the restart of the LHC program in 2015....
Max Fischer
(KIT - Karlsruhe Institute of Technology (DE))
4/14/15, 5:45 PM
Track8: Performance increase and optimization exploiting hardware features
oral presentation
With the second run period of the LHC, high energy physics collaborations will have to face increasing computing infrastructural needs.
Opportunistic resources are expected to absorb many computationally expensive tasks, such as Monte Carlo event simulation.
This leaves dedicated HEP infrastructure with an increased load of analysis tasks that in turn will need to process an increased volume...
260.
The Electronics, Online Trigger System and Data Acquisition System of the J-PARC E16 Experiment
Tomonori Takahashi
(Research Center for Nuclear Physics, Osaka University)
4/14/15, 5:45 PM
1. Introduction
The J-PARC E16 experiment aims to investigate the chiral symmetry restoration in cold nuclear matter and the origin of the hadron mass through the systematic study of the mass modification of vector mesons.
In the experiment,
$e^{+}e^{-}$ decay of slowly-moving $\phi$ mesons in the normal nuclear matter density are intensively studied using several nuclear targets (H,...
Andrew Hanushevsky
(STANFORD LINEAR ACCELERATOR CENTER)
4/14/15, 6:00 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
As more experiments move to a federated model of data access the environment becomes highly distributed and decentralized. In many cases this may pose obstacles in quickly resolving site issues; especially given vast time-zone differences. Spurred by ATLAS needs, Release 4 of XRootD incorporates a special mode of access to provide remote debugging capabilities. Essentially, XRootD allows a...
Mr
Andreas Joachim Peters
(CERN)
4/14/15, 6:00 PM
Archiving data to tape is a critical operation for any storage system, especially for the EOS system at CERN which holds production data from all major LHC experiments. Each collaboration has an allocated quota it can use at any given time therefore, a mechanism for archiving "stale" data is needed so that storage space is reclaimed for online analysis operations.
The archiving tool that we...
Dr
Tobias Stockmanns
(FZ Jülich GmbH)
4/14/15, 6:00 PM
Future particle physics experiments are searching more and more for rare decays which have similar signatures in the detector as the huge background. For those events usually simple selection criteria do not exist, which makes it impossible to implement a hardware-trigger based on a small subset of detector data.
Therefore all the detector data is read out continuously and processed...
Mr
Pawel Szostek
(CERN)
4/14/15, 6:00 PM
Track8: Performance increase and optimization exploiting hardware features
oral presentation
As Moore's Law drives the silicon industry towards higher transistor counts, processor designs are becoming more and more complex. The area of development includes core count, execution ports, vector units, uncore architecture and finally instruction sets. This increasing complexity leads us to a place where access to the shared memory is the major limiting factor, making feeding the cores...
Alexander Baranov
(ITEP Institute for Theoretical and Experimental Physics (RU))
4/14/15, 6:00 PM
Computational grid (or simply 'grid') infrastructures are powerful but restricted by several aspects: grids are incapable of running user jobs compiled with a non-authentic set of libraries and it is difficult to restructure grids to adapt to peak loads. At the same time if grids are not loaded with user-tasks, owners still have to pay for electricity and hardware maintenance. So a grid is not...
Mr
Eitaro Hamada
(High Energy Accelerator Research Organization (KEK))
4/14/15, 6:00 PM
**1. Introduction**
We developed a DAQ system of the J-PARC E16 Experiment by using the DAQ-Middleware. We evaluated the DAQ system and confirmed that the DAQ system can be applied to the experiment.
The DAQ system receives an average 660MB/spill of data (2-seconds spill per 6 seconds cycle). In order to receive such a large quantity of data, we need a network-distributed system....
Dr
Sergey Linev
(GSI DARMSTADT)
4/14/15, 6:15 PM
The *Data Acquisition Backbone Core* (*DABC*) is a C++ software framework that can implement and run various data acquisition solutions on Linux platforms. In 2013 version 2 of *DABC* has been released with several improvements. These developments have taken into account lots of practical experiences of *DABC v1* with detector test beams and laboratory set-ups since first release in 2009. The...
Mikhail Hushchyn
(Moscow Institute of Physics and Technology, Moscow)
4/14/15, 6:15 PM
The amount of data produced by the LHCb experiment every year consists of several petabytes. This data is kept on disk and tape storage systems. Disks are much faster than tapes, but are way more expensive and hence disk space is limited. It is impossible to fit the whole data taken during the experiment's lifetime on disk, but fortunately fast access to datasets are no longer needed after the...
Sameh Mannai
(Universite Catholique de Louvain (UCL) (BE))
4/14/15, 6:15 PM
The Semi-Digital Hadronic CALorimeter(SDHCAL) using Glass Resistive Plate Chambers (GRPCs) is
one of the two hadronic calorimeter options proposed by the ILD (International Large Detector) project for the future (ILC) International Linear Collider experiments.
It is a sampling calorimeter with 48 layers. Each layer has a size of 1 m² and
finely segmented into cells of 1 cm² ensuring a...
Christopher Hollowell
(Brookhaven National Laboratory)
4/14/15, 6:15 PM
Track8: Performance increase and optimization exploiting hardware features
oral presentation
Non-uniform memory access (NUMA) is a memory architecture for symmetric
multiprocessing (SMP) systems where each processor is directly connected
to separate memory. Indirect access to other CPU's (remote) RAM is still
possible, but such requests are slower as they must also pass through that
memory's controlling CPU. In concert with a NUMA-aware operating
system, the NUMA hardware...
Oliver Gutsche
(Fermi National Accelerator Lab. (US))
4/15/15, 9:00 AM
Dr
Maria Girone
(CERN)
4/15/15, 9:45 AM
Thorsten Sven Kollegger
(GSI - Helmholtzzentrum fur Schwerionenforschung GmbH (DE))
4/15/15, 11:00 AM
Mr
Tom Fifield
(OpenStack Foundation)
4/15/15, 11:30 AM
Manuel Delfino Reznicek
(Universitat Autònoma de Barcelona (ES))
4/16/15, 9:00 AM
Several scientific fields, including Astrophysics, Astroparticle Physics, Cosmology, Nuclear and Particle Physics, and Research with Photons, are estimating that by the 2020 decade they will require data handling systems with data volumes approaching the Zettabyte distributed amongst as many as 1018 individually addressable data objects (Zettabyte-Exascale systems). It may be...
Dr
Maria Grazia Pia
(Universita e INFN (IT))
4/16/15, 9:00 AM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
Testable physics by design
The validation of physics calculations requires the capability to thoroughly test them. The difficulty of exposing parts of the software to adequate testing can be the source of incorrect physics functionality, which in turn may generate hard to identify systematic effects in physics observables produced by the experiments.
Starting from real-life examples...
Glen Cowan
(Royal Holloway, University of London)
4/16/15, 9:00 AM
High Energy Physics has been using Machine Learning techniques (commonly known as Multivariate Analysis) since the 1990s with Artificial Neural Net and more recently with Boosted Decision Trees, Random Forest etc. Meanwhile, Machine Learning has become a full blown field of computer science. With the emergence of Big Data, data scientists are developing new Machine Learning algorithms to...
Martin Gasthuber
(Deutsches Elektronen-Synchrotron (DE))
4/16/15, 9:15 AM
Data taking and analysis infrastructures in HEP have evolved during many years to a well known problem domain. In contrast to HEP, third generations synchrotron light sources, existing and upcoming free electron laser are confronted an explosion in data rates which is primarily driven by recent developments in 2D pixel array detectors. The next generation will produce data in the region...
Ioannis Charalampidis
(CERN)
4/16/15, 9:15 AM
Lately there is a trend in scientific projects to look for computing resources in the volunteering community. In addition, to reduce the development effort required to port the scientific software stack to all the known platforms, the use of Virtual Machines (VMs) as end-projects is becoming increasingly popular. Unfortunately, the installation and the interfacing with the existing...
Elisabetta Ronchieri
(INFN)
4/16/15, 9:15 AM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
Geant4 is a widespread simulation system of "particles through matter" used in several experimental areas from high energy physics and nuclear experiments to medical studies. Some of its applications may involve critical use cases; therefore they would benefit from an objective assessment of the software quality of Geant4. The issue of maintainability is especially relevant for such a widely...
Andrea Perrotta
(Universita e INFN, Bologna (IT))
4/16/15, 9:15 AM
The CMS experiment has been designed with a 2-level trigger system. The first level is implemented using custom-designed electronics. The second level is the so-called High Level Trigger (HLT), a streamlined version of the CMS offline reconstruction software running on a computer farm. For Run II of the Large Hadron Collider, the increases in center-of-mass energy and luminosity will raise the...
Mr
Joao Correia Fernandes
(CERN)
4/16/15, 9:15 AM
We will present an overview of the current real-time video service offering for the LHC, in particular the operation of the CERN Vidyo service will be described in terms of consolidated performance and scale: The service is an increasingly critical part of the daily activity of the LHC collaborations, topping recently more than 50 million minutes of communication in one year, with peaks of up...
Dr
Patrick Fuhrmann
(DESY)
4/16/15, 9:30 AM
With the great success of the dCache Storage Technology in the framework of the World Wide LHC Computing Grid, an increasing number of non HEP communities were attracted to use dCache for their data management infrastructure. As a natural consequence, the dCache team was presented with new use-cases that stimulated the development of interesting dCache features.
Perhaps the most important...
Danilo Piparo
(CERN)
4/16/15, 9:30 AM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
The sixth release cycle of ROOT is characterised by a radical modernisation in
the core software technologies the tookit relies on: language standard,
interpreter, hardware exploitation mechanisms.
If on the one hand, the change offered the opportunity of consolidating the
existing codebase, in presence of such innovations, maintaing the balance
between full backward compatibility and...
Dominick Rocco
(urn:Google)
4/16/15, 9:30 AM
In this paper we present the Library Event Matching (LEM) classification technique for particle identification. The LEM technique was developed for the NOvA electron neutrino appearance analysis as an alternative but complimentary approach to standard multivariate methods. Traditional multivariate PIDs are based on high-level reconstructed quantities which can obscure or discard important...
Dr
Paul Millar
(Deutsches Elektronen-Synchrotron (DE))
4/16/15, 9:45 AM
The availability of cheap, easy-to-use sync-and-share cloud services has split the scientific storage world into the traditional big data management systems and the very attractive sync-and-share services. With the former, the location of data is well understood while the latter is mostly operated in the Cloud, resulting in a rather complex legal situation.
Beside legal issues, those two...
Bruno Silva De Sousa
(CERN)
4/16/15, 9:45 AM
The emergence of social media platforms in the consumer space unlocked new ways of interaction between individuals on the web. People develop now their social networks and relations based on common interests and activities with the choice to opt-in or opt-out on content of their interest. This kind of platforms have also an important place to fill inside large organizations and enterprises...
Stefan Gadatsch
(NIKHEF (NL))
4/16/15, 9:45 AM
In particle physics experiments data analyses generally use Monte Carlo (MC) simulation templates to interpret the observed data. These simulated samples may depend on one or multiple model parameters, such as a shifting mass parameter, and a set of such samples may be required to scan over the various parameter values. Since detailed detector MC simulation can be time-consuming, there is...
Tatiana Likhomanenko
(National Research Centre Kurchatov Institute (RU))
4/16/15, 9:45 AM
The main b-physics trigger algorithm used by the LHCb experiment is the so-called topological trigger. The topological trigger selects vertices which are a) detached from the primary proton-proton collision and b) compatible with coming from the decay of a b-hadron. In the LHC Run 1, this trigger utilized a custom boosted decision tree algorithm, selected an almost 100% pure sample of...
Philippe Canal
(Fermi National Accelerator Lab. (US))
4/16/15, 9:45 AM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
Following the release of version 6, ROOT has entered a new area of development. It will leverage the industrial strength compiler library shipping in ROOT 6 and its support of the C++11/14 standard, to significantly simplify and harden ROOT's interfaces and to clarify and substantially improve ROOT's support for multi-threaded environments.
This talk will also recap the most important new...
Steven Goldfarb
(University of Michigan (US))
4/16/15, 10:00 AM
The ATLAS Education and Outreach Group is in the process of migrating its public online content to a professionally designed set of web pages built on a Drupal-based content management system. Development of the front-end design passed through several key stages, including audience surveys, stakeholder interviews, usage analytics, and a series of fast design iterations, called sprints. ...
Alessandro Manzotti
(The University of Chicago)
4/16/15, 10:00 AM
CosmoSIS [http://arxiv.org/abs/1409.3409] is a modular system for
cosmological parameter estimation, based on Markov Chain Monte Carlo
(MCMC) and related techniques. It provides a series of samplers, which
drive the exploration of the parameter space, and a series of modules,
which calculate the likelihood of the observed data for a given physical
model, determined by the location of a...
Mr
Andreas Joachim Peters
(CERN)
4/16/15, 10:00 AM
EOS is an open source distributed disk storage system in production since 2011 at CERN. Development focus has been on low-latency analysis use cases for LHC and non-LHC experiments and life-cycle management using JBOD hardware for multi PB storage installations. The EOS design implies a split of hot and cold storage and introduced a change of the traditional HSM functionality based workflows...
Stefano Bagnasco
(Universita e INFN Torino (IT))
4/16/15, 10:00 AM
The present work aims at optimizing the use of computing resources available at the grid Italian Tier-2 sites of the ALICE experiment at CERN LHC by making them accessible to interactive distributed analysis, thanks to modern solutions based on cloud computing. The scalability and elasticity of the computing resources via dynamic (“on-demand”) provisioning is essentially limited by the size of...
Parag Mhashilkar
(Fermi National Accelerator Laboratory)
4/16/15, 10:15 AM
As part of the Fermilab/KISTI cooperative research project, Fermilab has successfully run an experimental simulation workflow at scale on a federation of Amazon Web Services (AWS), FermiCloud, and local FermiGrid resources. We used the CernVM-FS (CVMFS) file system to deliver the application software. We established Squid caching servers in AWS as well, using the Shoal system to let each...
Mr
Giulio Eulisse
(Fermi National Accelerator Lab. (US))
4/16/15, 10:15 AM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
In recent years the size and scale of scientific computing has grown
significantly. Computing facilities have grown to the point where energy
availability and costs have become important limiting factors
for data-center size and density. At the same time, power density
limitations in processors themselves are driving interest in more
heterogeneous processor architectures. Optimizing...
Mr
Joao Correia Fernandes
(CERN)
4/16/15, 10:15 AM
The CERN IT department has built over the years a performant and integrated ecosystem of collaboration tools, from videoconference and webcast services to event management software. These services have been designed and evolved in very close collaboration with the various communities surrounding the laboratory and have been massively adopted by CERN users. To cope with this very heavy usage,...
Dmytro Kresan
(GSI - Helmholtzzentrum fur Schwerionenforschung GmbH (DE))
4/16/15, 10:15 AM
The R3B (Reactions with Rare Radioactive Beams) experiment is one of the planned experiments at the future FAIR facility at GSI Darmstadt. R3B will cover experimental reaction studies with exotic nuclei far off stability, thus enabling a broad physics programs with rare-isotope beams with emphasis on nuclear structure and dynamics. Several different detection subsystems as well as...
Christoph Wissing
(Deutsches Elektronen-Synchrotron (DE))
4/16/15, 10:15 AM
The CMS experiment at the LHC relies on 7 Tier-1 centres of the WLCG to perform the majority of its bulk processing activity, and to archive its data. During the first run of the LHC, these two functions were tightly coupled as each Tier-1 was constrained to process only the data archived on its hierarchical storage. This lack of flexibility in the assignment of processing workflows...
Dr
Frederik Beaujean
(LMU Munich)
4/16/15, 10:15 AM
The Bayesian analysis toolkit (BAT)
is a C++ package centered around Markov-chain Monte Carlo sampling. It
is used in analyses of various particle-physics experiments such as
ATLAS and Gerda. The software has matured over the last few years to a
version 1.0. We will summarize the lessons learned and report on the
current developments of a complete redesign...
Dr
Samuel Cadellin Skipsey
4/16/15, 11:00 AM
The *Object Store* model has quickly become the de-facto basis of most commercially successful mass storage infrastructure, backing so-called "Cloud" storage such as Amazon S3, but also underlying the implementation of most parallel distributed storage systems.
Many of the assumptions in object store design are similar, but not identical, to concepts in the design of Grid Storage Elements,...
Vakho Tsulaia
(Lawrence Berkeley National Lab. (US))
4/16/15, 11:00 AM
Track8: Performance increase and optimization exploiting hardware features
oral presentation
High performance computing facilities present unique challenges and opportunities for HENP event processing. The massive scale of many HPC systems means that fractionally small utilizations can yield large returns in processing throughput. Parallel applications which can dynamically and efficiently fill any scheduling opportunities the resource presents benefit both the facility (maximal...
Stefan Nicolae Stancu
(CERN)
4/16/15, 11:00 AM
The LHC Optical Private Network, linking CERN and the Tier1s and the LHC Open Network Environment linking these to the Tier2 community successfully supported the data transfer needs of the LHC community during Run 1 and have evolved to better serve the networking requirements of the new computing models for Run 2. We present here the current status and the key changes, notably the delivered...
Dr
Andrew Norman
(Fermilab)
4/16/15, 11:00 AM
The NOvA experiment uses a continuous, free-running, dead-timeless data acquisition system to collect data from the 14 kT far detector. The DAQ system readouts the more than 344,000 detector channels and assembles the information into an raw unfiltered high bandwidth data stream. The NOvA trigger systems operate in parallel to the readout and asynchronously to the primary DAQ readout/event...
Oliver Keeble
(CERN)
4/16/15, 11:00 AM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
The overall success of LHC data processing depends heavily on stable, reliable and fast data distribution. The Worldwide LHC Computing Grid (WLCG) relies on the File Transfer Service (FTS) as the data movement middleware for moving sets of files from one site to another.
This paper describes the components of FTS3 monitoring infrastructure and how they are built to satisfy the common and...
Luca Mascetti
(CERN)
4/16/15, 11:15 AM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
Cernbox is a cloud synchronisation service for end-users: it allows to sync and share files on all major mobile and desktop platforms (Linux, Windows, MacOSX, Android, iOS) aiming to provide offline availability to any data stored in the CERN EOS infrastructure.
The successful beta phase of the service confirmed the high demand in the community for such easily accessible cloud storage...
Geert Jan Besjes
(Radboud University Nijmegen (NL))
4/16/15, 11:15 AM
We present a software framework for statistical data analysis, called *HistFitter*, that has been used extensively in the ATLAS Collaboration to analyze data of proton-proton collisions produced by the Large Hadron Collider at CERN. Most notably, HistFitter has become a de-facto standard in searches for supersymmetric particles since 2012, with some usage for Exotic and Higgs boson physics....
Sergey Panitkin
(Brookhaven National Laboratory (US))
4/16/15, 11:15 AM
Track8: Performance increase and optimization exploiting hardware features
oral presentation
The PanDA (Production and Distributed Analysis) workload management system (WMS) was developed to meet the scale and complexity of LHC distributed computing for the ATLAS experiment.
While PanDA currently uses more than 100,000 cores at well over 100 Grid sites with a peak performance of 0.3 petaFLOPS, next LHC data taking run will require more resources than Grid computing can possibly...
Mr
Michael Poat
(Brookhaven National Laboratory)
4/16/15, 11:15 AM
The STAR online computing environment is an intensive ever-growing system used for first-hand data collection and analysis. As systems become more sophisticated, they result in a more detailed dense collection of data output and inefficient limited storage systems have become an impediment to fast feedback to the online shift crews relying on data processing at near real-time speed. Motivation...
Mikolaj Krzewicki
(Johann-Wolfgang-Goethe Univ. (DE))
4/16/15, 11:15 AM
The ALICE High Level Trigger (HLT) is an online reconstruction, triggering and data compression system used in the ALICE experiment at CERN. Unique among the LHC experiments, it extensively uses modern coprocessor technologies like general purpose graphic processing units (GPGPU) and field programmable gate arrays (FPGA) in the data flow. Real-time data compression is performed using a cluster...
Parag Mhashilkar
(Fermi National Accelerator Laboratory)
4/16/15, 11:30 AM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
The FabrIc for Frontier Experiments (FIFE) program is an ambitious, major-impact initiative within the Fermilab Scientific Computing Division designed to lead the computing model development for Fermilab experiments and external projects. FIFE is a collaborative effort between physicists and computing professionals to provide computing solutions for experiments of varying scale, needs, and...
Dr
Tony Wildish
(Princeton University (US))
4/16/15, 11:30 AM
The LHC experiments have traditionally regarded the network as an unreliable resource, one which was expected to be a major source of errors and inefficiency at the time their original computing models were derived. Now, however, the network is seen as much more capable and reliable. Data are routinely transferred with high efficiency and low latency to wherever computing or storage resources...
Dr
Hironori Ito
(Brookhaven National Laboratory (US))
4/16/15, 11:30 AM
Ceph based storage solutions are becoming increasingly popular within the HEP/NP community over the last few years. With the current status of the Ceph project, both its object storage and block storage layers are production ready on a large scale, and even the Ceph file system (CephFS) storage layer is rapidly getting to that state as well. This contribution contains a thorough review of...
Manuel Martin Marquez
(CERN)
4/16/15, 11:30 AM
Data science is about unlocking valuable insights and obtaining deep knowledge out of the data. Its application enables more efficient daily-based operations and more intelligent decision-making processes. CERN has been very successful on developing custom data-driven control and monitoring systems. Several millions of control devices: sensors, front-end equipment, etc., make up these...
Dr
David Chamont
(LLR - École polytechnique)
4/16/15, 11:30 AM
Track8: Performance increase and optimization exploiting hardware features
oral presentation
The Matrix Element Method (MEM) is a well known powerful approach in particle physics to extract maximal information of the events arising from the LHC pp collisions. Compared to other methods requiring trainings, the MEM allows direct comparisons between a theory and the observation. Since the phase space has a higher dimensionality to explore, MEM is much more CPU time consuming at the...
Tom Uram
(urn:Google)
4/16/15, 11:45 AM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
HEP’s demand for computing resources has grown beyond the capacity of the Grid, and these demands will accelerate with the higher energy and luminosity planned for Run II. Mira, the ten petaflops supercomputer at the Argonne Leadership Computing Facility, is a potentially significant compute resource for HEP research. Through an award of fifty million hours on Mira, we have delivered millions...
Mr
Andreas Joachim Peters
(CERN)
4/16/15, 11:45 AM
In 2013, CERN IT evaluated then deployed a petabyte-scale Ceph cluster to support OpenStack use-cases in production. As of fall 2014, this cluster stores around 300 TB of data comprising more than a thousand VM images and a similar number of block device volumes. With more than a year of smooth operations, we will present our experience and tuning best-practices.
Beyond the cloud storage...
Taylor Childers
(Argonne National Laboratory (US))
4/16/15, 11:45 AM
Track8: Performance increase and optimization exploiting hardware features
oral presentation
Demand for Grid resources is expected to double during LHC Run II as compared to Run I; the capacity of the grid, however, will not double. The HEP community must consider how to bridge this computing gap. Two approaches to meeting this demand include targeting larger compute resources, and using the available compute resources as efficiently as possible. Argonne’s Mira, the fifth fastest...
Yu Higuchi
(High Energy Accelerator Research Organization (JP))
4/16/15, 11:45 AM
The ATLAS trigger has been used very successfully for the online event
selection during the first run of the LHC between 2009-2013 at a
centre-of-mass energy between 900 GeV and 8 TeV. The trigger system
consists of a hardware Level-1 (L1) and a software based high-level
trigger (HLT) that reduces the event rate from the design
bunch-crossing rate of 40 MHz to an average recording rate of...
Dave Kelsey
(STFC - Rutherford Appleton Lab. (GB))
4/16/15, 11:45 AM
The world is rapidly running out of IPv4 addresses; the number of IPv6 end systems connected to the internet is increasing; WLCG and the LHC experiments may soon have access to worker nodes and/or virtual machines (VMs) possessing only an IPv6 routable address. The HEPiX IPv6 Working Group (http://hepix-ipv6.web.cern.ch/) has been investigating, testing and planning for dual-stack services on...
Carlo Schiavi
(Universita e INFN Genova (IT))
4/16/15, 12:00 PM
Following the successful Run-1 LHC data-taking, the long shutdown gave the opportunity for significant improvements in the ATLAS trigger capabilities, as a result of the introduction of new or improved Level-1 trigger hardware and significant restructuring of the DAQ infrastructure. To make use of these new capabilities, the High-Level trigger (HLT) software has been to a large extent...
Ms
Xiaofeng LEI
(INSTITUE OF HIGH ENERGY PHYSICS, University of Chinese Academy of Sciences)
4/16/15, 12:00 PM
In the past years, we have successfully applied Hadoop to high-energy physics analysis. Although, we have not only improved the efficiency of data analysis, but also reduced the cost of cluster building so far, there are still some spaces to be optimized, like static pre-selection, low-efficient random data reading and I/O bottleneck caused by Fuse which is used to access HDFS. In order to...
Mr
Andreas Joachim Peters
(CERN)
4/16/15, 12:00 PM
The EOS storage software was designed to cover CERN disk-only storage use cases in the medium-term trading scalability against latency. To cover and prepare for long-term requirements the CERN IT data and storage services group (DSS) is actively conducting R&D and open source contributions to experiment with a next generation storage software based on CEPH.
CEPH provides a scale-out object...
Duncan Rand
(Imperial College Sci., Tech. & Med. (GB))
4/16/15, 12:00 PM
Named Data Networks (NDN) are an emerging network technology based around requesting data from a network rather than a specific host. Intermediate routers in the network cache the data. Each data packet must be signed to allow its provenance to be verified. Data blocks are addressed by a unique name which consists of a hierarchical path, a name and attributes. An example of a valid address...
Dr
Robert Andrew Currie
(Imperial College Sci., Tech. & Med. (GB))
4/16/15, 12:00 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
The DIRAC INTERWARE system was originally developed within the LHCb VO as a common interface to access distributed resources, i.e. grids, clouds and local batch systems. It has been used successfully in this context by the LHCb VO for a number of years. In April 2013 the GridPP consortium in the UK decided to offer a DIRAC service to a number of small VOs. The majority of these had been...
Robert Kutschke
(Femilab)
4/16/15, 12:15 PM
The art event processing framework is used by almost all new experiments at Fermilab, and by several outside of Fermilab. All use art as an external product in the same sense that the compiler, ROOT, Geant4, CLHEP and boost are external products. The art team has embarked on a campaign to document art and develop training materials for new users. Many new users of art have little or no...
Christos Papadopoulos
(Colorado State University)
4/16/15, 12:15 PM
Introduction
------------
The Computing Models of the LHC experiments continue to evolve from
the simple hierarchical MONARC model towards more agile models where
data is exchanged among many Tier2 and Tier3 sites, relying on both
strategic data placement, and an increased use of remote
access with caching through CMS's AAA and ATLAS' FAX projects, for example.
The challenges presented...
Dr
Andrew Norman
(Fermilab)
4/16/15, 12:15 PM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
oral presentation
As high energy physics experiments have grown, their operational needs and requirements they place on computing systems change. These changes often require new technical solutions to meet the increased demands and functionalities of the science. How do you affect sweeping change to core infrastructure, without causing major interruptions to the scientific programs?
This paper explores the...
Arnim Balzer
(Universiteit van Amsterdam)
4/16/15, 12:15 PM
The High Energy Stereoscopic System (H.E.S.S.) is an array of five imaging atmospheric Cherenkov telescopes located in the Khomas Highland in Namibia. Very high energy gamma rays are detected using the Imaging Atmospheric Cherenkov Technique. It separates the Cherenkov light emitted by the background of mostly hadronic air showers from the light emitted by air showers induced by gamma rays....
Niko Neufeld
(CERN)
4/16/15, 2:00 PM
Michael Ernst
(Unknown)
4/16/15, 2:45 PM
Jakob Blomer
(CERN)
4/16/15, 4:00 PM
Prof.
Daniele Bonacorsi
(University of Bologna)
4/16/15, 4:30 PM
Tadashi Maeno
(Brookhaven National Laboratory (US))
4/16/15, 4:40 PM
Track5: Computing activities and Computing models
Danilo Piparo
(CERN)
4/16/15, 5:30 PM
Track8: Performance increase and optimization exploiting hardware features
Marco Clemencic
(CERN)
4/17/15, 10:45 AM
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
Jose Flix Molina
(Centro de Investigaciones Energ. Medioambientales y Tecn. - (ES)
4/17/15, 11:10 AM
Andre Sailer
(CERN),
Andrew David Lahiff
(STFC - Rutherford Appleton Lab. (GB)),
Anna Elizabeth Woodard
(University of Notre Dame (US)),
Aram Santogidis
(CERN),
Christophe Haen
(CERN),
Dai Kobayashi
(Tokyo Institute of Technology (JP)), Mr
Erekle Magradze
(Georg-August-Universitaet Goettingen (DE)),
Yuji Kato
4/17/15, 11:35 AM
Dr
Richard Philip Mount
(SLAC National Accelerator Laboratory (US))
4/17/15, 12:30 PM
Prof.
Douglas Thain
(University of Notre Dame),
Haiyan Meng
(U), Prof.
Michael Hildreth
(University of Notre Dame)
Track5: Computing activities and Computing models
poster presentation
The reproducibility of scientific results increasingly depends upon the preservation of computational artifacts. Although preserving a computation to be used later sounds trivial, it is surprisingly difficult due to the complexity of existing software and systems. Implicit dependencies, networked resources, and shifting compatibility all conspire to break applications that appear to work well....
Shaun de Witt
(STFC)
Track3: Data store and access
poster presentation
For many years the Storage Resource Manager (SRM) has been the de-facto federation technology used by WLCG. This technology has, along with the rest of the middleware stack, mediated the transfer of many Petabytes of data since the start of data taking. In recent years, other technologies have become more popular as federation technologies because they offer additional functionalities that...
Dominick Rocco
(urn:Google)
Track2: Offline software
poster presentation
The NuMI Off-axis Neutrino Experiment (NOvA) is designed to study neutrino oscillations in the NuMI beam at Fermilab. Neutrinos at the Main Injector (NuMI) is currently being upgraded to provide 700 kW for NOvA. A 14 kt Far Detector in Ash River, MN and a functionally identical 0.3 kt Near Detector at Fermilab are positioned 810 km apart in the NuMI beam line. The fine granularity of the NOvA...
Domenico D'Urso
(Universita e INFN (IT))
Track2: Offline software
poster presentation
A flexible and modular data format implementation for HEP applications is presented.
Designed to face HEP data issues, the implementation is based on the CERN ROOT toolkit.
The design is aimed to create a data format as much as possible modular and easily upgradable and extendable.
Event informations are split into different files, that may contain different parts of the event (i.e....
Mr
Benjamin Farnham
(CERN), Mr
Piotr Pawel Nikiel
(CERN)
Track1: Online computing
poster presentation
This paper describes a new approach for generic design and efficient development of OPC UA servers. Development starts with creation of a design file, in XML format, describing an object-oriented information model of the target system or device. Using this model, the framework generates an executable OPC UA server application, which exposes the per-design OPC UA address space, without the...
Mr
Alessandro Italiano
(INFN-Bari)
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
poster presentation
This work presents the result of several tests that demonstrate the capabilities of HTCondor as batch system for a big computing farm serving both LHC use cases and others scientists.
The HTCondor testbed hosted at INFN-Bari is made of about 300 nodes and 15’000 CPU slots, and meant to sustain about 50’000 job in the queue. The computing farm is used both from Grid users of many VOs (HEP,...
Janusz Martyniak
(Imperial College London)
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
poster presentation
The international Muon Ionisation Cooling Experiment (MICE) is designed
to demonstrate the principle of muon ionisation cooling for the first
time, for application to a future Neutrino Factory or Muon Collider. The
experiment is currently under construction at the ISIS synchrotron at
the Rutherford Appleton Laboratory, UK. As presently envisaged, the
programme is divided into three Steps:...
Andrea Formica
(CEA/IRFU,Centre d'etude de Saclay Gif-sur-Yvette (FR))
Track3: Data store and access
poster presentation
Usage of Conditions Data in ATLAS is extensive for offline reconstruction and analysis (for example: alignment, calibration, data quality). The system is based on the LCG Conditions Database infrastructure, with read and write access via an ad hoc C++ API (COOL), a system which was developed before Run 1 data taking began. The infrastructure dictates that the data is organized into separate...
Andrea Biagioni
(INFN)
Track8: Performance increase and optimization exploiting hardware features
poster presentation
NaNet-10 is a four-ports 10GbE PCIe Network Interface Card designed for low-latency real-time operations with GPU systems.
For this purpose the design includes a UDP offload module, for a fast and deterministic to clock-cyle handling of transport layer protocol, plus a GPUDirect P2P/RDMA engine for low-latency communication with nVIDIA Tesla GPU devices.
A dedicate module (Merger) can...
Mike Hildreth
(University of Notre Dame (US))
Track2: Offline software
poster presentation
The CMS Simulation uses minimum bias events created by a "standard" event generator (e.g., Pythia) to simulate the additional interactions due to peripheral proton-proton collisions in each bunch crossing at the LHC (also known as pileup). Due to the inherent time constants of the CMS front-end electronics, many bunch crossings before and after the central bunch crossing of interest must be...
Alessandro De Salvo
(Universita e INFN, Roma I (IT)), Dr
Silvio Pardi
(INFN)
Track7: Clouds and virtualization
poster presentation
The advancements in technologies on provisioning end-to-end network services over geographical networks, together with the consolidation of Cloud Technologies, allow the creation of innovative scenarios for data centers.
In this work, we present the architecture and performance studies concerning a prototype of distributed Tier2 infrastructure for HEP, instantiated between the two...
Ian Peter Collier
(STFC - Rutherford Appleton Lab. (GB))
Track7: Clouds and virtualization
poster presentation
The RAL Tier-1 has been deploying production virtual machines for several years. As we move to providing a production private cloud, managed using OpenNebula, we have experimented with a range of different ways of deploying virtual machine images on hypervisors. We present a quantative comparison of a variety of virtual machine image and storage combinations, including monolithic Scientific...
Dr
Go Iwai
(KEK)
Track2: Offline software
poster presentation
We describe the development of an environment for Geant4 consisting of the application and data that enables users a faster and easier way to access the Geant4 applications without having to download and build the software locally. The environment is platform neutral and offers the users near-real time performance. The environment consists of data and Geant4 libraries built using the LLVM...
Dr
Dario Berzano
(CERN)
Track7: Clouds and virtualization
poster presentation
One of the most important steps of software lifecycle is Quality Assurance: this process comprehends both automatic tests and manual reviews, and all of them must pass successfully before the software is approved for production. Some tests, such as source code static analysis, are executed on a single dedicated service: in High Energy Physics, a full simulation and reconstruction chain on a...
Andrew John Washbrook
(University of Edinburgh (GB))
Track8: Performance increase and optimization exploiting hardware features
poster presentation
Multivariate training and classification methods using machine learning techniques are commonly applied in data analysis at HEP experiments. Despite their success in looking for signatures of new physics beyond the standard model it is known that some of these techniques are computationally bound when input sample size and model complexity are increased. Investigating opportunities for...
Dr
Domenico Giordano
(CERN)
Track7: Clouds and virtualization
poster presentation
Helix Nebula – the Science Cloud Initiative is a public-private-partnership between Europe's leading scientific research organisations (notably CERN, EMBL and ESA) and European IT cloud providers, that aims to establish a cloud-computing platform for data intensive science within Europe.
Over the past two years, Helix Nebula has built a federated cloud framework – the Helix Nebula...
Alexey Anisenkov
(Budker Institute of Nuclear Physics (RU))
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
poster presentation
The variety of the ATLAS Computing Infrastructure requires a central information system to define the topology of computing resources and to store the different parameters and configuration data which are needed by the various ATLAS software components.
The ATLAS Grid Information System (AGIS) is the system designed to integrate configuration and status information about resources, services...
Dr
Stephane Guillaume Poss
(Alpes Lasers SA)
Track7: Clouds and virtualization
poster presentation
We provide a report on ALDIRAC, the first DIRAC extension for a commercial application. DIRAC is a complete distributed computing solution, initially implemented for the LHCb experiment but now used by a wider community. The ALDIRAC extension is designed for the Alpes Lasers SA company in Neuchatel, Switzerland, to perform the simulation of the properties of Quantum Cascade Lasers on a Cloud...
Mr
Tadeas Bilka
(Charles University in Prague)
Track2: Offline software
poster presentation
The Belle II experiment will start taking data in 2017. The SuperKEKB accelerator will deliver a factor 40 higher luminosity in comparison to its predecessor, KEKB, to acquire a 50 times larger data sample of B-B̅ events. In order to manage higher occupancy and background, a new silicon vertex detector consisting of two inner layers of DEPFET pixel sensors surrounded by four layers of...
Javier Jimenez Pena
(Instituto de Fisica Corpuscular (ES))
Track2: Offline software
poster presentation
ATLAS is equipped with a tracking system built using different technologies, silicon planar sensors (pixel and micro-strip) and gaseous drift- tubes, all embedded in a 2T solenoidal magnetic field. For the LHC Run II, the system has been upgraded with the installation of a new pixel layer, the Insertable B-layer (IBL). Offline track alignment of the ATLAS tracking system has to deal with about...
Dr
Vladimir Sapunenko
(INFN-CNAF (IT))
Track3: Data store and access
poster presentation
Data management constitutes one of the major challenges that a geographically-distributed data centre has to face, especially when remote data access is involved. We discuss an integrated solution which enables transparent and efficient access to online and nearline data through high latency networks. This is based on the joint use of the General Parallel File System (GPFS) and of the Tivoli...
Dr
Huiyoung Ryu
(KISTI), Dr
Junghyun Kim
(KISTI), Prof.
Kihyeon Cho
(KISTI)
Track2: Offline software
poster presentation
For the solution of searching for new physics beyond the standard model, we do a research and development of simulation tool kit based on the evolving computing architecture with international collaboration. Using the tools, we study particle physics beyond the standard model. We present the current status of the research and development for these.
Robert Fischer
(Rheinisch-Westfaelische Tech. Hoch. (DE))
Track5: Computing activities and Computing models
poster presentation
Within CERN's new open data portal, the CMS collaboration provides a substantial fraction of its recorded data to the public. To explore and analyse the data, computing resources, an analysis framework, and documentation are required as well. While scientists can use C++ and the experiment software CMSSW in virtual machines, a simpler approach is needed, e.g. for university students who are in...
Lukas Alexander Heinrich
(New York University (US))
Track5: Computing activities and Computing models
poster presentation
Long before data taking ATLAS established a policy that all analyses need to be preserved. In the initial data-taking period, this has been achieved by various tools and techniques. ATLAS is now reviewing the analysis preservation with the aim to bring coherence and robustness to the process and with a clearer view of the level of reproducibility that is reasonably achievable. The secondary...
Mr
Phil Demar
(Fermilab), Dr
Wenji Wu
(Fermi National Accelerator Laboratory)
Track6: Facilities, Infrastructure, Network
poster presentation
Software-Defined Networking (SDN) has emerged as a major development direction in network technology. Conceptually, SDN enables customization of forwarding through network infrastructure on a per-flow basis. With SDN, a high impact LHC data flow could be allocated a “slice” of the network infrastructure. Functionally, the data flow would have a private path through the network infrastructure,...
Lu Wang
Track6: Facilities, Infrastructure, Network
poster presentation
The cluster of CC-IHEP is a middle sized computing system providing 10 thousands CPU cores, 3 PB disk storage, and 40 GB /s IO throughput. Its 1000+ users come from serials of HEP experiments including ATLAS, BESIII, CMS, DYB, JUNO, YBJ etc. In such a system, job statistics is necessary to find performance bottlenecks, locate software pitfalls, identify suspicious behaviors and make resource...
Michi Hostettler
(Universitaet Bern (CH))
Track8: Performance increase and optimization exploiting hardware features
poster presentation
The Piz Daint Cray XC30 HPC system at CSCS, the Swiss National Supercomputing centre, is in 2014 the highest ranked European system on TOP500, also featuring GPU accelerators. Event generation and detector simulation for the ATLAS experiment has been enabled for this machine. We report on the technical solutions, performance, HPC policy challenges and possible future opportunities for HEP on...
Graeme Stewart
(University of Glasgow (GB))
Track2: Offline software
poster presentation
To deal with Big Data flood from the ATLAS detector most events have to be rejected in the trigger system. the trigger rejection is complicated by the presence of a large number of minimum-bias events – the pileup. To limit pileup effects in the high luminosity environment of the LHC Run-2, ATLAS relies on full tracking provided by the Fast TracKer (FTK) implemented with custom...
Shima Shimizu
(Kobe University (JP))
Track1: Online computing
poster presentation
The immense rate of proton-proton collisions at the Large Hadron
Collider (LHC) must be reduced from the nominal bunch-crossing rate of
40 MHz to approximately 1 kHz before the data can be written on disk
offline. The ATLAS Trigger System performs real-time selection of
these events in order to achieve this reduction. Dedicated selection
of events containing jets is uniquely challenging...
Andreu Pacheco Pages
(Institut de Física d'Altes Energies - Barcelona (ES))
Track2: Offline software
poster presentation
In this presentation we will review the ATLAS Monte Carlo production setup including the different production steps involved in full and fast detector simulation. A report on the Monte Carlo production campaigns during Run-I, Long Shutdown 1 (LS1) and status of the production for Run-2 will be presented. The presentation will include the details on various performance aspects....
James Catmore
(University of Oslo (NO)),
Roger Jones
(Lancaster University (GB))
Track2: Offline software
poster presentation
Based on experience gained from run-I of the LHC, the ATLAS vertex reconstruction group has developed a refined primary vertex reconstruction strategy for run-II. With instantaneous luminosity exceeding 10^34 cm-2 s-1, an average of 40 to 50 pp collisions per bunch crossing are expected. Together with the increase of the center-of-mass collision energy from 8 TeV to 13 TeV, this will create...
Andreas Salzburger
(CERN)
Track2: Offline software
poster presentation
The successful physics program of Run-1 of the LHC with the discovery of the higgs boson in 2012 has put a strong emphasis on design studies for future upgrades of the existing LHC detectors and for future accelerators. Ideas how to cope with instantaneous luminosities way beyond the current specifications of the LHC in future tracking detectors are emerging and need sufficiently accurate...
Fabian Glaser
(Georg-August-Universitaet Goettingen (DE))
Track7: Clouds and virtualization
poster presentation
User analysis job demands can exceed available computing resources, especially before major conferences. ATLAS physics results might be slowed down due to this lack of resources available. For these reasons, cloud R&D activities are now included in the skeleton of the ATLAS computing model, which has been extended by using resources from commercial and private cloud providers to satisfy the...
Gianluca Cerminara
(CERN)
Track2: Offline software
poster presentation
Fast and efficient methods for the calibration and the alignment of the detector are a key asset to exploit the physics potential of the Compact Muon Solenoid (CMS) detector and to ensure timely preparation of results for conferences and publications.
To achieve this goal, the CMS experiment has set up a powerful framework. This includes automated workflows in the context of a prompt...
Mr
Erekle Magradze
(Georg-August-Universitaet Goettingen (DE))
Track6: Facilities, Infrastructure, Network
poster presentation
High-throughput computing platforms consist of complex infrastructure and provide a number of services apt to failures. To mitigate the impact of failures on the quality of the provided services, a constant monitoring and in time reaction is required, which is impossible without automation of the system administration
processes. This paper introduces a way of automation of the process of...
Oliver Schulz
(MPI for Physics, Munich)
Track2: Offline software
poster presentation
GERDA is an ultra-low background experiment, designed to search for the
neutrinoless double beta decay of Ge-76. The main background sources of
such an experiment are minute radioactive ontaminations of the experimental
setup itself. Gaining a good understanding of the individual contributions
to this radioactive background is vital not only for data analysis, but also
guides the design...
Katsuaki Tomoyori
(Japan Atomic Energy Agency)
Track2: Offline software
poster presentation
In neutron protein crystallography, it should be also emphasized that the weak Bragg reflections due to the large unit cells may be buried beneath the strong background caused by the incoherent scattering of hydrogen atoms. Therefore, the background estimation from the source is more reliable to improve the accuracy of Bragg integral intensity. We propose the adoption of Statistics-sensitive...
Mr
Olivier Couet
(CERN)
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
poster presentation
The ROOT reference guide is part of the code. Classes description, methods usage, examples etc...
are all embedded with the code itself. Doxygen is the reference model allowing to extract the
documentation from such self described system. The ROOT documentation requires the development
of specific tools (scripts) in the Doxygen context. The proposed project is these tools writing.
Rafal Zbigniew Grzymkowski
(Polish Academy of Sciences (PL))
Track7: Clouds and virtualization
poster presentation
The role of cloud computing technology in the distributed computing for HEP experiments grows rapidly. Some experiments (Atlas, BES-III, LHCb,…) already exploits private and public cloud resources for data processing. Future experiments such as Belle II or upgraded LHC experiments will largely rely on the availability of cloud resources and therefore their computing models have to be adjusted...
Dr
Xiaomei Zhang
(Institute of High Energy Physics)
Track5: Computing activities and Computing models
poster presentation
Distributed computing is necessary nowadays for high energy physics experiments to organize heterogeneous computing resources all over the world to process enormous amounts of data.
The BESIII experiment in China, which has aggregated about 3 PB of data over the last 5 years, has established its own distributed computing system, based on DIRAC, as a supplement to local clusters, collecting...
Manuel Martin Marquez
(CERN)
Track3: Data store and access
poster presentation
CERN’s accelerator complex is an extreme data generator, every second an important amount of comprehensively heterogeneous data coming from control equipment and monitoring agents is persisted and needs to be analysed. Over the decades, CERN’s researching and engineering teams have applied different approaches, techniques and technologies. This situation has minimized the necessary...
Luca Mazzaferro
(Universita e INFN Roma Tor Vergata (IT))
Track8: Performance increase and optimization exploiting hardware features
poster presentation
The possible usage of HPC resources by ATLAS is now becoming viable due to the changing nature of these systems and it is also very attractive due to the need for increasing amounts of simulated data.
In recent years the architecture of HPC systems has evolved, moving away from specialized monolithic systems, to a more generic linux type platform. This change means that the deployment of...
Andreas Petzold
(KIT - Karlsruhe Institute of Technology (DE))
Track7: Clouds and virtualization
poster presentation
The possibilities of cloud storage for use in HEP computing have been the topic of many studies and trials. The typical cloud storage values, easily accessible and expandable, relatively cheap and with a light weight interface have become available for local storage as well. Initially as part of larger environments like Open Nebula or OpenStack Swift, vendors now offer value storage with...
Lorenzo Moneta
(CERN)
Track2: Offline software
poster presentation
Differentiation is ubiquitous in high energy physics, for instance for minimization algorithms used for fitting and statistical analysis, detector alignment and calibration, and for theoretical physics. Automatic differentiation (AD) avoids well-known limitations in round-offs and speed, which symbolic and numerical differentiation suffer from, by transforming the source code of...
Mrs
Silvia Arezzini
(INFN - Pisa)
Track8: Performance increase and optimization exploiting hardware features
poster presentation
"CLUSTERALIVE"
"Clusteralive" is an integrated system developed in order to monitor and manage few important tasks in our HPC environment.
We have also other management systems, but now, with “Clusteralive” we can know immediately, just seeing our screen, if Clusters are up and running and we are sure that the most important functionality are well instanced.
"Clusteralive" is a php...
Marian Zvada
(University of Nebraska (US))
Track3: Data store and access
poster presentation
Over the past three years, the CMS Collaboration has developed the “Any Data, Anytime, Anywhere” technology to make use of a global data federation that is based on the XrootD protocol. The federation is now deployed across virtually all Tier-1 and Tier-2 sites in the CMS distributed computing system. This data federation gives workflows greater flexibility for location of execution, which has...
Clint Allan Richardson
(Boston University (US))
Track1: Online computing
poster presentation
The two-level trigger system employed by CMS consists of the Level 1 (L1) Trigger, which is implemented using custom-built electronics, and the High Level Trigger (HLT), a farm of commercial cpus running a streamlined version of the offline CMS reconstruction software. The operational L1 output rate of 100 kHz, together with the number of cpus in the HLT farm, imposes a fundamental constraint...
Mrs
Natalia Ratnikova
(Fermilab)
Track3: Data store and access
poster presentation
Storage capacity at CMS Tier-1 and Tier-2 sites reached over 100 Petabytes in 2014, and will be substantially increased during Run 2 data taking. The allocation of storage for the individual users analysis data, which is not accounted as a centrally managed storage space, will be increased to up to 40%. For comprehensive tracking and monitoring of the storage utilization across all...
Mr
Igor Mandrichenko
(FNAL)
Track3: Data store and access
poster presentation
Abstract: Conditions or calibration data are an important part of High Energy experiments. This kind of data is typically organized in terms of intervals of validity that require a special type of database table schema and API structure. At Fermilab we have designed and developed ConDB, a general tool to store, manage and retrieve conditions data organized into validity intervals in a...
Mr
Michael Poat
(Brookhaven National Laboratory)
Track6: Facilities, Infrastructure, Network
poster presentation
The STAR online computing environment is an intensive ever-growing system used for real-time data collection and analysis. Composed of heterogeneous and sometimes custom-tuned machine groups (Data Acquisition or DAQ computing, Trigger group, Sow Control or user-end data quality monitoring resources do not have the same requirements) the computing infrastructure was managed by manual...
Marko Slyz
(FNAL)
Track5: Computing activities and Computing models
poster presentation
The Dark Energy Survey (DES) uses a CCD camera installed in the Blanco
telescope in Cerro Tololo, Chile. The goal of the survey is to study
the effect known as Dark Energy.
DES uses Fermigrid for nightly processing, for quality assessement of
images, and for the detection of type 1A Super Novae. Nighly
processing needs to be carried out for each of the 105 nights in a
season that DES...
Adam Aurisano
(University of Cincinnati)
Track3: Data store and access
poster presentation
During operations, NOvA produces between 5,000 and 7,000 raw files per day with peaks in excess of 12,000. These files must be processed in several stages to produce fully calibrated and reconstructed analysis files. In addition, many simulated neutrino interactions must be produced and processed through the same stages as data. To accommodate the large volume of data and Monte Carlo,...
Osamu Tatebe
(University of Tsukuba)
Track3: Data store and access
poster presentation
Files in storage are often corrupted silently without any explicit
error. This is typically due to file system software bug, RAID
controller firmware bug, and some other reasons. Most critical issue
is damaged data is read without any error. Although there are several
mechanisms to detect data corruption in different layers such as ECC
in disk and memory and TCP checksum, the data may...
Christophe Haen
(CERN)
Track3: Data store and access
poster presentation
The DIRAC Interware provides a development framework and a complete set of components for building distributed computing systems. The DIRAC Data Management System (DMS) offers all the necessary tools to ensure data handling operations for small and large user communities. It supports transparent access to storage resources based on multiple technologies, and is easily expandable. The...
Dr
Federico Colecchia
(Brunel University London)
Track2: Offline software
poster presentation
The contamination from low-energy strong interactions is a major issue for data analysis at the Large Hadron Collider, particularly with reference to pileup, i.e. to proton-proton collisions from other bunch crossings. With a view to improving on the performance of pileup subtraction in higher-luminosity regimes, particle weighting methods have recently been proposed whereby the weights are...
Ruben Domingo Gaspar Aparicio
(CERN)
Track3: Data store and access
poster presentation
Inspired on different database as a service, DBaas, providers, the database group at CERN has developed a platform to allow CERN user community to run a database instance with database administrator privileges providing a full toolkit that allows the instance owner to perform backup/ point in time recoveries, monitoring specific database metrics, start/stop of the instance and...
Mr
Karsten Schwank
(DESY)
Track3: Data store and access
poster presentation
Increasingly, sites are using dCache to support communities that have different requirements from WLCG; as an example, DESY facilities and services now support photon sciences and biology groups. This presents new use-cases for dCache. Of particular interest is the chaotic file size distribution with a peak towards small files. This is problematic because tertiary storage systems, and tape in...
Mr
Tigran Mkrtchyan Mkrtchyan
(Deutsches Elektronen-Synchrotron DESY)
Track3: Data store and access
poster presentation
For over a decade, dCache.ORG has provided software which is used at more than 80 sites around the world, providing reliable services for WLCG experiments and others. This can be achieved only with a well established process from white board, where ideas are created, through packages, installed on the production systems. Since early 2013 we have moved to git as out source code management...
Thomas Lindner
(TRIUMF)
Track1: Online computing
poster presentation
DEAP-3600 is a dark matter experiment located at SNOLAB in Ontario, Canada. The DEAP detector uses 3600kg of liquid argon to search for the interactions of Weakly Interacting Massive Particles (WIMPs), a dark matter candidate. Light from the WIMP interactions is imaged using an array of 255 PMTs. A critical challenge for the DEAP experiment is the large background from 39Argon beta decays...
Wim Lavrijsen
(Lawrence Berkeley National Lab. (US))
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
poster presentation
The language improvements in C++11/14 greatly reduce the amount of boilerplate code required and allow resource ownership to be clarified in interfaces. On top, the Cling C++ interpreter brings a truly interactive experience and real dynamic behavior to the language. Taken together, these developments bring C++ much closer to Python in ability, allowing the combination of PyROOT/cppyy and...
Brian Davies
(STFC (RAL) (GB))
Track6: Facilities, Infrastructure, Network
poster presentation
perfSonar is a network monitoring tool set which enables performance of wide area communications to be analysed and eases problem identification across distributed centres. It has been widely used within WLCG since 2012 and has been crucial in identifying network problems and confirming network changes have the desired effect. We report on examples of this within this presentation.
In...
Jingyan Shi
(IHEP)
Track7: Clouds and virtualization
poster presentation
Batch system is a common way for a local cluster to schedule jobs running on work nodes. In some cases, some jobs have to stay in queue without suitable work nodes while some job slots have to keep free without suitable jobs running. The reasons for such case might be various. One of the main reasons is that operating system running on the free work nodes is different from the one that jobs in...
Franco Brasolin
(Universita e INFN (IT))
Track7: Clouds and virtualization
poster presentation
During the LHC long shutdown period (LS1), that started in 2013, the simulation in Point1 (Sim@P1) project takes advantage in an opportunistic way of the trigger and data acquisition (TDAQ) farm of the ATLAS experiment. The farm provides more than 1500 computer nodes, and they are particularly suitable for running event generation and Monte Carlo production jobs that are mostly CPU and not...
Andre Zibell
(Bayerische Julius Max. Universitaet Wuerzburg (DE))
Track1: Online computing
poster presentation
Ourania Sidiropoulou
on behalf of the ATLAS Muon Collaboration
A Micromegas (MM) quadruplet prototype with an active area of {0.5 m$^2$} that adopts the general design foreseen for the upgrade of the innermost forward muon tracking systems (Small Wheels) of the ATLAS detector in 2018-2019, has been built at CERN and is going to be tested in the ATLAS cavern environment...
Yuki Obara
(University of Tokyo)
Track1: Online computing
poster presentation
The purpose of the J-PARC E16 experiment is to investigate the origin of
hadron mass through the chiral symmetry restoration in nuclear matter.
In the experiment, we measure mass spectra of vector mesons in nuclei
in the $e^{+}e^{-}$ decay channel
with high precision and high statistics.
We use a 30 GeV proton beam with high intensity of $10^{10}$ per spill
to achieve high...
Andrew John Washbrook
(University of Edinburgh (GB)),
David Crooks
(University of Glasgow (GB))
Track7: Clouds and virtualization
poster presentation
The field of analytics, the process of analysing data to visualise meaningful patterns and trends, has become increasingly important to a wide range of scientific applications as the volume and variety of accessible data available to process (so called Big Data) has significantly increased. There are a number of scalable analytic platforms and services which have risen in prominence (such as...
Wataru Nakai
(University of Tokyo / RIKEN)
Track2: Offline software
poster presentation
The J-PARC E16 experiment will be performed to measure the mass modification of vector mesons in nuclear matter at J-PARC in order to study the origin of hadron mass.
In the experiment, we will measure invariant mass spectra of vector mesons with the electron and positron decay channel.
We will use 30 GeV proton beam with an intensity of $1\times10^{10}$ protons per pulse at High-momentum...
Dr
Tony Wildish
(Princeton University (US))
Track5: Computing activities and Computing models
poster presentation
We present an abstract view of data-transfer architectures in use in ATLAS and CMS. We use this to classify data-transfer tools not in terms of their technology, but in terms of their more basic features, such as the properties of the traffic they handle and the use-cases they serve.
This classification moves the focus from programming interfaces and technologies back into the original...
Geun Chul Park
(KiSTi Korea Institute of Science & Technology Information (KR))
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
poster presentation
AMGA (ARDA Metadata Grid Application) is a grid metadata catalog system that has been developed as a component of the EU FP7 EMI consortium based on the requirements of the HEP (High-Energy Physics) and the Biomed user communities. Currently, AMGA is exploited to manage the metadata in the gBasf2 framework at the Belle II which is one of the largest particle physics experiments in the world....
Andrew McNab
(University of Manchester (GB))
Track2: Offline software
poster presentation
The LHCb experiment has recorded the world’s largest sample of charmed meson decays. The search for matter-antimatter asymmetries in charm sector requires high precision analysis and thus intensive computing. This contribution will present a powerful method to measure matter-antimatter asymmetries in multi-body decays where GPU systems have been successfully exploited. In this method, local...
Christopher Jung
(KIT - Karlsruhe Institute of Technology (DE))
Track8: Performance increase and optimization exploiting hardware features
poster presentation
An ARM cluster, CEPH, ROOT and the energy balance
The total cost of ownershipt (TCO) of todays computer centres are
increasingly driven the power consumption of computing equipment. The
question arises if Intel based CPUs are still the best choice for analysis
tasks. Furthermore, data-driven computing models are emerging.
This contribution compares performance, TCO, power and energy...
Silvia Arezzini
(INFN Italy)
Track7: Clouds and virtualization
poster presentation
A large scale computing center, when not dedicated to a single/few users, has to face the problem of meeting ever changing user needs with respect to operating system version, architecture, availability of attached data volumes and logins. While clouds are a typical answer to these types of questions, they introduce resource problems like higher usage of RAM, difficulty to expose bare metal...
Steven Andrew Farrell
(Lawrence Berkeley National Lab)
Track5: Computing activities and Computing models
poster presentation
The ATLAS analysis model has been overhauled for the upcoming run of data collection in 2015 at 13 TeV. One key component of this upgrade was the Event Data Model (EDM), which now allows for greater flexibility in the choice of analysis software framework and provides powerful new features that can be exploited by analysis software tools. A second key component of the upgrade is the...
Stefano Dal Pra
(INFN)
Track7: Clouds and virtualization
poster presentation
The WLCG community and many groups in the HEP community have based
their computing strategy on the Grid paradigm, which proved successful
and still ensues its goals. However, Grid technology has not spread
much over other communities; in the commercial world, the cloud
paradigm is the emerging way to provide computing services.
WLCG experiments aim to achieve integration of their...
Dr
Gen Kawamura
(International Cloud Cooperation)
Track7: Clouds and virtualization
poster presentation
Grid computing enables deployments of large scale distributed computational infrastructures among different research facilities. It has been recently proposed that the Grid infrastructure be based on cloud computing. Provisioning systems and automated management frameworks using Cobbler, Rocks, Cfengine and Puppet are being successfully applied to many systems. Having implemented these new...
Stefano Dal Pra
(INFN)
Track6: Facilities, Infrastructure, Network
poster presentation
Tier-1 sites providing computing power for HEP experiments are
usually tightly designed for high throughput performances. This
is pursued by reducing the variety of supported usecases and
tuning for performances those ones, the most important of which
have been that of single-core jobs. Moreover, the usual workload
is saturation: each available core in the farm is in use and
there are...
Afiq Aizuddin Bin Anuar
(University of Malaya (MY))
Track1: Online computing
poster presentation
The CMS experiment has been designed with a 2-level trigger system. The first level is implemented using custom-designed electronics. The second level is the so-called High Level Trigger (HLT), a streamlined version of the CMS offline reconstruction software running on a computer farm. For Run II of the Large Hadron Collider, the increase in center-of-mass energy and luminosity will raise the...
Mr
Romain Wartel
(CERN)
Track5: Computing activities and Computing models
poster presentation
Federated identity management (FIM) is an arrangement made among multiple organisations that lets subscribers use the same identification data, e.g. account names & credentials, to obtain access to the secured resources and computing services of all other organisations in the group. Specifically in the various research communities there is an increased interest in a common approach as there is...
Christopher Jung
(KIT - Karlsruhe Institute of Technology (DE))
Track3: Data store and access
poster presentation
Most analyses in experimental high-energy physics (HEP) are based on the data analysis framework ROOT. Therefore, simulated as well as measured events are stored in ROOT trees. A typical analysis loops over events in ROOT files and selects relevant events for further processing according to certain selection criteria.
The emergence of NoSQL databases provide a new mean for large scale data...
Vikas Singhal
(Department of Atomic Energy (IN))
Track2: Offline software
poster presentation
The Compressed Baryonic Matter (CBM) experiment at the Facility for Anti-Proton and Ion Research (FAIR)in Darmstadt, Germany, is going to produce about 1 TByte per second of raw data at an interaction rate of 10 MHz for the measurement of very rare particles. Until now, all the HEP experiments are based on traditional hardware trigger approach; therefore all simulation and reconstruction...
Jerome Odier
(Centre National de la Recherche Scientifique (FR))
Track3: Data store and access
poster presentation
The ATLAS Metadata Interface (AMI) can be considered to be a mature application because it has existed for at least 10 years. Over the years, the number of users and the number of functions provided for these users has increased.
It has been necessary to adapt the hardware infrastructure in a seamless way so that the Quality of Service remains high.
We will describe the evolution of...
Brian Paul Bockelman
(University of Nebraska (US)), Dr
Jose Caballero Bejar
(Brookhaven National Laboratory (US))
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
poster presentation
The Open Science Grid Application Software Installation Service (OASIS)
provides an application installation service for Open Science Grid (OSG) virtual organizations (VOs) built on top of the CERN Virtual Machine File System (CVMFS).
This paper provides an overview and progress report of the OASIS service, which has been in production for over 18 months.
OASIS can be used either...
Thomas Lindner
(TRIUMF)
Track5: Computing activities and Computing models
poster presentation
ND280 is the off-axis near detector for the T2K neutrino experiment; it is designed to characterize the unoscillated T2K neutrino beam and measure neutrino cross-sections. We have developed a complicated system for processing and simulating the ND280 data, using computing resources from North America, Europe and Japan.
Recent work has concentrated on unifying our computing framework across...
Andrew David Lahiff
(STFC - Rutherford Appleton Lab. (GB))
Track7: Clouds and virtualization
poster presentation
Today it is becoming increasingly common for WLCG sites to provide both grid and cloud compute resources. In order to avoid the inefficiencies caused by static partitioning of resources it is necessary to integrate grid and cloud resources. There are two options to consider when doing this. The simplest option is to have the cloud manage all the physical hardware and use entirely virtualised...
Dr
Maria Grazia Pia
(Universita e INFN (IT)),
Sung Hun Kim
(H)
Track2: Offline software
poster presentation
Geant4 recommends a set of PhysicsLists and related classes (Builders, PhysicsConstructors) to its user community to facilitate the use of Geant4 functionality despite its intrinsic physics complexity. Relative limited documentation is available in the literature regarding Geant4 physics configuration tools, especially concerning the quantification of their accuracy, their computational...
Anna Elizabeth Woodard
(University of Notre Dame (US)),
Matthias Wolf
(University of Notre Dame (US))
Track5: Computing activities and Computing models
poster presentation
Individual scientists in high energy physics experiments like the Compact Muon Solenoid (CMS) experiment at the Large Hadron Collider require extensive use of computing resources for analysis of massive data sets. The majority of this analysis work is done at dedicated grid-enabled CMS computing facilities. University campuses offer considerable additional computing resources, but these are...
Arturos Sanchez Pineda
(Universita e INFN, Napoli (IT))
Track2: Offline software
poster presentation
We explore the potentialities of current web applications to create online interfaces that allow the visualization, interaction and real physics cut-based analysis and monitoring of processes trough a web browser. The project consists in the initial development of web-based and cloud computing services to allow students and researches to perform fast and very useful cut-based analysis on a...
Dr
Marc Paterno
(Fermilab)
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
poster presentation
The scientific discovery process can be advanced by the integration of
independently-developed programs run on disparate computing facilities
into coherent workflows usable by scientists who are not experts in
computing. For such advancement, we need a system which scientists can
use to formulate analysis workflows, to integrate new components to
these workflows, and to execute different...
Dr
Samuel Cadellin Skipsey
Track3: Data store and access
poster presentation
The state of the art in Grid style data management is to achieve increased resilience of data via multiple complete replicas of data files across multiple storage endpoints. While this is effective, it is not the most space-efficient approach to resilience, especially when the reliability of individual storage endpoints is sufficiently high that only a few will be inactive at any point in...
Dr
Helge Meinhard
(CERN)
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
poster presentation
We will present how CERN's services around Issue Tracking and Version Control have evolved, and what the plans for the future are. We will describe the services' main design, integration and structure, giving special attention to the new requirements from the community of users in terms of collaboration and integration tools and how we address this challenge when defining new services based on...
Dr
Bockjoo Kim
(University of Florida (US)), Dr
Dimitri Bourilkov
(University of Florida (US)),
Jorge Luis Rodriguez
(Florida International University (US)),
Paul Ralph Avery
(University of Florida (US)),
Yu Fu
(University of Florida (US))
Track3: Data store and access
poster presentation
One of the CMS Tier2 centers, the Florida CMS Tier2 center, has been using the Lustre filesystem for its data storage backend system since 2004. Recently, the data access pattern at our site has changed greatly due to various new access methods that include file transfers through the GridFTP servers, read access from the worker nodes, and remote read access through xrootd. In order to optimize...
Brian Davies
(STFC (RAL) GB)
Track3: Data store and access
poster presentation
The Rutherford Appleton Laboratory (RAL) operates the UK WLCG Tier1 facility on behalf of GridPP. Tier 1's provide persistent archival storage (on tape at RAL) and
online storage for fast access data analysis. RAL is one of the few Tier-1s which
supports data management for all the major LHC experiments, as well as a number of
smaller Virtual Organisations. This allows us to compare usage...
Prof.
Alberto Aloisio
(Universita' di Napoli Federico II and INFN)
Track1: Online computing
poster presentation
We present a feasibility study of a RF transmitters and modulators
based on parametric softcores fully embedded in a general purpose FPGA fabric,
without using external components.
This architecture aims at providing wireless physical layers
to I-o-T and NFC protocols with programmable hardware.
We show preliminary results with latest generation 7-series XILINX FPGA.
Manuel Delfino Reznicek
(Universitat Autònoma de Barcelona (ES))
Track6: Facilities, Infrastructure, Network
poster presentation
Energy consumption is an increasing concern for data centers. This paper summarizes recent energy efficiency upgrades at the Port d’Informació Científica (PIC) in Barcelona, Spain which have considerably lowered energy consumption. The upgrades were particularly challenging, as they involved modifying the already existing machine room, which is shared by PIC with the general IT services of the...
Alejandro Alvarez Ayllon
(CERN)
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
poster presentation
FTS3 is the service responsible for the distribution of the LHC data across the WLCG Infrastructure. To facilitate its use outside the traditional grid environment we have provided a web application - known as WebFTS - fully oriented towards final users, and easily usable within a browser.
This web application is completely decoupled from the core service, and interfaces with it via a REST...
Dr
Sebastien Binet
(IN2P3/LAL)
Track2: Offline software
poster presentation
`fwk`: a go-based concurrent control framework
============================================
Current HEP control frameworks have been designed and written in the
early 2000's, when multi-core architectures were not yet pervasive.
As a consequence, an inherently sequential event processing design
emerged.
Evolving current frameworks' APIs and data models encouraging global
states,...
Soon Yung Jun
(Fermi National Accelerator Lab. (US))
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
poster presentation
Performance evaluation and analysis of large-scale computing
applications is essential for optimizing the use of resources. As detector
simulation is one of the most compute-intensive tasks and Geant4 is the
simulation toolkit most widely used in contemporary high energy
physics (HEP) experiments, it is important to monitor Geant4
through its development cycle for changes in computing...
Dr
Tamborini Aurora
(University of Pavia - INFN Section of Pavia)
Track2: Offline software
poster presentation
**Purpose**
The aim of this work is a study of a possible use of carbon ion pencil beams (delivered with active scanning modality) for the treatment of ocular melanomas at the National Centre for Oncological Hadrontherapy (CNAO). The promising aspect of carbon ions radiotherapy for the treatment of this disease lies in its superior relative radiobiological effectiveness (RBE).
The Monte...
Dr
Simone Campana
(CERN)
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
poster presentation
The ATLAS Experiment at the Large Hadron Collider has collected data during Run 1 and is ready to collect data in Run 2. The ATLAS data are distributed, processed and analysed at more than 130 grid and cloud sites across the world. At any given time, there are more than 150,000 concurrent jobs running and about a million jobs are submitted on a daily basis on behalf of thousands of physicists...
Dr
Randy Sobie
(University of Victoria (CA))
Track7: Clouds and virtualization
poster presentation
The HEP community is increasingly using clouds that are distributed around the world for running its applications. The stringent software criteria of HEP experiments require that we use the identical (secure) virtual machine (VM) image at all sites with a minimal set of site-specific customizations. Nearly all cloud systems (such as OpenStack) require that the VM image to be instantiated must...
Haykuhi Musheghyan
(Georg-August-Universitaet Goettingen (DE))
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
poster presentation
The importance of monitoring on HEP grid computing systems is growing due to a significant increase in their complexities. Computer scientists and administrators have been studying and building effective ways to gather information on and clarify a status of each local grid infrastructure.
The HappyFace project aims at making the above-mentioned workflow possible. It aggregates, processes and...
Mr
Suman Sau
(Calcutta University)
Track1: Online computing
poster presentation
The Compressed Baryonic Matter (CBM) experiment is a part of the Facility for Antiproton and Ion Research (FAIR) in Darmstadt at the GSI. This experiment will examine heavy-ion collisions in fixed target geometry and will be able to measure hadrons, electrons and muons. Muon Chamber (MUCH) is used to detect low momentum muons in an environment of high particle densities. Basic read out chain...
Vincenzo Daponte
(Universite de Geneve (CH))
Track1: Online computing
poster presentation
The CMS High Level Trigger (HLT) is implemented running a streamlined version of the CMS offline reconstruction software running on thousands of CPUs. The CMS software is written mostly in C++, using Python as its configuration language through an embedded CPython interpreter. The configuration of each process is made up of hundreds of "modules", organized in "sequences" and "paths". As an...
Dr
Maria Grazia Pia
(Universita e INFN (IT)),
Peter Steinbach
(MPI-CBG),
Stefan Kluth
(Max-Planck-Institut fuer Physik (Werner-Heisenberg-Institut) (D), Dr
Thomas Schoerner-Sadenius
(DESY),
Thomas Velz
(Universitaet Bonn (DE))
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
poster presentation
The ability to judge, use and develop code efficiently and successfully is a key ingredient in modern particle physics. Software design plays a fundamental role in the software development process and is instrumental to many critical aspects in the life-cycle of an experiment: the transparency of software design enables the validation of physics results, contributes to the effective use of...
Edgar Fajardo Hernandez
(Univ. of California San Diego (US))
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
poster presentation
The HTCondor batch system is heavily used in the HEP community as the batch system for several WLCG resources. Moreover it is the backbone of the GlideInWMS, the main pilot system used by CMS. To prepare for LHC Run 2, we are probing the scalability limits of new versions and configurations of HTCondor with the goal of reaching at least 200,000 simultaneous running jobs in a single pool.
A...
Michal Husejko
(CERN)
Track8: Performance increase and optimization exploiting hardware features
poster presentation
In this paper we present our findings gathered during the evaluation and testing of Windows Server High Performance Computing (Windows HPC) in view of potentially using it as a production HPC system for engineering applications. The Windows HPC package, an extension of Microsoft's Windows Server product, provides all essential interfaces, utilities and management functionality for creating,...
Shawn Mc Kee
(University of Michigan (US))
Track6: Facilities, Infrastructure, Network
poster presentation
In today's world of distributed scientific collaborations, there are many challenges to providing reliable inter-domain network infrastructure. Network operators use a combination of active monitoring and trouble tickets to detect problems. However, some of these approaches do not scale to wide area inter-domain networks due to unavailability of data. The Pythia Network Diagnostic...
Stefano Dal Pra
(INFN)
Track7: Clouds and virtualization
poster presentation
While in the business world the cloud paradigm is typically implemented purchasing resources and services from third party providers (e.g. Amazon), in the scientific environment there's usually the need of on-premises IaaS infrastructures which allow efficient usage of the hardware distributed among (and owned by) different scientific administrative domains. In addition, the requirement of...
Mr
Robert Mina
(U. Virginia)
Track1: Online computing
poster presentation
The NOvA collaboration has constructed a 14,000 ton, fine-grained, low-Z, total absorption tracking calorimeter at an off-axis angle to an upgraded NuMI neutrino beam. This detector, with its excellent granularity and energy resolution, and relatively low-energy neutrino thresholds was designed to observe electron neutrino appearance in a muon neutrino beam but it also has unique capabilities...
380.
Implementing a Domain Specific Language to configure and run LHCb Continuous Integration builds
Marco Clemencic
(CERN)
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
poster presentation
The new LHCb nightly build system described at CHEP 2013 was limited by the use of JSON files for its configuration. JSON had been chosen as a temporary solution to maintain backward compatibility towards the old XML format by means of a translation function.
Modern languages like Python leverage on meta-programming techniques to enable the development of Domain Specific Languages...
Michael Boehler
(Albert-Ludwigs-Universitaet Freiburg (DE))
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
poster presentation
Every day hundreds of tests are run on the Worldwide LHC Computing Grid for the ATLAS, CMS, and LHCb experiments in order to evaluate the performance and reliability of the different computing sites. All this activity is steered, controlled, and monitored by the HammerCloud testing infrastructure. Sites with failing functionality tests are auto-excluded from the ATLAS computing grid, therefore...
Ben Couturier
(CERN),
Marco Clemencic
(CERN)
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
poster presentation
The purpose of this paper is to describe the steps that led to an improved interface for LHCb's Nightly Builds Dashboard. The goal was to have an efficient application that meets the needs of both the project developers, by providing them with a user friendly interface, as well as those of the computing team supporting the system by providing them with a dashboard allowing for better...
Jae-Hyuck Kwak
(KISTI)
Track3: Data store and access
poster presentation
This paper describes the recent improvement of AMGA python client library for the Belle II Experiment. We were drawn to the action items about library improvement after in-depth discussions with the developer of the Belle II distributed computing group. It includes GSI support, client-side metadata federation support and atomic operation support. Some of the improvements were already applied...
Alessandro De Salvo
(Universita e INFN, Roma I (IT)),
Domenico Elia
(INFN Bari),
Laura Perini
(Università degli Studi e INFN Milano (IT)),
Tommaso Boccali
(Sezione di Pisa (IT))
Track5: Computing activities and Computing models
poster presentation
In 2012, 14 Italian Institutions participating in all major LHC Experiments won a grant from the ITALIAN MINISTRY OF RESEARCH (MIUR), to optimise Analysis activities and in general the Tier2/Tier3 infrastructure. We report on the activities being researched upon, on the considerable improvement in the ease of access to resources by physicists, also those with no specific computing interests....
Andrew David Lahiff
(STFC - Rutherford Appleton Lab. (GB))
Track7: Clouds and virtualization
poster presentation
Today the primary method by which the LHC and other experiments run computing work at WLCG sites is grid job submission. Jobs are submitted to computing element middleware which in turn submits jobs to a batch system managing the local compute resources. With the increasing interest and usage of cloud technology, a new challenge facing sites which support multiple experiments in recent years...
Dr
Luca Mazzaferro
(Max-Planck-Institut fuer Physik (Werner-Heisenberg-Institut) (D)
Track6: Facilities, Infrastructure, Network
poster presentation
In a grid computing infrastructure tasks such as continuous upgrades, services installations and software deployments are part of an admins daily work. In such an environment tools to help with the management, provisioning and monitoring of the deployed systems and services have become crucial.
As experiments such as the LHC increase in scale, the computing infrastructure also becomes...
Andre Sailer
(CERN)
Track2: Offline software
poster presentation
The DD4hep detector description toolkit offers a flexible and easy to use solution for the consistent and complete description of particle physics detectors in one single system. It provides software components addressing visualisation, simulation, reconstruction and analysis of high energy physics data.
The Linear Collider community has adopted DD4hep early on in the development phase and...
Dr
Alexey Poyda
(NATIONAL RESEARCH CENTRE "KURCHATOV INSTITUTE"),
Eygene Ryabinkin
(National Research Centre Kurchatov Institute (RU); Moscow Institute for Physics and Technology, Applied computational geophysics lab), Dr
Ruslan Mashinistov
(NATIONAL RESEARCH CENTRE "KURCHATOV INSTITUTE"; P.N. Lebedev Institute of Physics (Russian Academy of Sciences))
Track8: Performance increase and optimization exploiting hardware features
poster presentation
During LHC Run1 ATLAS and ALICE produced more than 30 Petabytes of data, That rate outstripped any other scientific effort going on, even in data-rich fields such as genomics and climate science. To address an unprecedented multi-petabyte data processing challenge, the experiments are relying on the computational grid infrastructure deployed by the Worldwide LHC Computing Grid (WLCG).
LHC...
Dr
Dario Barberis
(Università e INFN Genova (IT))
Track3: Data store and access
poster presentation
The ATLAS EventIndex System, developed for use in LHC Run 2, is designed to index every processed event in ATLAS, replacing the TAG System used in Run 1. Its storage infrastructure, based on Hadoop, necessitates revamping how information in this system relates to other ATLAS systems. In addition, the scope of this new application is different from that of the TAG System. It will store fewer...
Alec Habig
(Univ. of Minnesota Duluth)
Track1: Online computing
poster presentation
The NOvA experiment, with a baseline of 810 km, samples Fermilab's upgraded NuMI beam with a Near Detector on-site and a Far Detector (FD) at Ash River, MN, to observe oscillations of muon neutrinos. The 344,064 liquid scintillator-filled cells of the 14 kton FD provide high granularity of a large detector mass and enable us to also study non-accelerator based neutrinos with our Data Driven...
Eygene Ryabinkin
(National Research Centre Kurchatov Institute (RU))
Track7: Clouds and virtualization
poster presentation
Cloud technologies allow easy load balancing between different tasks and projects. From the viewpoint of the data analysis in the ALICE experiment, cloud allows to deploy software using Cern Virtual Machine (CernVM) and CernVM File System (CVMFS), to run different (including outdated) versions of software for long term data preservation and to dynamically allocate resources for different...
Dr
Maria Girone
(CERN)
Track8: Performance increase and optimization exploiting hardware features
poster presentation
High energy physics experiments are experiencing a growth in the number of collected and processed events that exceeds the rate of growth in computing resources sustainable by technology improvements at a flat yearly cost. This trend is expected to continue into the foreseeable future, and as the field is not expecting a big increase in support, innovative approaches are needed. In areas of...
Michal Husejko
(CERN)
Track1: Online computing
poster presentation
High-Level Synthesis (HLS) for Field-Programmable Logic Array (FPGA) programming is becoming a practical alternative to well-established VHDL and Verilog languages. This paper describes a case study in the use of HLS tools to design an FPGA-based data acquisition systems (DAQ). We will present the implementation of the CERN CMS detector ECAL Data Concentrator Card (DCC) functionality in HLS...
Yuji Kato
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
poster presentation
The BelleII is an asymmetric energy $e^{+}e^{-}$ collider experiment at SuperKEKB in Japan. One of the main goals of BelleII is to search for physics beyond the Standard Model with a data set of about $5 \times 10^{10}$ $B\bar{B}$ pairs. In order to store such huge amount of data including MC events and analyze it in a timely manner, BelleII adopted distributed computing model with DIRAC...
Dr
Sergey Linev
(GSI DARMSTADT)
Track4: Middleware, software development and tools, experiment frameworks, tools for distributed computing
poster presentation
This is further development of JSRootIO project. Code was mostly rewritten to make it modular; I/O part was clearly separated from the graphics.
Many new interactive features were implemented:
- loading of required functionality on the fly;
- possibility of dynamic update of objects drawings;
- automatic resize of drawings when browser window is resized;
- move/resize of...