Frank Linde
(NIKHEF (NL))
10/14/13, 9:20 AM
Dr
Robert Lupton
(Princeton)
10/14/13, 11:00 AM
Many of the scientific computing frameworks used in 'big science' have several million lines of source code, and software engineering challenges are amongst the most prominent challenges, be it in high-energy physics, astronomy, or other sciences. Dr Robert Lupton of Princeton University will talk the software engineering challenges that face scientific computing and how large scale systems...
Dr
Kostas Glinos
(European Commission)
10/14/13, 11:45 AM
Through joint efforts between the HEP community in the early days of the EU DataGrid project, through EGEE, and via EGI-InSPIRE today, the European Commission has had a profound impact in the way computing and data management for high energy physics is done.
Kostas Glinos, Head of Unit eInfrastructures of the European Commission, has been with the European Commission since 1992. He leads...
Dr
Randy Sobie
(University of Victoria (CA))
10/14/13, 1:30 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Oral presentation to parallel session
The computing model of the ATLAS experiment was designed around the concept of grid computing and, since the start of data taking, this model has proven very successful. However, new cloud computing technologies bring attractive features to improve the operations and elasticity of scientific distributed computing. ATLAS sees grid and cloud computing as complementary technologies that will...
Tomasz Rybczynski
(AGH University of Science and Technology (PL))
10/14/13, 1:30 PM
Data Stores, Data Bases, and Storage Systems
Oral presentation to parallel session
The LHCb experiment records millions of proton collisions every second, but only a fraction of them are useful for LHCb physics.
In order to filter out the "bad events" a large farm of x86-servers (~2000 nodes) has been put in place. These servers boot from and run from NFS, however they use their local disk to temporarily store data, which cannot be processed in real-time...
Claudio Kopper
10/14/13, 1:30 PM
Software Engineering, Parallelism & Multi-Core
Oral presentation to parallel session
The IceCube Neutrino Observatory is a cubic kilometer-scale neutrino detector built into the ice sheet at the geographic South Pole. Light propagation in glacial ice is an important component of IceCube detector simulation that requires a large number of embarrassingly parallel calculations. The IceCube collaboration recently began using GPUs in order to simulate direct propagation of...
Marco Cattaneo
(CERN)
10/14/13, 1:30 PM
Distributed Processing and Data Handling B: Experiment Data Processing, Data Handling and Computing Models
Oral presentation to parallel session
The LHCb experiment has taken data between December 2009 and February 2013. The data taking conditions and trigger rate have been adjusted several times to make optimal use of the luminosity delivered by the LHC and to extend the physics potential of the experiment.
By 2012, LHCb was taking data at twice the instantaneous luminosity and 2.5 times the high level trigger rate than originally...
Rainer Schwemmer
(CERN)
10/14/13, 1:30 PM
Data acquisition, trigger and controls
Oral presentation to parallel session
The LHCb Data Acquisition system reads data from over 300 read-out boards and distributes them to more than 1500 event-filter servers. It uses a simple push-protocol over Gigabit Ethernet. After filtering, the data is consolidated into files for permanent storage using a SAN-based storage system.
Since the beginning of data-taking many lessons have been learned and the reliability and...
Alessandro Di Girolamo
(CERN)
10/14/13, 1:30 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Oral presentation to parallel session
The WLCG information system is just one of the many information sources that are required to populate a VO configuration database. Other sources include central portals such as the GOCDB and the OIM from EGI and OSG respectively. Providing a coherent view of all this information that has been synchronized from many different sources is a challenging activity and has been duplicated to various...
Frank-Dieter Gaede
(Deutsches Elektronen-Synchrotron (DE))
10/14/13, 1:30 PM
Event Processing, Simulation and Analysis
Oral presentation to parallel session
One of the key requirements for Higgs physics at the International Linear Collider ILC is excellent track reconstruction with very good momentum and impact parameter resolution. ILD is one of the two detector concepts at the ILC.
Its central tracking system comprises of a highly granular TPC, an intermediate silicon tracker and a pixel vertex detector, and it is complemented by silicon...
Jim Kowalkowski
(Fermilab)
10/14/13, 1:50 PM
Data acquisition, trigger and controls
Oral presentation to parallel session
The artdaq data acquisition software toolkit has been developed within the Fermilab Scientific Computing Division to meet the needs of current and future experiments. At its core, the toolkit provides data transfer, event building, and event analysis functionality, the latter using the art event analysis framework.
In the last year, functionality has been added to the toolkit in the areas...
Thomas Kuhr
(KIT)
10/14/13, 1:52 PM
Distributed Processing and Data Handling B: Experiment Data Processing, Data Handling and Computing Models
Oral presentation to parallel session
The Belle II experiment, a next-generation B factory experiment at KEK, is expected to record a two orders of magnitude larger data volume than its predecessor, the Belle experiment. The data size and rate are comparable to or more than the ones of LHC experiments and requires to change the computing model from the Belle way, where basically all computing resources were provided by KEK, to a...
Marcos Seco Miguelez
(Universidade de Santiago de Compostela (ES)),
Victor Manuel Fernandez Albor
(Universidade de Santiago de Compostela (ES))
10/14/13, 1:52 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Oral presentation to parallel session
The Datacenter at the Galician Institute of High Energy Physics(IGFAE) of the Santiago de Compostela University (USC) is a computing cluster with about 150 nodes and 1250 cores that hosts the LHCb Tiers 2 and 3. In this small datacenter, and of course in similar or bigger ones, it is very important to keep optimal conditions of temperature, humidity and pressure. Therefore, it is a necessity...
Dmytro Karpenko
(University of Oslo (NO))
10/14/13, 1:52 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Oral presentation to parallel session
During three years of LHC data taking, the ATLAS collaboration completed three petascale data reprocessing campaigns on the Grid, with up to 2 PB of data being reprocessed every year. In reprocessing on the Grid, failures can occur for a variety of reasons, while Grid heterogeneity makes failures hard to diagnose and repair quickly. As a result, Big Data processing on the Grid must tolerate a...
Prof.
Gang CHEN
(INSTITUTE OF HIGH ENERGY PHYSICS), Dr
Wenjing Wu
(IHEP, CAS)
10/14/13, 1:53 PM
Data Stores, Data Bases, and Storage Systems
Oral presentation to parallel session
The limitation of scheduling modules and the gradual addition of disk pools in distributed storage systems often result in imbalances among their disk pools in terms of both available space and number of files. This can cause various problems to the storage system such as single point of failure, low system throughput and imbalanced resource utilization and system loads. An algorithm named...
Philippe Canal
(Fermi National Accelerator Lab. (US))
10/14/13, 1:53 PM
Software Engineering, Parallelism & Multi-Core
Oral presentation to parallel session
We will present massively parallel high energy electromagnetic particle transportation through a finely segmented detector in the Graphic Processor Unit (GPU). Simulating events of energetic particle decay in a general-purpose high energy physics (HEP) detector requires intensive computing resources, due to the complexity of the geometry as well as physics processes applied to particles...
Leo Piilonen
(Virginia Tech)
10/14/13, 1:55 PM
Event Processing, Simulation and Analysis
Oral presentation to parallel session
I will describe the charged-track extrapolation and the muon identification modules in the Belle II data analysis code library. These modules use GEANT4E to extrapolate reconstructed charged tracks outward from the Belle II Central Drift Chamber into the outer particle-identification detectors, the electromagnetic calorimeter, and the K-long and muon (KLM) detector embedded in the iron yoke...
Kael Hanson
(Université Libre de Bruxelles)
10/14/13, 2:10 PM
Data acquisition, trigger and controls
Oral presentation to parallel session
The IceCube Neutrino Observatory is a cubic kilometer-scale neutrino detector built into the ice sheet at the geographic South Pole. The online system for IceCube comprises subsystems for data acquisition, online filtering, supernova detection, and experiment control and monitoring. The observatory records astrophysical and cosmic ray events at a rate of approximately 3 kHz and selects the...
Simone Campana
(CERN)
10/14/13, 2:14 PM
Distributed Processing and Data Handling B: Experiment Data Processing, Data Handling and Computing Models
Oral presentation to parallel session
The ATLAS Distributed Computing project (ADC) was established in 2007 to
develop and operate a framework, following the ATLAS computing model, to enable
data storage, processing and bookkeeping on top of the WLCG distributed
infrastructure. ADC development has always been driven by operations and this
contributed to its success. The system has fulfilled the demanding requirements of...
Mr
Alexandr Zaytsev
(Brookhaven National Laboratory (US)), Mr
Kevin CASELLA
(Brookhaven National Laboratory (US))
10/14/13, 2:14 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Oral presentation to parallel session
RHIC & ATLAS Computing Facility (RACF) at BNL is a 15000 sq. ft. facility hosting the IT equipment of the BNL ATLAS WLCG Tier-1 site, offline farms for the STAR and PHENIX experiments operating at the Relativistic Heavy Ion Collider (RHIC), BNL Cloud installation, various Open Science Grid (OSG) resources, and many other small physics research oriented IT installations. The facility originated...
Gerardo Ganis
(CERN)
10/14/13, 2:15 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Oral presentation to parallel session
The advent of private and commercial cloud platforms has opened the question of evaluating the cost-effectiveness of such solution for computing in High Energy Physics .
Google Compute Engine (GCE) is a IaaS product launched by Google as an experimental platform during 2012 and now open to the public market.
In this contribution we present the results of a set of CPU-intensive and...
Xavier Espinal Curull
(CERN)
10/14/13, 2:16 PM
Data Stores, Data Bases, and Storage Systems
Oral presentation to parallel session
Data Storage and Services (DSS) group at CERN stores and provides access to the data coming from the LHC and other physics experiments. We implement specialized storage services to provide tools for an optimal data management, based on the evolution of data volumes, the available technologies and the observed experiment and users usage patterns. Our current solutions are CASTOR for...
Qiming Lu
(Fermi National Accelerator Laboratory)
10/14/13, 2:16 PM
Software Engineering, Parallelism & Multi-Core
Oral presentation to parallel session
Synergia is a parallel, 3-dimensional space-charge particle-in-cell code that is widely used by the accelerator modeling community. We present our work of porting the pure MPI-based code to a hybrid of CPU and GPU computing kernels. The hybrid code uses the CUDA platform, in the same framework as the pure MPI solution. We have implemented a lock-free collaborative charge-deposition algorithm...
Kunihiro Nagano
(High Energy Accelerator Research Organization (JP))
10/14/13, 2:30 PM
Data acquisition, trigger and controls
Oral presentation to parallel session
The ATLAS trigger system has been used for the online event selection for three years of LHC data-taking and is preparing for the next run. The trigger system consists of a hardware level-1 (L1) and a software high-level trigger (HLT). The high-level trigger is currently implemented in a region-of-interest based level-2 (L2) stage and a event filter (EF) operating after even building with...
Claudio Grandi
(INFN - Bologna)
10/14/13, 2:36 PM
Distributed Processing and Data Handling B: Experiment Data Processing, Data Handling and Computing Models
Oral presentation to parallel session
The CMS Computing Model was developed and documented in 2004. Since then the model has evolved to be more flexible and to take advantage of new techniques, but many of the original concepts remain and are in active use. In this presentation we will discuss the changes planned for the restart of the LHC program in 2015. We will discuss the changes planning in the use and definition of the...
Dr
Tony Wong
(Brookhaven National Laboratory)
10/14/13, 2:36 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Oral presentation to parallel session
The advent of cloud computing centers such as Amazon's EC2 and Google's Computing Engine has elicited comparisons with dedicated computing clusters. Discussions on appropriate usage of cloud resources (both academic and commercial) and costs have ensued. This presentation discusses a detailed analysis of the costs of operating and maintaining the RACF (RHIC and ATLAS Computing Facility)...
Dr
Jerome LAURET
(BROOKHAVEN NATIONAL LABORATORY)
10/14/13, 2:36 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Oral presentation to parallel session
User Centric Monitoring (or UCM) has been a long awaited feature in STAR, whereas programs, workflows and system “events” could be logged, broadcast and later analyzed. UCM allows to collect and filter available job monitoring information from various resources and present it to users in a user-centric view rather than and administrative-centric point of view. The first attempt and...
Christos Filippidis
(Nat. Cent. for Sci. Res. Demokritos (GR))
10/14/13, 2:39 PM
Data Stores, Data Bases, and Storage Systems
Oral presentation to parallel session
Given the current state of I/O and storage systems in petascale systems, incremental solutions in most aspects are unlikely to provide the required capabilities in exascale systems. Traditionally I/O has been considered as a separate activity that is performed before or after the main simulation or analysis computation, or periodically for activities such as check-pointing, but still as...
Dr
Tareq AbuZayyad
(University of Utah)
10/14/13, 2:39 PM
Software Engineering, Parallelism & Multi-Core
Oral presentation to parallel session
The Telescope Array Cosmic Rays Detector located in the Western Utah Desert is used for the observation of ultra-high energy cosmic rays. The simulation of a fluorescence detector response to cosmic rays initiated air showers presents many opportunities for parallelization. In this presentation we report on the Monte Carlo program used for the simulation of the Telescope Array fluorescence...
Slava Krutelyov
(Texas A & M University (US))
10/14/13, 2:40 PM
Event Processing, Simulation and Analysis
Oral presentation to parallel session
In 2012 the LHC increased both the beam energy and intensity. The former made obsolete all of the simulation data generated for 2011; the latter increased the rate of multiple proton-proton collisions (piluep) in a single event, significantly increasing the complexity of both the reconstructed and matching simulated events. Once the pileup surpassed 10, the resources needed for the software to...
Gero Müller
(RWTH Aachen University)
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
Many programs in experimental particle physics do not yet have a graphical interface, or demand strong platform and software requirements. With the most recent development of the VISPA project, we provide graphical interfaces to existing software programs and access to multiple computing clusters through standard web browsers. The scalable client-server system allows analyses to be performed...
John Bland
(University of Liverpool)
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
Liverpool is consistently amongst the top Tier-2 sites in Europe in terms of efficiency and cluster utilisation. This presentation will cover the work done at Liverpool over the last six years to maximise and maintain efficiency and productivity at their Tier 2 site, with an overview of the tools used (including established, emerging, and locally developed solutions) for monitoring, testing,...
Philipp Sitzmann
(Goethe University Frankfurt)
10/14/13, 3:00 PM
CMOS Monolithic Active Pixel Sensors (MAPS) have demonstrated excellent performances as tracking detectors for charged particles. Their outstanding spatial resolution (few µm), ultra-light material budget (50 µm) and advanced radiation tolerance (> 1Mrad, >1e13 neq/cm²). They were therefore chosen for the vertex detectors of STAR and CBM and are foreseen to equip the upgraded ALICE-ITS. They...
Vincenzo Spinoso
(Universita e INFN (IT)),
Vincenzo Spinoso
(Universita e INFN (IT))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
Running and monitoring simulations usually involves several different aspects of the entire workflow: the configuration of the job, the site issues, the software deployment at the site, the file catalogue, the transfers of the simulated data. In addition, the final product of the simulation is often the result of several sequential steps. This project tries a different approach to monitoring...
Daniel Hugo Campora Perez
(CERN)
10/14/13, 3:00 PM
The LHCb Software Infrastructure is built around a flexible, extensible, single-process, single-threaded framework named Gaudi. One way to optimise the overall usage of a multi-core server, which is used for example in the Online world, is running multiple instances of Gaudi-based applications concurrently. For LHCb, this solution has been shown to work well up to 32 cores and is expected...
Andrea Formica
(CEA/IRFU,Centre d'etude de Saclay Gif-sur-Yvette (FR))
10/14/13, 3:00 PM
The ATLAS muon alignment system is composed of about 6000 optical sensors for the Barrel muon spectrometer and the same number for the 2 Endcaps wheels.
The system is acquiring data from every sensor continuously , with a whole read-out cycle of about 10 minutes. The read-out chain stores data inside an Oracle DB. These data are used as input from the alignment algorithms (C++ based) in...
Mr
MA Binsong
(IPN Orsay France)
10/14/13, 3:00 PM
The PANDA (AntiProton ANnihilation at DArmstadt) experiment is one of the key projects at the future Facility for Antiproton and Ion Research (FAIR), which is currently under construction at Darmstadt. This experiment will perform precise studies of antiproton-proton and antiproton-nucleus annihilation reactions. The aim of the rich experimental program is to improve our knowledge of the...
Mr
Peter Waller
(University of Liverpool (GB))
10/14/13, 3:00 PM
The focus in many software architectures of the LHC experiments is to deliver a well-designed Event Data Model (EDM). Changes and additions to the stored data are often very expensive, requiring large amounts of CPU time, disk storage and man-power. At the ATLAS experiment, such a reprocessing has only been undertaken once for data taken in 2012.
However, analysts have to develop and apply...
Alessandro De Salvo
(Universita e INFN, Roma I (IT))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
In the Atlas experiment, the calibration of the precision tracking chambers of the
muon detector is very demanding, since the rate of muon tracks required to get a
complete calibration in homogeneous conditions and to feed prompt reconstruction
with fresh constants is very high (several hundreds Hz for 8-10 hours runs). The
calculation of calibration constants is highly CPU consuming. In...
Dr
Salman Toor
(Helsinki Institute of Physics (FI))
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
The challenge of providing a resilient and scalable computational and data management solution for massive scale research environments, such as the CERN HEP analyses, requires continuous exploration of new technologies and techniques. In this article we present a hybrid solution of an open source cloud with a network file system for CMS data analysis. Our aim has been to design a scalable and...
Andrea Formica
(CEA/IRFU,Centre d'etude de Saclay Gif-sur-Yvette (FR))
10/14/13, 3:00 PM
ATLAS Conditions data include about 2 TB in a relational database and 400 GB of files referenced from the database. Conditions data is entered and retrieved using COOL, the API for accessing data in the LCG Conditions Database infrastructure. It is managed using an ATLAS-customized python based tool set.
Conditions data are required for every reconstruction and simulation job, so access to...
Gareth Roy
(U),
Mark Mitchell
(University of Glasgow)
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
With the current trend towards "On Demand Computing" in big data environments it becomes crucial that the deployment of services and resources becomes increasingly automated. With opensource projects such as Canonicals MaaS and Redhats Spacewalk; automated deployment is available for large scale data centre environments but these solutions can be too complex and heavyweight for smaller,...
Derek John Weitzel
(University of Nebraska (US))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
Bosco is a software project developed by the Open Science Grid to help scientists better utilize their on-campus computing resources. Instead of submitting jobs through a dedicated gatekeeper, as most remote submission mechanisms use, it uses the built-in SSH protocol to gain access to the cluster. By using a common access method, SSH, we are able to simplify the interaction with the...
Alexey Anisenkov
(Budker Institute of Nuclear Physics (RU))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
In this paper we describe the ATLAS Grid Information System (AGIS), the system designed to integrate configuration and status information about resources, services and topology of the computing infrastructure used by ATLAS Distributed Computing (ADC) applications and services.
The Information system centrally defines and exposes the topology of the ATLAS computing infrastructure including...
Ian Collier
(UK Tier1 Centre), Mr
Matthew James Viljoen
(STFC - Science & Technology Facilities Council (GB))
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
In this paper we shall introduce the service deployment framework based on Quattor and Microsoft HyperV at the RAL Tier 1. As an example, we will explain how the framework has been applied to CASTOR in our test
infrastructure and outline our plans to roll it out into full production. CASTOR is a relatively complicated open source hierarchical storage management system in production use at...
Qiyan Li
(Goethe University Frankfurt)
10/14/13, 3:00 PM
CBM aims to measure open charm particles from 15-40 AGeV/c heavy ion collisions by means of secondary vertex reconstruction. The measurement concept includes the use of a free-running DAQ, real time tracking, primary and secondary vertex reconstruction and a tagging of open charm candidates based on secondary vertex information. The related detector challenge will be adressed with an...
Fabrizio Furano
(CERN)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
In this contribution we present a vision for the use of the HTTP protocol for data management in the context of HEP, and we present demonstrations of the use of HTTP-based protocols for storage access & management, cataloguing, federation and transfer.
The support of HTTP/WebDAV, provided by frameworks for scientific data access like DPM, dCache, STORM, FTS3 and foreseen for XROOTD, can be...
Francesco Giacomini
(INFN CNAF)
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
The success of a scientific endeavor depends, often significantly, on
the ability to collect and later process large amounts of data in an
efficient and effective way. Despite the enormous technological
progress in areas such as electronics, networking and storage, the
cost of the computing factor remains high. Moreover the limits reached
by some historical directions of hardware...
Carlos Solans Sanchez
(CERN)
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
The Tile calorimeter is one of the sub-detectors of ATLAS. In order to ensure its proper operation and assess the quality of data, many tasks are to be performed by means of many tools which were developed independently to satisfy different needs. Thus, these systems are commonly implemented without a global perspective of the detector and lack basic software features. Besides, in some cases...
Dr
Giacinto Donvito
(INFN-Bari)
10/14/13, 3:00 PM
The italian community in CMS has built a geographically distributed network in which all the data stored in the italian region are available to all the users for their everyday work. This activity involves at different level all the CMS centers: the Tier1 at CNAF, all the four Tier2s (Bari, Rome, Legnaro and Pisa), and few Tier3s (Trieste, Perugia, etc). The federation uses the new network...
Dr
Samuel Cadellin Skipsey
10/14/13, 3:00 PM
Of the three most widely used implementations of the WLCG Storage Element specification, Disk Pool Manager (DPM) has the simplest implementation of file placement balancing (StoRM doesn't attempt this, leaving it up to the underlying filesystem, which can be very sophisticated in itself). DPM uses a round-robin algorithm (with optional filesystem weighting), for placing files across...
Shaun De Witt
(STFC - Science & Technology Facilities Council (GB))
10/14/13, 3:00 PM
At the RAL Tier 1 we have successfully been running a CASTOR HSM instance for a number of years. While it performs well for disk-only storage for analysis and processing jobs, it is heavily optimised for tape usage. We have been investigating alternative technologies which could be used for online storage for analysis. We present the results of our preliminary selection and test results for...
Dr
Massimiliano Nastasi
(INFN Milano-Bicocca)
10/14/13, 3:00 PM
Measurements of radioactive sources, in order to reach an optimum level of accuracy, require an accurate determination of the detection efficiency of the experimental setup. In gamma ray spectroscopy, in particular, the high level of sensitivity reached nowadays implies a correct evaluation of the detection capability of source emitted photons. The standard approach, based
on an analytical...
David Cameron
(University of Oslo (NO))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
Grid middleware suites provide tools to perform the basic tasks of job
submission and retrieval and data access, however these tools tend to
be low-level, operating on individual jobs or files and lacking in
higher-level concepts. User communities therefore generally develop
their own application-layer software catering to their specific
communities' needs on top of the Grid middleware....
Dr
Roberto Ammendola
(INFN Roma Tor Vergata)
10/14/13, 3:00 PM
Modern Graphics Processing Units (GPUs) are now considered accelerators for general purpose computation. A tight interaction between the GPU and the interconnection network is the strategy to express the full potential on capability computing of a multi-GPU system on large HPC clusters; that is why an efficient and scalable interconnect is a key technology to finally deliver GPUs for...
Dr
Jörg Meyer
(KIT - Karlsruher Institute of Technology)
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
After analysis and publication, there is no need to keep experimental data online on spinning disks. For reliability and costs inactive data is moved to tape and put into a data archive. The data archive must provide reliable access for at least ten years following a recommendation of the German Science Foundation (DFG), but many scientific communities wish to keep data available much longer....
Jaroslava Schovancova
(Brookhaven National Laboratory (US))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
The ATLAS Distributed Computing (ADC) Monitoring targets three groups of customers: ADC Operations, ATLAS Management, and ATLAS sites and ATLAS funding agencies. The main need of ADC Operations is to identify malfunctions early and then escalate issues to an activity or a service expert. The ATLAS Management use visualisation of long-term trends and accounting information about the ATLAS...
Alexey Sedov
(Universitat Autònoma de Barcelona)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
ATLAS Distributed Computing Operation Shifts were evolved to meet new requirements. New monitoring tools as well as new operational changes led to modifications in organization of shifts. In this paper we describe the roles and the impacts of the shifts to smooth operation of complex computing grid employed in ATLAS, the influence of Discovery of Higgs like particle on shift operations, the...
Dr
Alexander Undrus
(Brookhaven National Laboratory (US))
10/14/13, 3:00 PM
The ATLAS Nightly Build System is a facility for automatic production of software releases. Being the major component of ATLAS software infrastructure, it supports more than 50 multi-platform branches of nightly releases and provides vast opportunities for testing new packages, for verifying patches to existing software, and for migrating to new platforms and compilers. The Nightly System...
Grigori Rybkin
(Universite de Paris-Sud 11 (FR))
10/14/13, 3:00 PM
The ATLAS software code base is over 7 million lines organised in about 2000 packages. It makes use of some 100 external software packages, is developed by more than 400 developers and used by more than 2500 physicists from over 200 universities and laboratories in 6 continents. To meet the challenge of configuration and building of this software, the Configuration Management Tool (CMT) is...
Jason Alexander Smith
(Brookhaven National Laboratory (US))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
Public clouds are quickly becoming cheap and easy methods for dynamically adding more computing resources to your local site to help handle peak computing demands. As cloud use continues to grow, the HEP community is looking to run more than just simple homogeneous VM images, which run basic data analysis batch jobs. The growing demand for heterogeneous server configurations demands better...
Jason Alexander Smith
(Brookhaven National Laboratory (US))
10/14/13, 3:00 PM
Running a stable production service environment is important in any field. To accomplish this, a proper configuration management system is necessary along with good change management policies. Proper testing and validation is required to protect yourself against software or configuration changes to production services that can cause major disruptions. In this paper, we discuss how we extended...
Dr
Jorge Luis Rodriguez
(UNIVERSITY OF FLORIDA)
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
With the explosion of big data in many fields, the efficient
management of knowledge about all aspects of the data analysis gains
in importance. A key feature of collaboration in large scale projects
is keeping a log of what and how is being done - for private use and
reuse and for sharing selected parts with collaborators and peers,
often distributed geographically on an increasingly...
John Hover
(Brookhaven National Laboratory (BNL)-Unknown-Unknown)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
AutoPyFactory (APF) is a next-generation pilot submission framework that has been used as part of the ATLAS workload management system (PanDA) for two years. APF is reliable, scalable, and offers easy and flexible configuration. Using a plugin-based architecture, APF polls for information from configured information and batch systems (including grid sites), decides how many additional pilot...
Ludmila Marian
(CERN)
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
The volume of multimedia material produced by CERN is growing rapidly, fed by the increase of dissemination activities carried out by the various outreach teams, such as the central CERN Communication unit and the Experiments Outreach committees. In order for this multimedia content to be stored digitally for the long term, to be made available to end-users in the best possible conditions and...
Ian Peter Collier
(STFC - Science & Technology Facilities Council (GB))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
In the last three years the CernVM Filesystem (CernVM-FS) has transformed the distribution of experiment software to WLCG grid sites. CernVM-FS removes the need for local installations jobs and performant network fileservers at sites, in addition it often improves performance at the same time. Furthermore the use of CernVM-FS standardizes the computing environment across the grid and removes...
Stefano Dal Pra
(Unknown)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
At the Italian Tier1 Center at CNAF we are evaluating the possibility to change the current production batch system. This activity is motivated mainly because we are looking for a more flexible licensing model as well as to avoid vendor lock-in.
We performed a technology tracking exercise and among many possible solutions we chose to evaluate Grid Engine as an alternative because its...
Victor Manuel Fernandez Albor
(Universidade de Santiago de Compostela (ES))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
Communities of different locations are running their computing jobs on dedicated infrastructures without the need to worry about software, hardware or even the site where their programs are going to be executed. Nevertheless, this usually implies that they are restricted to use certain types or versions of an Operating System because either their software needs an definite version of a system...
Kenneth Bloom
(University of Nebraska (US))
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
To impart hands-on training in physics analysis, CMS experiment initiated the concept of CMS Data Analysis School (CMSDAS). It was born three years ago at the LPC (LHC Physics Center), Fermilab and is based on earlier workshops held at the LPC and CLEO Experiment. As CMS transitioned from construction to the data taking mode, the nature of earlier training also evolved to include more of...
Mr
Igor Sfiligoi
(University of California San Diego)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
The CMS experiment at the Large Hadron Collider is relying on the HTCondor-based glideinWMS batch system to handle most of its distributed computing needs. In order to minimize the risk of disruptions due to software and hardware problems, and also to simplify the maintenance procedures, CMS has set up its glideinWMS instance to use most of the attainable High Availability (HA) features. The...
Mrs
Ianna Osborne
(Fermi National Accelerator Lab. (US))
10/14/13, 3:00 PM
CMS faces real challenges with upgrade of the CMS detector through 2020. One of the challenges, from the software point of view, is managing upgrade simulations with the same software release as the 2013 scenario. We present the CMS geometry description software model, its integration with the CMS event setup and core software. The CMS geometry configuration and selection is implemented in...
Dr
Tony Wildish
(Princeton University (US))
10/14/13, 3:00 PM
During the first LHC run, CMS saturated one hundred petabytes of storage resources with data. Storage accounting and monitoring help to meet the challenges of storage management, such as efficient space utilization, fair share between users and groups, and further resource planning. We present newly developed CMS space monitoring system based on the storage dumps produced at the sites. Storage...
111.
CMS users data management service integration and first experiences with its NoSQL data storage
Marco Mascheroni
(Universita & INFN, Milano-Bicocca (IT))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
The distributed data analysis workflow in CMS assumes that jobs run in a different location to where their results are finally stored. Typically the user outputs must be transferred from one site to another by a dedicated CMS service, AsyncStageOut. This new service is originally developed to address the inefficiency in using the CMS computing resources when transferring the analysis job...
Dr
Edward Karavakis
(CERN)
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
The ATLAS Experiment at the Large Hadron Collider has been collecting data for three years. The ATLAS data are distributed, processed and analysed at more than 130 grid and cloud sites across the world. The total throughput of transfers is more than 5 GB/s and data occupies more than 120 PB on disk and tape storage. At any given time, there are more than 100,000 concurrent jobs running and...
Moritz Kretz
(Ruprecht-Karls-Universitaet Heidelberg (DE))
10/14/13, 3:00 PM
In 2014 the Insertable B-Layer (IBL) will extend the existing Pixel Detector of the ATLAS experiment at CERN by 12 million additional pixels. As with the already existing pixel layers, scanning and tuning procedures need to be employed for the IBL to account for aging effects and guarantee a unified response across the detector. Scanning the threshold or time-over-threshold of a front-end...
Carlos Solans Sanchez
(CERN)
10/14/13, 3:00 PM
After two years of operation of the LHC, the ATLAS Tile Calorimeter is undergoing the consolidation process of its front-end electronics. The first layer of certification of the repairs is performed in the experimental area with a portable test-bench which is capable of controlling and reading out all the inputs and outputs of one front-end module through dedicated cables. This testbench has...
Line Everaerts
(CERN)
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
Using the framework of ITIL best practises, the service managers within CERN-IT have engaged into a continuous improvement process, mainly focusing on service operation. This implies an explicit effort to understand and improve all service management aspects in order to increase efficiency and effectiveness. We will present the requirements, how they were addressed and share our experiences....
Mr
Hiroyuki Maeda
(Hiroshima Institute of Technology)
10/14/13, 3:00 PM
DAQ-Middleware is a software framework for a network-distributed data acquisition (DAQ) system that is based on the Robot Technology Middleware (RTM). The framework consists of a DAQ-Component and a DAQ-Operator. The basic functionalities such as transferring data, starting and stopping the system, and so on, are already prepared in the DAQ-Components and DAQ-Operator. The DAQ-Component is...
Ruslan Asfandiyarov
(Universite de Geneve (CH)),
Yordan Ivanov Karadzhov
(Universite de Geneve (CH))
10/14/13, 3:00 PM
The Electron-Muon Ranger (EMR) is a totally active scintillator detector which will be installed in the muon beam of the Muon Ionization Cooling Experiment (MICE), the main R&D project for a future neutrino factory. It is designed to measure the properties of a low energy beam composed of muons, electrons and pions, and to perform an identification on a particle by particle basis. The EMR is...
Evan Niner
(Indiana University), Mr
Zukai Wang
(University of Virginia)
10/14/13, 3:00 PM
The NOvA experiment at Fermi National Accelerator Lab, due to its unique readout and buffering design, is capable of accessing physics beyond the core neutrino oscillations program for which it was built. In particular the experiment is able to search for evidence of relic cosmic magnetic monopoles and for the signs of the neutrino flash from a near by supernova through uses of a specialized...
katarzyna wichmann
(DESY)
10/14/13, 3:00 PM
The data preservation project at DESY was established in 2008, shortly after data taking ended at the HERA ep collider, soon after coming under the umbrella of the DPHEP global initiative. All experiments are implementing data preservation schemes to allow long term analysis of their data, in cooperation with the DESY-IT division. These novel schemes include software validation and...
Dr
Bodhitha Jayatilaka
(Fermilab)
10/14/13, 3:00 PM
The Fermilab Tevatron collider's data-taking run ended in September 2011, yielding a dataset with rich scientific potential. The CDF experiment has nearly 9 PB of collider and simulated data stored on tape. A large computing infrastructure consisting of tape storage, disk cache, and distributed grid computing for physics analysis with the CDF data is present at Fermilab.
The Fermilab Run II...
Dr
Michael Kirby
(Fermi National Accelerator Laboratory)
10/14/13, 3:00 PM
The Tevatron experiments have entered their post-data-taking phases but are still producing physics output at a high rate.
The D0 experiment has initiated efforts to preserve both data access and full analysis capability for the collaboration members through at least 2020. These efforts will provide useful lessons in ensuring long-term data access for numerous experiments throughout...
Mr
Tao Lin
(Institute of High Energy Physics)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
Data Transfer is an essential part in grid. In the BESIII experiment, the result of Monte Carlo Simulation should be transfered back from other sites to IHEP and the DST files for physics analysis should be tranfered from IHEP to other sites. A robust transfer system should
make sure all data are transfered correctly.
DIRAC consists of cooperation distributed services and light-weight...
Kai Leffhalm
(Deutsches Elektronen-Synchrotron (DE))
10/14/13, 3:00 PM
The dCache storage system writes billing data into flat files or a relational database.
For a midsize dCache installation there are one million entries - representing 300 MByte - per day.
Gathering accounting information for a longer time interval about transfer rates per group, per file type or per user results in increasing load on the servers holding the billing information.
Speeding up...
Gancho Dimitrov
(CERN)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
The ATLAS experiment at CERN is one of the four Large Hadron Collider
experiments. The DCS Data Viewer (DDV) is an application that provides
access to historical data of the ATLAS Detector Control System (DCS)
parameters and their corresponding alarm information. It features a
server-client architecture: the pythonic server serves as interface to
the Oracle-based conditions database and...
Andreas Petzold
(KIT)
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
GridKa, the German WLCG Tier-1 site hosted by Steinbuch Centre for Computing at Karlsruhe Institute of Thechnology, is a collaboration partner in the HEPIX-IPv6 testbed. A special IPv6-enabled gridftp server has been installed previously. In 2013, the IPv6 efforts will be increased. Already the installation of a new Mini-Grid site has been started. This Mini-Grid installation is planned as a...
Franco Brasolin
(Universita e INFN (IT))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
With the LHC collider at CERN currently going through the period of Long Shutdown 1 (LS1) there is a remarkable opportunity to use the computing resources of the large trigger farms of the experiments for other data processing activities.
In the case of ATLAS experiment the TDAQ farm, consisting of more than 1500 compute nodes, is particularly suitable for running Monte Carlo production jobs...
Tai Sakuma
(Texas A & M University (US))
10/14/13, 3:00 PM
We describe the creation of 3D models of the CMS detector and events using SketchUp, a 3D modelling program. SketchUp provides a Ruby API with which we interface with the CMS Detector Description, the master source of the CMS detector geometry, to create detailed 3D models of the CMS detector. With the Ruby API we also interface with the JSON-based event format used for the iSpy event display...
Sergey Belogurov
(ITEP Institute for Theoretical and Experimental Physics (RU))
10/14/13, 3:00 PM
Detector geometry exchange between CAD systems and physical Monte Carlo (MC), packages ROOT and Geant4 is a labor-consuming process necessary for fine design optimization. CAD and MC geometries have completely different structure and hierarchy. For this reason automatic conversion is possible only for very simple shapes.
CATIA-GDML Geometry Builder is a tool which allows to facilitate...
Xavier Espinal Curull
(CERN)
10/14/13, 3:00 PM
After the strategic decision in 2011 to separate tier-0 activity from analysis, CERN-IT developed EOS as a new petascale disk-only solution to address the fast-growing needs for high-performance low-latency data access. EOS currently holds around 22PB usable space for the four big experiment (ALICE, ATLAS, CMS, LHCb), and we expect to grow to >30PB this year. EOS is one of the first production...
Luisa Arrabito
(LUPM Université Montpellier 2, IN2P3/CNRS)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
DIRAC (Distributed Infrastructure with Remote Agent Control) is a general framework for the management of tasks over distributed heterogeneous computing environments. It has been originally developed to support the production activities of the LHCb (Large Hadron Collider Beauty) experiment and today is extensively used by several particle physics and biology communities. Current (Fermi-LAT,...
Dr
Armando Fella
(INFN Pisa), Mr
Bruno Santeramo
(INFN Bari),
Cristian De Santis
(Universita degli Studi di Roma Tor Vergata (IT)), Dr
Giacinto Donvito
(INFN-Bari),
Marcin Jakub Chrzaszcz
(Polish Academy of Sciences (PL)), Mr
Milosz Zdybal
(Institute of Nuclear Physics, Polish Academy of Science),
Rafal Zbigniew Grzymkowski
(P)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
In HEP computing context, R&D studies aiming to the definition of the data and workload models were brought forward by the SuperB community beyond the experiment life itself.
This work is considered of great interest for a generic mid- and small size VO to fulfil Grid exploiting requirements involving CPU-intensive tasks.
We present the R&D line achievements in the design, developments...
Dr
Tony Wong
(Brookhaven National Laboratory)
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
The RHIC and ATLAS Computing Facility (RACF) at Brookhaven Lab is a dedicated data center serving the needs of the RHIC and US ATLAS community. Since it began
operations in the mid-1990's, it has operated continuously with few unplanned downtimes. In the last 24 months, Brookhaven Lab has been affected by two hurricanes and a record-breaking snowstorm. In
this presentation, we discuss...
Justin Lewis Salmon
(University of the West of England (GB))
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
The Extended ROOT Daemon (XRootD) is a distributed, scalable system for low-latency clustered data access. XRootD is mature and widely used in HEP, both standalone and as core functionality for the EOS system at CERN, and hence requires extensive testing to ensure general stability. However, there are many difficulties posed by distributed testing, such as cluster initialization,...
Stefano Piano
(INFN (IT))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
Since 2003 the computing farm hosted by the INFN T3 facility in Trieste supports the activities of many scientific communities. Hundreds of jobs from 45 different VOs, including those of the LHC experiments, are processed simultaneously. The currently available shared disk space amounts to about 300 TB, while the computing power is provided by 712 cores for a total of 7400 HEP-SPEC06. Given...
Dr
Jorge Luis Rodriguez
(UNIVERSITY OF FLORIDA)
10/14/13, 3:00 PM
We have developed remote data access for large volumes of data over the Wide Area Network based on the Lustre filesystem and Kerberos authentication for security. It this paper we explore a prototype for two-step data access from worker nodes at Florida T3 centers, located behind a firewall and using a private network, to data hosted on the Lustre filesystem at the University of Florida CMS T2...
Ian Gable
(University of Victoria (CA))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
It has been shown possible to run HEP workloads on remote IaaS cloud resources. Typically each running Virtual Machine (VM) makes use of the CERN VM Filesystem (CVMFS), a caching HTTP file system, to minimize the size of the VM images, and to simplify software installation. Each VM must be configured with a HTTP web cache, usually a Squid Cache, in proximity in order to function efficiently....
Dr
raul lopes
(School of Design and Engineering - Brunel University, UK)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
The performance of hash function computations can impose a significant workload on SSL/TLS authentication servers. In the WLCG this workload shows also in the computation of data transfers checksums. It has been shown in the EGI grid infrastructure that the checksum computation can double the IO load for large file transfers leading to an increase in re-transfers and timeout errors. Storage...
Tomas Kouba
(Acad. of Sciences of the Czech Rep. (CZ))
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
The production usage of the new IPv6 protocol is becoming reality in the HEP community and the Computing Centre of the Institute of Physics in Prague participates in many IPv6 related activities. Our contribution will present experience with monitoring in HEPiX
distributed IPv6 testbed which includes 11 remote sites. We use Nagios
to check availability of services and Smokeping for...
Mr
Igor Sfiligoi
(University of California San Diego)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
The basic premise of pilot systems is to create an overlay scheduling system on top of leased resources. And by definition, leases have a limited lifetime, so any job that is scheduled on such resources must finish before the lease is over, or it will be killed and all the computation wasted. In order to effectively schedule jobs to resources, the pilot system thus requires the expected...
Ian Fisk
(Fermi National Accelerator Lab. (US))
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
The Fermilab CMS Tier-1 facility provides processing, networking, and storage as one of seven Tier-1 facilities for the CMS experiment. The storage consists of approximately 15 PB of online/nearline disk managed by the dCache file system, and 22 PB of tape managed by the Enstore mass storage system. Data is transferred to and from computing centers worldwide using the CMS-developed PhEDEx...
Guenter Duckeck
(Experimentalphysik-Fakultaet fuer Physik-Ludwig-Maximilians-Uni), Dr
Johannes Ebke
(Ludwig-Maximilians-Univ. Muenchen (DE)),
Sebastian Lehrack
(LMU Munich)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
The Apache Hadoop software is a Java based framework for distributed
processing of large data sets across clusters of computers using the
Hadoop file system (HDFS) for data storage and backup and MapReduce as a processing platform.
Hadoop is primarily designed for processing large textual data sets
which can be processed in arbitrary chunks, and must be adapted to the use case of...
Ian Fisk
(Fermi National Accelerator Lab. (US))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
The physics event reconstruction in LHC/CMS is one of the biggest challenges for computing.
Among the different tasks that computing systems perform, the reconstruction takes most of the CPU resources that are available. The reconstruction time of a single event varies according to the event complexity. Measurements were done in order to find precisely this correlation, creating means to...
Ian Fisk
(Fermi National Accelerator Lab. (US))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
CMS production and analysis job submission is based largely on glideinWMS and pilot submissions. The transition from multiple different submission solutions like gLite WMS and HTCondor-based implementations was carried out over years and is coming now to a conclusion. The historically explained separate glideinWMS pools for different types of production jobs and analysis jobs are being unified...
Prof.
Jesus Marco
(IFCA (CSIC-UC) Santander Spain)
10/14/13, 3:00 PM
The strategy at the end of the LEP era for the long term
preservation of physics results and data processing framework
was not obvious.
One of the possibilities analyzed at the time, previously to the
generalization of virtualization techniques, was the setup of
a dedicated farm, to be conserved in its original state for
medium-long term, at least until the new data from LHC could...
Wim Lavrijsen
(Lawrence Berkeley National Lab. (US))
10/14/13, 3:00 PM
Intel recently released the first commercial boards of its Many Integrated Core (MIC) Architecture. MIC is Intel's solution for the domain of throughput computing, currently dominated by general purpose programming on graphics processors (GPGPU). MIC allows the use of the more familiar x86 programming model and supports standard technologies such as OpenMP, MPI, and Intel's Threading Building...
Laura Sargsyan
(ANSL (Yerevan Physics Institute) (AM))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
The organization of the distributed user analysis on the Worldwide LHC Computing Grid (WLCG) infrastructure is one of the most challenging tasks among the computing activities at the Large Hadron Collider. The Experiment Dashboard offers a solution that not only monitors but also manages (kill, resubmit) user tasks and jobs via a web interface. The ATLAS Dashboard Task Monitor provides...
Boris Wagner
(University of Bergen (NO) for the ALICE Collaboration)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
The Nordic Tier-1 for the LHC is distributed over several, sometimes smaller, computing centers. In order to minimize administration effort, we are interested in running different grid jobs over one common grid middleware. ARC is selected as the internal middleware in the Nordic Tier-1. The AliEn grid middleware, used by ALICE has a different design philosophy than ARC. In order to use most of...
Jakub Cerkala
(Technical University of Košice),
Slávka Jadlovská
(Department of Cybernetics and Artificial Intelligence, Faculty of Electrical Engineering and Informatics, Technical University of Košice)
10/14/13, 3:00 PM
ALICE Controls data produced by commercial SCADA system WINCCOA is
stored in ORACLE database on the private experiment network. The SCADA
system allows for basic access and processing of the historical data.
More advanced analysis requires tools like ROOT and needs therefore a
separate access method to the archives.
The present scenario expects that detector experts create simple...
Max Fischer
(KIT - Karlsruhe Institute of Technology (DE))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
The CMS collaboration is successfully using glideInWMS for managing grid resources within the WLCG project. The GlideIn mechanism with HTCondor underneath provides a clear separation of responsibilities between administrators operating the service and users utilizing computational resources.
German CMS collaborators (dCMS) have explored modern capabilities of the glideInWMS and aiming at...
Dennis Box
(F)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
The Fermilab Intensity Frontier Experiments use an integrated submission system known as FIFE-jobsub, part of the FIFE (Fabric for Frontier Experiments) initiative, to submit batch jobs to the Open Science Grid. FIFE-jobsub eases the burden on experimenters by integrating data transfer and site selection details in an easy to use and well documented format. FIFE-jobsub automates tedious...
Johannes Philipp Grohs
(Technische Universitaet Dresden (DE))
10/14/13, 3:00 PM
The readout of the trigger signals of the ATLAS Liquid Argon (LAr) calorimeters is foreseen to be upgraded in order to prepare for operation during the first high-luminosity phase of the Large Hadron Collider (LHC). Signals with improved spatial granularity are planned to be received from the detector by a Digitial Processing System (DPS) in ATCA technology and will be sent in real-time to the...
Laurent Garnier
(LAL-IN2P3-CNRS)
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
Geant4 application in a web browser
Geant4 is a toolkit for the simulation of the passage of particles through matter. The Geant4 visualization system supports many drivers including OpenGL, OpenInventor, HepRep, DAWN, VRML, RayTracer, gMocren and ASCIITree, with diverse and complementary functionalities.
Web applications have an increasing role in our work, and thanks to emerging...
Dr
Thomas Kittelmann
(European Spallation Source ESS AB)
10/14/13, 3:00 PM
The construction of the European Spallation Source ESS AB, which will become the worlds most powerful source of cold and thermal neutrons (meV scale), is about to begin in Lund, Sweden, breaking ground in 2014 and coming online towards the end of the decade. Currently 22 neutron-scattering instruments are planned as the baseline suite at the facility, and a crucial part of each such beam-line...
Prof.
Vladimir Ivantchenko
(CERN)
10/14/13, 3:00 PM
Electromagnetic physics sub-package of the Geant4 Monte Carlo toolkit is an important component of LHC experiment simulation and other Geant4 applications. In this work we present recent progress in Geant4 electromagnetic physics modeling, with an emphasis on the new refinements for the processes of multiple and single scattering, ionisation, high energy muon interactions, and gamma induced...
Aurelie Pascal
(CERN)
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
CERN has recently renewed its obsolete VHF firemen’s radio network and replaced it by a digital one based on TETRA technology. TETRA already integrates an outdoor GPS localization system, but it appeared essential to look for a solution to also locate TETRA users in CERN’s underground facilities.
The system which answers this problematic and which has demonstrated a good resistance to...
Oliver Keeble
(CERN)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
The GLUE 2 information schema is now fully supported in the production EGI/WLCG information system. However, to make the schema usable and allow clients to rely on the information it is important that the meaning of the published information is clearly defined, and that information providers and site configurations are validated to ensure as far as possible that what they publish is correct....
Dr
Yaodong CHENG
(Institute of High Energy Physics,Chinese Academy of Sciences)
10/14/13, 3:00 PM
Gluster file system adopts no metadata architecture, which theoretically eliminates both a central point of failure and a performance bottleneck of metadata server. Firstly, this talk will introduce gluster compared to lustre or hadoop. However, its some mechanisms are not so good in current version. For example, it has to read the extend attributes of all bricks to locate one file. And it is...
Dr
Sebastien Binet
(IN2P3/LAL)
10/14/13, 3:00 PM
Current HENP libraries and frameworks were written before multicore
systems became widely deployed and used.
From this environment, a 'single-thread' processing model naturally
emerged but the implicit assumptions it encouraged are greatly
impairing our abilities to scale in a multicore/manycore world.
Thanks to C++11, C++ is finally slowly catching up with regard to
concurrency...
Michelle Perry
(Florida State University)
10/14/13, 3:00 PM
The search for new physics has typically been guided by theoretical models with relatively few parameters. However, recently, more general models, such as the 19-parameter phenomenological minimal supersymmetric standard model (pMSSM), have been used to interpret data at the Large Hadron Collider. Unfortunately, due to the complexity of the calculations, the predictions of these models are...
Derek John Weitzel
(University of Nebraska (US))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
During the last decade, large-scale federated distributed infrastructures have continually developed and expanded. One of the crucial components of a cyber-infrastructure is an accounting service that collects data related to resource utilization and identity of users using resources. The accounting service is important for verifying pledged resource allocation per particular groups and users,...
Johannes Elmsheuser
(Ludwig-Maximilians-Univ. Muenchen (DE))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
With the exponential growth of LHC (Large Hadron Collider) data in 2012, distributed computing has become the established way to analyze collider data. The ATLAS grid infrastructure includes more than 130 sites worldwide, ranging from large national computing centers to smaller university clusters. HammerCloud was previously introduced with the goals of enabling VO- and site-administrators to...
Dr
Gabriele Garzoglio
(FERMI NATIONAL ACCELERATOR LABORATORY)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
Fermilab supports a scientific program that includes experiments and scientists located across the globe. To better serve this community, in 2004, the (then) Computing Division undertook the strategy of placing all of the High Throughput Computing (HTC) resources in a Campus Grid known as FermiGrid, supported by common shared services. In 2007, the FermiGrid Services group deployed a service...
Maria Dimou
(CERN)
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
In the Wordwide LHC Computing Grid (WLCG) project the Tier centres are of paramount importance for storing and accessing experiment data and for running the batch jobs necessary for experiment production activities.
Although Tier2 sites provide a significant fraction of the resources a non-availability of resources at the Tier0 or the Tier1s can seriously harm not only WLCG Operations but...
Steven Goldfarb
(University of Michigan (US))
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
On July 4, 2012, particle physics became a celebrity. Around 1,000,000,000 people (yes, 1 billion) saw rebroadcasts of two technical presentations announcing discovery of a new boson. The occasion was a joint seminar of the CMS and ATLAS collaborations, and the target audience were members of those collaborations plus interested experts in the field of particle physics. Yet, the world ate it...
Ramon Medrano Llamas
(CERN)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
The recent paradigm shift toward cloud computing in IT, and general interest in "Big Data" in particular, have demonstrated that the computing requirements of HEP are no longer globally unique. Indeed, the CERN IT department and LHC experiments have already made significant R&D investments in delivering and exploiting cloud computing resources. While a number of technical evaluations of...
Wahid Bhimji
(University of Edinburgh (GB))
10/14/13, 3:00 PM
“Big Data” is no longer merely a buzzword, but is business-as-usual in the private sector. High Energy Particle Physics is often cited as the archetypal Big Data use case, however it currently shares very little of the toolkit used in the private sector or other scientific communities.
We present the initial phase of a programme of work designed to bridge this technology divide by both...
Alex Mann
(Ludwig-Maximilians-Univ. Muenchen (DE)),
Alexander Mann
(Ludwig-Maximilians-Universität)
10/14/13, 3:00 PM
The ATLAS detector operated during the three years of the run 1 of the Large
Hadron Collider collecting information on a large number of proton-proton events.
One the most important results obtained so far is the discovery of one Higgs
boson. More precise measurements of this particle must be performed as well as
there are other very important physics topics still to be explored. One of...
Stefan Kluth
(Max-Planck-Institut fuer Physik (Werner-Heisenberg-Institut) (D)
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
We benchmarked an ARM Cortex A9 based server system with a four-core CPU running at 1.1 GHz. The system used Ubuntu 12.04 as operating system and the hepspec 2006 (HS06) benchmarking suite was compiled natively with gcc-4.4 on the system. The benchmark was run for various settings of the relevant gcc compiler options. We did not find significant influence from the compiler options on the...
Andre Sailer
(CERN),
Christian Grefe
(CERN),
Stephane Guillaume Poss
(Centre National de la Recherche Scientifique (FR))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
ILCDIRAC was initially developed in the context of the CLIC Conceptual
Design Report (CDR), published in 2012-2013. It provides a convenient interface for the mass production of the simulation events needed for the physics performance studies of the two detectors concepts considered, ILD and SID. It was since used in the ILC Detailed Baseline Detector (DBD) studies of the SID detector...
Dr
Alexei Strelchenko
(FNAL)
10/14/13, 3:00 PM
Lattice Quantum Chromodynamics (LQCD) simulations are critical for understanding the validity of the Standard Model and the results of the High-Energy and Nuclear Physics experiments. Major improvements in the calculation and prediction of physical observables, such as nucleon form factors or flavor singlet meson mass, require large amounts of computer resources, of the order of hundreds of...
Kati Lassila-Perini
(Helsinki Institute of Physics (FI))
10/14/13, 3:00 PM
Implementation of the CMS policy on long-term data preservation, re-use and open access has started. Current practices in providing data additional to published papers and distributing simplified data-samples for outreach are promoted and consolidated. The first measures have been taken for the analysis and data preservation for the internal use of the collaboration and for the open access to...
Enrico Mazzoni
(INFN-Pisa)
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
The INFN-Pisa Tier2 infrastructure is described, optimized not only for GRID CPU and Storage access, but also for a more interactive use of the resources in order to provide good solutions for the final data analysis step. The Data Center, equipped with about 5000 production cores, permits the use of
modern analysis techniques realized via advanced statistical tools (like RooFit and RooStat)...
Donato De Girolamo
(INFN CNAF), Mr
Lorenzo Chiarelli
(INFN CNAF), Mr
Stefano Zani
(INFN CNAF)
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
The computing models of HEP experiments, starting from the LHC ones, are facing an evolution with the relaxation of the data locality paradigm: the possibility of a job accessing data files over the WAN is becoming more and more common.
One of the key factors for the success of this change is the ability
to use the network in the most efficient way: in the best scenario,
the network...
Andrew Malone Melo
(Vanderbilt University (US))
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
The LHC experiments have always depended upon a ubiquitous, highly-performing network infrastructure to enable their global scientific efforts. While the experiments were developing their software and physical infrastructures, parallel development work was occurring in the networking communities responsible for interconnecting LHC sites. During the LHC's Long Shutdown \#1 (LS1) we have an...
Dr
Tony Wildish
(Princeton University (US))
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
The ever-increasing amount of data handled by the CMS dataflow and workflow management tools poses new challenges for cross-validation among different systems within CMS experiment at LHC. To approach this problem we developed an integration test suite based on the LifeCycle agent, a tool originally conceived for stress-testing new releases of PhEDEx, the CMS data-placement tool. The LifeCycle...
Ivana Hrivnacova
(Universite de Paris-Sud 11 (FR))
10/14/13, 3:00 PM
g4tools, that is originally part of the inlib and exlib packages [1], provides a very light and easy to install set of C++ classes that can be used to perform analysis in a Geant4 batch program. It allows to create and manipulate histograms and ntuples, and write them in supported file formats (ROOT, AIDA XML, CSV and HBOOK).
It is integrated in Geant4 through analysis manager classes,...
Dmitry Nilsen
(Karlsruhe Institute of Technology), Dr
Pavel Weber
(KIT - Karlsruhe Institute of Technology (DE))
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
The complexity of the heterogeneous computing resources, services and recurring infrastructure changes at the GridKa WLCG Tier-1 computing center require a structured approach to configuration management and optimization of interplay between functional components of the whole system. A set of tools deployed at GridKa, including Puppet, Redmine, Foreman, SVN and Icinga, provides the...
Dr
Andreas Gellrich
(DESY)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
The vast majority of jobs in the Grid are embarrassingly parallel. In
particular HEP tasks are divided into atomic jobs without need for
communication between them. Jobs are still neither multi-threaded nor
multi-core capable. On the other hand, resource requirements reach
from CPU-dominated Monte Carlo jobs to network intense analysis jobs.
The main objective of any Grid site is to...
Vidmantas Zemleris
(Vilnius University (LT))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
Background: The goal of the virtual data service integration is to provide a coherent interface for querying a number of heterogenous data sources (e.g., web services, web forms, proprietary systems, etc.) in cases where accurate results are necessary. This work explores various aspects of its usability.
Problem: Querying is usually carried out through a structured query language, such as...
Victoria Sanchez Martinez
(Instituto de Fisica Corpuscular (IFIC) UV-CSIC (ES))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
In this contribution we expose the performance of the Iberian (Spain and Portugal) ATLAS cloud during the first LHC running period (March 2010-January 2013) in the framework of the GRID Computing and Data Model. The evolution of the resources for CPU, disk and tape in the Iberian Tier1 and Tier2s is summarized. The data distribution over all ATLAS destinations is shown, focusing in the number...
Andrew John Washbrook
(University of Edinburgh (GB))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
High Performance Computing (HPC) provides unprecedented computing power for a diverse range of scientific applications. As of November 2012, over 20 supercomputers deliver petaflop peak performance with the expectation of "exascale" technologies available in the next 5 years. Despite the sizeable computing resources on offer there are a number of technical barriers that limit the use of HPC...
Eygene Ryabinkin
(National Research Centre Kurchatov Institute (RU))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
The review of the distributed grid computing infrastructure for LHC experiments in Russia is given. The emphasis is placed on the Tier-1 site construction at the National Research Centre "Kurchatov Institute" (Moscow) and the Joint Institute for Nuclear Research (Dubna).
In accordance with the protocol between CERN, Russia and the Joint Institute for Nuclear Research (JINR) on participation...
Luca dell'Agnello
(INFN-CNAF)
10/14/13, 3:00 PM
Long-term preservation of experimental data (intended as both raw and derived formats) is one of the emerging requirements coming from scientific collaborations. Within the High Energy Physics community the Data Preservation in High Energy Physics (DPHEP) group coordinates this effort.
CNAF is not only one of the Tier-1s for the LHC experiments, it is also a computing center providing...
Shaun De Witt
(STFC - Science & Technology Facilities Council (GB))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
WLCG is moving towards greater use of xrootd. While this will in general optimise resource usage on the grid, it can create load problems at sites when storage elements are unavailable. We present some possible methods of mitigating these problems and the results from experiments at STFC
Andrew John Washbrook
(University of Edinburgh (GB))
10/14/13, 3:00 PM
A number of High Energy Physics experiments have successfully run feasibility studies to demonstrate that many-core devices such as GPGPUs can be used to accelerate algorithms for trigger systems and data analysis. After this exploration phase experiments on the Large Hadron Collider are now investigating how these devices can be incorporated into key areas of their software framework in...
Mr
Stephen Lloyd
(University of Edinburgh)
10/14/13, 3:00 PM
The Matrix Element Method has been used with great success in the past several years, notably for the high precision top quark mass determination, and subsequently the single top quark discovery, at the Tevatron. Unfortunately, the Matrix Element method is notoriously CPU intensive due to the complex integration performed over the full phase space of the final state particles arising from...
DIMITRIOS ZILASKOS
(STFC)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
The WLCG uses HEP-SPEC as its benchmark for measuring CPU performance. This provides a consistent and repeatable CPU benchmark to describe experiment requirements, lab commitments and existing resources. However while HEP-SPEC has been customized to represents WLCG applications it is not a perfect measure.
The Rutherford Appleton Laboratory (RAL), is the UK Tier 1 site and provides CPU and...
Dr
Jean-Roch Vlimant
(CERN)
10/14/13, 3:00 PM
The analysis of the LHC data at the CMS experiment requires the production of a large number of simulated events. In 2012, CMS has produced over 4 Billion simulated events in about 100 thousands of datasets. Over the past years a tool (PREP) has been developed for managing such a production of thousands of samples.
A lot of experience working with this tool has been gained, and conclusions...
Dr
Janusz Martyniak
(Imperial College London)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
The international Muon Ionisation Cooling Experiment (MICE) is designed to demonstrate the principle of muon ionisation cooling for the first time, for application to a future Neutrino Factory or Muon Collider. The experiment is currently under construction at the ISIS synchrotron at the Rutherford-Appleton Laboratory, UK.
The configuration/condition of the experiment during each run is...
Yordan Ivanov Karadzhov
(Universite de Geneve (CH))
10/14/13, 3:00 PM
The Muon Ionization Cooling Experiment (MICE) is under development at the Rutherford Appleton Laboratory (UK). The goal of the experiment is to build a section of a cooling channel that can demonstrate the principle of ionization cooling and to verify its performance in a muon beam. The final setup of the experiment will be able to measure a 10% reduction in emittance (transverse phase space...
Dr
Patricia Mendez Lorenzo
(CERN)
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
The large potential and flexibility of the ServiceNow infrastructure based on "best practices" methods is allowing the migration of some of the ticketing systems traditionally used for the tracing of the servers and services available at the CERN IT Computer Center. This migration enables a standardization and globalization of the ticketing and control systems implementing a generic system...
Mark Mitchell
(University of Glasgow)
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
The monitoring of a grid cluster (or of any piece of reasonably scaled IT infrastructure) is a key element in the robust and consistent running of that site. There are several factors which are important to the selection of a useful monitoring framework, which include ease of use, reliability, data input and output. It is critical that data can be drawn from different instrumentation packages...
Alexandre Beche
(CERN)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
The computing models of the LHC experiments are gradually moving from hierarchical data models with centrally managed data pre-placement towards federated storage which provides seamless access to data files independently of their location and dramatically improved recovery due to fail-over mechanisms. Enabling loosely coupled data clusters to act as a single storage resource should increase...
Bogdan Lobodzinski
(DESY, Hamburg, Germany)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
Small Virtual Organizations (VO) employ all components of the EMI or gLite Middleware. In this framework, a monitoring system is designed for the H1 Experiment to identify and recognize within the GRID the best suitable resources for execution of CPU-time consuming Monte Carlo (MC) simulation tasks (jobs). Monitored resources are Computer Elements (CEs), Storage Elements (SEs), WMS-servers...
Georg Weidenspointner
(MPE Garching)
10/14/13, 3:00 PM
An extensively documented, quantitative study of software evolution resulting in deterioration of physical accuracy over the years is presented. The analysis concerns the energy deposited by electrons in various materials produced by Geant4 versions released between 2007 and 2013.
The evolution of the functional quality of the software is objectively quantified by means of a rigorous...
Dr
Maria Grazia Pia
(Universita e INFN (IT))
10/14/13, 3:00 PM
A large-scale project is in progress, which validates the basic constituents of the electromagnetic physics models implemented in major Monte Carlo codes (EGS, FLUKA, Geant4, ITS, MCNP, Penelope) against extensive collections of experimental data documented in the literature. These models are responsible for the physics observables and the signal generated in particle detectors, including...
Ian Gable
(University of Victoria (CA))
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
We review the demonstration of next generation high performance 100 Gbps networks for HEP that took place at the Supercomputing 2012 (SC12) conference in Salt Lake City. Three 100 Gbps circuits were established from the California Institute of Technology, the University of Victoria and the University of Michigan to the conference show floor. We were able to to efficiently utilize these...
Paul Nilsson
(University of Texas at Arlington (US))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
The Production and Distributed Analysis system (PanDA) has been in use in the ATLAS Experiment since 2005. It uses a sophisticated pilot system to execute submitted jobs on the worker nodes. While originally designed for ATLAS, the PanDA Pilot has recently been refactored to facilitate use outside of ATLAS. Experiments are now handled as plug-ins, and a new PanDA Pilot user only has to...
Dr
Peter Van Gemmeren
(Argonne National Laboratory (US))
10/14/13, 3:00 PM
The ATLAS event store employs a persistence framework with extensive navigational capabilities. These include real-time back navigation to upstream processing stages, externalizable data object references, navigation from any data object to any other both within a single file and across files, and more. The 2013-2014 shutdown of the Large Hadron Collider provides an opportunity to enhance...
Anastasia Karavdina
(University Mainz)
10/14/13, 3:00 PM
Precise luminosity determination is crucial for absolute cross-section measurements and scanning experiments with the fixed target PANDA experiment at the planned antiproton accelerator HESR (FAIR, Germany). For the determination of the luminosity we will exploit the elastic antiproton-proton scattering. Unfortunately there are no or only a few data with large uncertainties available in the...
Christopher John Walker
(University of London (GB))
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
The WLCG, and high energy physics in general, relies on remote Tier-2
sites to analyse the large quantities of data produced. Transferring
this data in a timely manner requires significant tuning to make
optimum usage of expensive WAN links.
In this paper we describe the techniques we have used at QMUL to
optimise network transfers. Use of the FTS with settings and
appropriate TCP...
Dr
Giacinto Donvito
(INFN-Bari),
Tommaso Boccali
(Sezione di Pisa (IT))
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
The Italian Ministry of Research (MIUR) funded in the past years research projects aimed to an optimization of the analysis activities in the Italian CMS computing Centers. A new grant started in 2013, and activities are already ongoing in 9 INFN sites, all hosting local CMS groups. Main focus will be on the creation of an italian storage federation (via Xrootd initially, and later HTTP) which...
Egor Ovcharenko
(ITEP Institute for Theoretical and Experimental Physics (RU))
10/14/13, 3:00 PM
One of the current problems in HEP computing is the development of particle propagation algorithms capable of efficient work at parallel architectures. An interesting approach in this direction has been recently introduced by the GEANT5 group at CERN [1]. Our report will be devoted to realization of similar functionality using Intel Threading Building Blocks (TBB) library.
In the prototype...
Stewart Martin-Haugh
(University of Sussex (GB))
10/14/13, 3:00 PM
We present a description of the algorithms and the performance of the ATLAS Inner Detector trigger for LHC run I, as well as prospects for a redesign of the tracking algorithms in run 2. The Inner Detector trigger algorithms are vital for many trigger signatures at ATLAS. The performance of the algorithms for muons, electrons, taus and b-jets is presented.
The ATLAS trigger software after...
Enrico Bonaccorsi
(CERN),
Francesco Sborzacchi
(Istituto Nazionale Fisica Nucleare (IT)),
Niko Neufeld
(CERN)
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
The virtual computing is often run to satisfy different needs: reduce
costs, reduce resources, simplify maintenance and the last but not the
least add flexibility.
The use of Virtualization in a complex system such as a farm of PCs that
control the hardware of an experiment (PLC, power supplies ,gas,
magnets..) put as in a condition where not only an High Performance
requirements...
Eduardo Bach
(UNESP - Universidade Estadual Paulista (BR))
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
Distributed storage systems have evolved from providing a simple means to store data remotely to offering advanced services like system federation and replica management. This evolution have made possible due to the advancement of the underlying communication technology, that plays a vital role in determining the communication efficiency of the distributed systems. The dCache system, which has...
Dr
Dmytro Kovalskyi
(Univ. of California Santa Barbara (US))
10/14/13, 3:00 PM
Databases are used in many software components of the HEP computing, from monitoring and task scheduling to data storage and processing. While the database design choices have a major impact on the system performance, some solutions give better results out of the box than the others. This paper presents detailed comparison benchmarks of the most popular Open Source systems for a typical class...
Christophe Haen
(Univ. Blaise Pascal Clermont-Fe. II (FR))
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
The backbone of the LHCb experiment is the Online system, which is a very large and heterogeneous computing center. Making sure of the proper behavior of the many different tasks running on the more than 2000 servers represents a huge workload for the small expert-operator team and is a 24/7 task. At the occasion of CHEP 2012, we presented a prototype of a framework that we designed in order...
Dr
Dirk Hoffmann
(Centre de Physique des Particules de Marseille, CNRS/IN2P3)
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
PLUME - FEATHER is a non-profit project created to Promote economicaL, Useful and Maintained softwarE For the Higher Education And THE Research communities. The site references software, mainly Free/Libre Open Source Software (FLOSS) from French universities and national research organisations, (CNRS, INRA...), laboratories or departments. Plume means feather in French. The main goals of PLUME...
Graeme Andrew Stewart
(CERN)
10/14/13, 3:00 PM
This paper describes a popularity prediction tool for data-intensive data management systems, such as the ATLAS distributed data management (DDM) system. The tool is fed by the DDM popularity system, which produces historical reports about ATLAS data usage and provides information about the files, datasets, users and sites where data was accessed. The tool described in this contribution uses...
Nathalie Rauschmayr
(CERN)
10/14/13, 3:00 PM
Due to the continuously increasing number of cores on modern CPUs, it is important to adapt HEP applications. This must be done at different levels: the software which must support parallelization and the scheduling has to differ between multicore and singlecore jobs. The LHCb software framework (GAUDI) provides a parallel prototype (GaudiMP), based on the multiprocessing approach. It allows a...
Simone Coscetti
(Sezione di Pisa (IT))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
The ALEPH Collaboration took data at the LEP (CERN) electron-positron collider in the period 1989-2000, producing more than 300 scientific papers. While most of the Collaboration activities stopped in the last years, the data collected still has physics potential, with new theoretical models emerging, and needing a check with data at the Z and WW production energies. An attempt to revive and...
Dr
Dirk Hoffmann
(Centre de Physique des Particules de Marseille, CNRS/IN2P3)
10/14/13, 3:00 PM
We are developing the prototype of a high speed data acquisition (DAQ) system for the Cherenkov Telescope Array. This experiment will be the next generation ground-based gamma-ray instrument. It will be made up of approximately 100 telescopes of at least three different sizes, from 6 to 24 meters in diameter.
Each camera equipping the telescopes is composed of hundreds of light detecting...
Semen Lebedev
(Justus-Liebig-Universitaet Giessen (DE))
10/14/13, 3:00 PM
The software framework of the CBM experiment at FAIR - CBMROOT - has been continuously growing over the years. The increasing complexity of the framework and number of users require improvements in maintenance, reliability and in overall software development process. In this report we address the problem of the software quality assurance (QA) and testing. Two main problems are considered in...
Dr
Armando Fella
(INFN Pisa), Mr
Domenico Diacono
(INFN Bari), Dr
Giacinto Donvito
(INFN-Bari), Mr
Giovanni Marzulli
(GARR),
Paolo Franchini
(Universita e INFN (IT)), Dr
Silvio Pardi
(INFN)
10/14/13, 3:00 PM
In HEP computing context, R&D studies aiming to the definition of the data and workload models were brought forward by the SuperB community beyond the experiment life itself. This work is considered of great interest for a generic mid- and small size VO during its Computing Model definition phase.
Data-model R&D work we are presenting, starts with the general design
description of the...
Dr
Tony Wildish
(Princeton University (US))
10/14/13, 3:00 PM
PhEDEx. the data-placement tool used by the CMS experiment at the LHC, was conceived in a more trusting time. The security model was designed to provide a safe working environment for site agents and operators, but provided little more protection than that. CMS data was not sufficiently protected against accidental loss caused by operator error or software bugs or from loss of data caused by...
Adrian Buzatu
(University of Glasgow (GB))
10/14/13, 3:00 PM
In high-‐energy physics experiments, online selection is crucial to reject most uninteresting collisions and to focus on interesting physical signals.
The b-‐jet selection is part of the trigger strategy of the ATLAS experiment and is meant to select hadronic final states with heavy-‐flavor content. This is important for the selection of physics channels with more than one b-‐jet in the...
Christian Veelken
(Ecole Polytechnique (FR))
10/14/13, 3:00 PM
An algorithm for reconstruction of the Higgs mass in $H \rightarrow \tau\tau$ decays is presented. The algorithm computes for each event a likelihood function $P(M_{\tau\tau})$ which quantifies the level of compatibility of a Higgs mass hypothesis $M_{\tau\tau}$, given the measured momenta of visible tau decay products plus missing transverse energy reconstructed in the event. The algorithm is...
Mr
Igor Mandrichenko
(Fermilab)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
RESTful web services are popular solution for distributed data access and information management. Performance, scalability and reliability of such services is critical for the success of data production and analysis in High Energy Physics as well as other areas of science.
At FNAL, we have been successfully using REST HTTP-based data access architecture to provide access to various types...
Dr
Tony Wildish
(Princeton University (US))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
PhEDEx has been serving CMS community since 2004 as the data broker. Every PhEDEx operation is initiated by a request, such as request to move data, request to delete data, and so on. A request has it own life cycle, including creation, approval, notification, and book keeping and the details depend on its type. Currently, only two kinds of requests, transfer and deletion, are fully integrated...
Bertrand Bellenot
(CERN)
10/14/13, 3:00 PM
In my poster I'll present a new graphical back-end for ROOT that has been developed for the Mac OS X operating system as an alternative to the more than 15 year-old X11-based version. It represents a complete implementation of ROOT's GUI, 2D and 3D graphics based on Apple's native APIs/frameworks, written in Objective-C++.
Daniela Remenska
(NIKHEF (NL))
10/14/13, 3:00 PM
A big challenge in concurrent software development is early discovery of design errors which can lead to deadlocks or race-conditions. Traditional testing does not always expose such problems in complex distributed applications. Performing more rigorous formal analysis, like model-checking, typically requires a model which is an abstraction of the system. For object-oriented software, UML is...
Mr
Igor Mandrichenko
(Fermilab)
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
Over several years, we have developed a number of collaborative tools used by groups and collaborations at FNAL, which is becoming a Suite of Scientific Collaborative Tools. Currently, the suite includes:
- Electronic Logbook (ECL),
- Shift Scheduler,
- Speakers Bureau and
- Members Database.
These product organize and help run the collaboration at every stage of its life...
Federico Stagni
(CERN),
Mario Ubeda Garcia
(CERN)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
Within this paper we present an autonomic Computing resources management system used by LHCb for assessing the status of their Grid resources. Virtual Organizations Grids include heterogeneous resources. For example, LHC experiments very often use resources not provided by WLCG and Cloud Computing resources will soon provide a non-negligible fraction of their computing power.
The lack of...
Giovanni Zurzolo
(Universita e INFN (IT))
10/14/13, 3:00 PM
Artificial Neural Networks (ANN) are widely used in High Energy Physics, in particular as software for data analysis. In the ATLAS experiment that collects proton-proton and heavy ion collision data at the Large Hadron Collider, ANN are mostly applied to make a quantitative judgment on the class membership of an event, using a number of variables that are supposed to discriminate between...
Mr
Ajay Kumar
(Indian Institute of Technology Indore)
10/14/13, 3:00 PM
Ajay Kumar and Ankhi Roy
For the PANDA collaboration
Indian Institute of Technology Indore, Indore-4520017, India
Email- ajayk@iiti.ac.in
The PANDA experiment is one of the main experiments at the future accelerator facility FAIR which is currently under construction in Darmstadt, Germany. Experiments will be performed with intense, phase space cooled antiproton beams incident on a...
Dr
Guy Barrand
(Universite de Paris-Sud 11 (FR))
10/14/13, 3:00 PM
Softinex names a software environment targeted to data analysis and visualization. It covers the C++ inlib and exlib "header only" libraries that permit, through GL-ES and a maximum of common code, to build applications deliverable on the AppleStore (iOS), GooglePlay (Android), traditional laptops/desktops under MacOSX, Linux and Windows, but also deliverable as a web service able to display...
Dr
Alexander Moibenko
(Fermi NAtiona Accelerator Laboratoy)
10/14/13, 3:00 PM
Enstore is a tape based Mass Storage System originally designed for Run II Tevatron experiments at FNAL (CDF, D0). Over the years it has proven to be reliable and scalable data archival and delivery solution, which meets diverse requirements of variety of applications including US CMS Tier 1, High Performance Computing, Intensity Frontier experiments as well as data backups. Data intensive...
Dr
Simon Patton
(LAWRENCE BERKELEY NATIONAL LABORATORY)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
The SPADE application was first used by the IceCube experiment to move its data files from the South Pole to Wisconsin. Since then is has been adapted by the DayaBay experiment to move its data files from its experiment, just outside Hong Kong, to both Beijing an LBNL. The aim of this software is to automate much of the data movement and warehousing that is often done by hand or home-grown...
Alastair Dewhurst
(STFC - Science & Technology Facilities Council (GB))
10/14/13, 3:00 PM
During the early running of the LHC , multiple collaborations began to include Squid caches in their distributed computing models. The two main use cases are: for remotely accessing conditions data via Frontier, which is used by ATLAS and CMS; and serving collaboration software via CVMFS, which is used by ATLAS, CMS, and LHCb, and is gaining traction with some non-LHC collaborations. As a...
Dr
Xavier Espinal Curull
(CERN)
10/14/13, 3:00 PM
This contribution describes the evolution of the main CERN storage system, CASTOR, as it manages the bulk data stream of the LHC and other CERN experiments, achieving nearly 100 PB of stored data by the end of LHC Run 1.
Over the course of 2012 the CASTOR service has addressed the Tier-0 data management requirements, focusing on a tape-backed archive solution, ensuring smooth operations of...
Dr
Catherine Biscarat
(LPSC/IN2P3/CNRS France)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
We describe the synergy between CIMENT (a regional multidisciplinary HPC centre) and the infrastructures used for the analysis of data recorded by the ATLAS experiment at the LHC collider and the D0 experiment at the Tevatron.
CIMENT is the High Performance Computing (HPC) centre developed by Grenoble University. It is a federation of several scientific departments and it is based on the...
Daniele Francesco Kruse
(CERN)
10/14/13, 3:00 PM
Disk access and tape migrations compete for network bandwidth in CASTOR’s disk servers, over various protocols: RFIO, Xroot, root and GridFTP. As there are a limited number of tape drives, it is important be keep them busy all the time, at their nominal speed. With potentially 100s of user read streams per server, the bandwidth for the tape migrations has to be guaranteed to a controlled...
Thomas Lindner
(T)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
ND280 is the off-axis near detector for the T2K neutrino experiment. ND280 is a sophisticated, multiple sub-system detector designed to characterize the T2K neutrino beam and measure neutrino cross-sections. We have developed a complicated system for processing and simulating the ND280 data, using computing resources from North America, Europe and Japan. The first key challenge has been...
michele pezzi
(Infn-cnaf)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
In large computing centers, such as the INFN-CNAF Tier1, is essential to be able to set all the machines, depending on use, in an automated way. For several years at the Tier1 has been used Quattor, a server provisioning tool, which is currently used in production.
Nevertheless we have recently started a comparison study involving other tools able to provide specific server installation...
Robert Fay
(University of Liverpool)
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
A key aspect of ensuring optimum cluster reliability and productivity lies in keeping worker nodes in a healthy state. Testnodes is a lightweight node testing solution developed at Liverpool. While Nagios has been used locally for general monitoring of hosts and services, Testnodes is optimised to answer one question: is there any reason this node should not be accepting jobs? This tight focus...
Jason Webb
(Brookhaven National Lab)
10/14/13, 3:00 PM
The STAR experiment has adopted an Abstract Geometry Modeling Language (AgML) as the primary description of our geometry model. AgML establishes a level of abstraction, decoupling the definition of the detector from the software libraries used to create the concrete geometry model. Thus, AgML allows us to support both our legacy GEANT3 simulation application and our ROOT/TGeo based...
Mr
Barthelemy Von Haller
(CERN)
10/14/13, 3:00 PM
ALICE (A Large Ion Collider Experiment) is a detector designed to study the physics of strongly interacting matter and the quark-gluon plasma produced in heavy-ion collisions at the CERN Large Hadron Collider (LHC). Due to the complexity of ALICE in terms of number of detectors and performance requirements, Data Quality Monitoring (DQM) plays an essential role in providing an online feedback...
Dr
Dario Barberis
(Università e INFN Genova (IT))
10/14/13, 3:00 PM
Modern scientific experiments collect vast amounts of data that must be cataloged to meet multiple use cases and search criteria. In particular, high-energy physics experiments currently in operation produce several billion events per year. A database with the references to the files including each event in every stage of processing is necessary in order to retrieve the selected events from...
Martin Woudstra
(University of Manchester (GB))
10/14/13, 3:00 PM
CERN’s Large Hadron Collider (LHC) is the highest energy proton-proton collider, providing also the highest instantaneous luminosity as a hadron collider. Bunch crossings occurred every 50 ns in 2012 runs. Amongst of which the online event selection system should reduce the event recording rate down to a few 100 Hz, while events are in a harsh condition with many overlapping proton-proton...
Rafal Zbigniew Grzymkowski
(P)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
In the multidisciplinary institutes the traditional way of computations is highly ineffective. A computer cluster dedicated to a single research group is typically exploited at a rather low level. The private cloud model enables various groups to share computing resources. It can boost the efficiency of the infrastructure usage by a large factor and at the same time reduce maintenance costs....
Dr
Federico De Guio
(CERN)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
The Data Quality Monitoring (DQM) Software proved to be a central tool in the CMS experiment. Its flexibility allowed its integration in several environments: Online, for real-time detector monitoring; Offline, for the final, fine-grained Data Certification; Release-Validation, to constantly validate our reconstruction software; in Monte Carlo productions. The central tool to deliver Data...
Shima Shimizu
(Kobe University (JP))
10/14/13, 3:00 PM
The ATLAS jet trigger is an important element of the event selection process,
providing data samples for studies of Standard Model physics and searches for new
physics at the LHC. The ATLAS jet trigger system has undergone substantial
modifications over the past few years of LHC operations, as experience developed
with triggering in a high luminosity and high event pileup environment. In...
Dr
WooJin Park
(KIT)
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
The GridKa computing center, hosted by Steinbuch Centre for Computing at the Karlsruhe Institute for Technology (KIT) in Germany, is serving as the largest Tier-1 center used by the ALICE collaboration at the LHC. In 2013, GridKa provides 30k HEPSEPC06, 2.7 PB of disk space, and 5.25 PB of tape storage to ALICE. The 10Gbit/s network connections from GridKa to CERN, several Tier-1 centers and...
Norman Anthony Graf
(SLAC National Accelerator Laboratory (US))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
The International Linear Collider (ILC) physics and detector
community recently completed an exercise to demonstrate the
physics capabilities of detector concepts. The Detailed
Baseline Design (DBD) involved the generation, simulation, reconstruction and analysis of large samples of Monte Carlo datasets. The detector simulations utilized extremely detailed Geant4 implementations of...
Thomas Baron
(CERN)
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
For a long time HEP has been ahead of the curve in its usage of remote collaboration tools, like videoconference and webcast, while the local CERN collaboration facilities were somewhat behind the expected quality standards for various reasons. This time is now over with the creation by the CERN IT department in 2012 of an integrated conference room service which provides guidance and...
Mr
Massimo Sgaravatto
(INFN Padova)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
The Legnaro-Padova Tier-2 is a computing facility serving the
ALICE and CMS LHC experiments. It also supports other High Energy Physics experiments and other virtual organizations of different disciplines, which can opportunistically harness idle resources if
available.
The unique characteristic of this Tier-2 is its topology: the computational resources are spread in two different...
Sandra Saornil Gamarra
(Universitaet Zuerich (CH))
10/14/13, 3:00 PM
The experiment control system of the LHCb experiment is continuously evolving and improving. The guidelines and structure initially defined are kept, and more common tools are made available to all sub-detectors. Although the main system control is mostly integrated and actions are executed in common for the whole LHCb experiment, there is some degree of freedom for each sub-system to...
Pierrick Hanlet
(Illinois Institute of Technology)
10/14/13, 3:00 PM
The Muon Ionization Cooling Experiment (MICE) is a demonstration
experiment to prove the feasibility of cooling a beam of muons for
use in a Neutrino Factory and/or Muon Collider. The MICE cooling
channel is a section of a modified Study II cooling channel which
will provide a 10% reduction in beam emittance. In order to ensure a
reliable measurement, MICE will measure the beam emittance...
Joern Mahlstedt
(NIKHEF (NL))
10/14/13, 3:00 PM
The LHC is the world's highest energy and luminosity proton-proton (p-p) collider. During 2012 luminosities neared 10^34 cm-2 s-1, with bunch crossings occurring every 50 ns. The online event selection system of the ATLAS detector must reduce the event recording rate to only a few hundred Hz and, at the same time, selecting events considered interesting. This presentation will specifically...
Pierrick Hanlet
(Illinois Institute of Technology)
10/14/13, 3:00 PM
The international Muon Ionisation Cooling Experiment (MICE) is designed to demonstrate the principle of muon ionisation cooling for the first time, for application to a future Neutrino Factory or Muon Collider. In order to measure the change in beam emittance, MICE is equipped with a pair of high precision scintillating fibre trackers. The trackers are required to measure a 10% change in...
Daniele Francesco Kruse
(CERN)
10/14/13, 3:00 PM
Physics data stored in CERN tapes is quickly reaching the 100 PB milestone. Tape is an ever-changing technology that is still following Moore's law in terms of capacity. This means we can store every year more and more data in the same amount of tapes. However this doesn't come for free: the first obvious cost is the new higher capacity media. The second less known cost is related to moving...
Mr
Andrey SHEVEL
(Petersburg Nuclear Physics Institute)
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
A small physics group (3-15 persons) might use a number of computing facilities for the analysis/simulation, developing/testing, teaching. It is discussed different types of computing facilities: collaboration computing facilities, group local computing cluster (including colocation), cloud computing. The author discuss the growing variety of different computing options for small groups and...
Bob Cowles
(BrightLite Information Security)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
As HEP collaborations grow in size (10 years ago, BaBar was 600 scientists; now, both CMS and ATLAS are on the order of 3000 scientists), the collaboratory has become a key factor in allowing identity management (IdM), once confined to individual sites, to scale with the number of members, number of organizations, and the complexity of the science collaborations. Over the past two decades (at...
Jason Webb
(Brookhaven National Lab)
10/14/13, 3:00 PM
The STAR experiment pursues a broad range of physics topics in pp, pA and AA collisions produced by the Relativistic Heavy Ion Collider (RHIC). Such a diverse experimental program demands a simulation framework capable of supporting an equally diverse set of event generators, and a flexible event record capable of storing the (common) particle-wise and (varied) event-wise information provided...
Oliver Holme
(ETH Zurich, Switzerland)
10/14/13, 3:00 PM
The Electromagnetic Calorimeter (ECAL) is one of the sub-detectors of the Compact Muon Solenoid (CMS) experiment of the Large Hadron Collider (LHC) at CERN. The Detector Control System (DCS) that has been developed and implemented for the CMS ECAL was deployed in accordance with the LHC schedule and has been supporting the detector data-taking since LHC physics runs started in 2009. During...
Andrew David Lahiff
(STFC - Science & Technology Facilities Council (GB))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
While migration from the grid to the cloud has been gaining increasing momentum in recent times, WLCG sites are currently still expected to accept grid job submission, and this is likely to continue for the foreseeable future. Furthermore, sites which support multiple experiments may need to provide both cloud and grid-based access to resources for some time, as not all experiments may be...
Shaun De Witt
(STFC - Science & Technology Facilities Council (GB))
10/14/13, 3:00 PM
LHC experiments are moving away from a traditional HSM solution for Tier 1's in order to separate long term tape archival from disk only access, using the tape as a true archive (write once, read rarely). In this poster we present two methods by which this is being achieved at two distinct sites, ASGC and RAL, which have approached this change in very different ways.
Robert Fay
(University of Liverpool)
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
As the number of cores on chip continues to trend upwards and new CPU architectures emerge, increasing CPU density and diversity presents multiple challenges to site administrators.
These include scheduling for massively multi-core systems (potentially including GPU (integrated and dedicated) and many integrated core (MIC)) to ensure a balanced throughput of jobs while preserving overall...
Daniel Hugo Campora Perez
(CERN)
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
The LHCb Online Network is a real time high performance network, in which 350 data sources send data over a Gigabit Ethernet LAN to more than 1500 receiving nodes. The aggregated throughput of the application, called Event Building, is more than 60 GB/s. The protocol employed by LHCb makes the sending nodes transmit simultaneously portions of events to one receiving node at a time, which is...
Dr
Daniel van der Ster
(CERN), Dr
Jakub Moscicki
(CERN)
10/14/13, 3:00 PM
AFS is a mature and reliable storage service at CERN, having worked for more than 20 years as the provider of Linux home directories and application areas. Recently, our AFS service has been growing at unprecedented rates (300% in the past year), thanks to innovations in both the hardware and software components of our file servers.
This work will present how AFS is used at CERN and how...
Daniele Gregori
(Istituto Nazionale di Fisica Nucleare (INFN)),
Luca dell'Agnello
(INFN-CNAF),
Pier Paolo Ricci
(INFN CNAF),
Tommaso Boccali
(Sezione di Pisa (IT)), Dr
Vincenzo Vagnoni
(INFN Bologna), Dr
Vladimir Sapunenko
(INFN)
10/14/13, 3:00 PM
The Mass Storage System installed at the INFN CNAF Tier-1 is one of the biggest hierarchical storage facilities in Europe. It currently provides storage resources for about 12% of all LHC data, as well as to other High Energy Physics experiments.
The Grid Enabled Mass Storage System (GEMSS) is the present solution implemented at the INFN CNAF Tier-1 and it is based on a custom integration...
Ivan Antoniev Dzhunov
(University of Sofia)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
Given the distributed nature of the grid and the way CPU resources are pledged and scared around the globe, VO's are facing the challenge to monitor the use of these resources. For CMS and the operation of centralized workflows the monitoring of how many production jobs are running and pending in the Glidein WMS production pools is very important. The Dashboard SSB (Site Status Board) provides...
Dr
Tomoaki Nakamura
(University of Tokyo (JP))
10/14/13, 3:00 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Poster presentation
The Tokyo Tier2 center, which is located at International Center for Elementary Particle Physics (ICEPP) in the University of Tokyo, was established as a regional analysis center in Japan for the ATLAS experiment. The official operation with WLCG was started in 2007 after the several years development since 2002. In December 2012, we have replaced almost all hard wares as the third system...
Jetendr Shamdasani
(University of the West of England (GB))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
Efficient, distributed and complex software is central in the analysis of high energy physics (HEP) data. One area that has been somewhat overlooked in recent years has been the tracking of the development of the HEP software and of its use in data analyses and its evolution over time. This area of tracking analyses to provide records of actions performed, outcomes achieved and (re-)design...
Daniele Francesco Kruse
(CERN)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
Administrating a large-scale, multi-protocol, hierarchical tape storage infrastructure like the one at CERN, which stores around 30PB / year, requires an adequate monitoring system for quick spotting of malfunctions, easier debugging and on demand report generation. The main challenges for such system are: to cope with log format diversity and its information scattered among several log files,...
Morten Dam Joergensen
(Niels Bohr Institute (DK))
10/14/13, 3:00 PM
The ATLAS offline data quality monitoring infrastructure functioned successfully during the 2010-2012 run of the LHC. During the 2013-14 long shutdown, a large number of upgrades will be made in response to user needs and to take advantage of new technologies - for example, deploying richer web applications, improving dynamic visualization of data, streamlining configuration, and moving...
Rahmat Rahmat
(University of Mississippi (US))
10/14/13, 3:00 PM
HFGFlash is a very fast simulation of electromagnetic showers using parameterizations of the profiles in Hadronic Forward Calorimeter. HF GFlash has good agreement to Collision Data and previous Test Beam results. In addition to good agreement to Data and previous Test Beam results, HFGFlash can simulate about 10000 times faster than Geant4. We will report the latest development of HFGFlash...
Robin Eamonn Long
(Lancaster University (GB))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
The need to maximize computing facilities whilst maintaining versatile and flexible setups leads to the need for on demand virtual machines through the use of cloud computing. GridPP is currently investigating the role that Cloud Computing, in the form of Virtual Machines, can play in supporting Particle Physics analyses. As part of this research we look at the ability of VMWare's ESXi...
Igor Sfiligoi
(University of California San Diego)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
Monitoring is an important aspect of any job scheduling environment, and Grid computing is no exception. Writing quality monitoring tools is however a hard proposition, so the Open Science Grid decided to leverage existing enterprise-class tools in the context of the glideinWMS pilot infrastructure, which powers a large fraction of its Grid computing. The product chosen is the CycleServer,...
Carl Henrik Ohman
(Uppsala University (SE))
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
With the advent of commercial as well as institutional and national clouds, new opportunities for on-demand computing resources for the HEP community become available. With the new cloud technologies come also new challenges, and one such is the contextualization of cloud resources with regard to requirements of the user and his experiment. In particular on Google's new cloud platform Google...
Igor Sfiligoi
(University of California San Diego)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
The HTCondor based glideinWMS has become the product of choice for exploiting Grid resources for many communities. Unfortunately, its default operational model expects users to log into a machine running a HTCondor schedd before being able to submit their jobs. Many users would instead prefer to use their local workstation for everything.
A product that addresses this problem is rcondor, a...
Antanas Norkus
(Vilnius University (LT))
10/14/13, 3:00 PM
The scrutiny and validation of the software and of the calibrations used to simulate and reconstruct the collision events, have been key elements to the physics performance of the CMS experiment.
Such scrutiny is performed in stages by approximately one hundred experts who master specific areas of expertise, ranging from the low-level reconstruction and calibration which specific to a...
Stephen Jones
(Liverpool University)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
VomsSnooper is a tool that provides an easy way to keep documents and sites up to date with the newest VOMS records from the Operations Portal, and removes the need for manual edits to security configuration files.
Yaim is used to configure the middle-ware at grid sites. Specifically, Yaim processes variables that define which VOMS services are used to authenticate users of any VO. The data...
Alexandre Beche
(CERN),
David Tuckett
(CERN)
10/14/13, 3:00 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Poster presentation
The Worldwide LHC Computing Grid provides resources for the four main virtual organizations. Along with data processing, data distribution is the key computing activity on the WLCG infrastructure. The scale of this activity is very large, the ATLAS virtual organization (VO) alone generates and distributes more than 40 PB of data in 100 million files per year. Another challenge is the...
Matevz Tadel
(Univ. of California San Diego (US))
10/14/13, 3:00 PM
Following the smashing success of XRootd-based USCMS data-federation, AAA project investigated extensions of the federation architecture by developing two sample implementations of an XRootd, disk-based, caching-proxy. The first one simply starts fetching a whole file as soon as a file-open request is received and is suitable when completely random file access is expected or it is already...
Dr
Peter Elmer
(Princeton University (US))
10/14/13, 3:45 PM
Software Engineering, Parallelism & Multi-Core
Oral presentation to parallel session
Modern HEP software stacks, such as those used by the LHC experiments
at CERN, involve many millions of lines of custom code per experiment,
as well as a number of similarly sized shared packages (ROOT, Geant4,
etc.) Thousands of people have made contributions over time to these
code bases, including graduate students, postdocs, professional
researchers and software/computing...
Dr
Daniel van der Ster
(CERN)
10/14/13, 3:45 PM
Data Stores, Data Bases, and Storage Systems
Oral presentation to parallel session
Emerging storage requirements, such as the need for block storage for both OpenStack VMs and file services like AFS and NFS, have motivated the development of a generic backend storage service for CERN IT. The goals for such a service include (a) vendor neutrality, (b) horizontal scalability with commodity hardware, (c) fault tolerance at the disk, host, and network levels, and (d) support for...
Dr
Antonio Maria Perez Calero Yzquierdo
(Centro de Investigaciones Energ. Medioambientales y Tecn. - (ES)
10/14/13, 3:45 PM
Distributed Processing and Data Handling B: Experiment Data Processing, Data Handling and Computing Models
Oral presentation to parallel session
In the next years, processor architectures based on much larger numbers of cores will be most likely the model to continue "Moore's Law" style throughput gains. This not only results in many more jobs in parallel running the LHC Run 1 era monolithic applications. Also the memory requirements of these processes push the workernode architectures to the limit. One solution is parallelizing the...
Ramon Medrano Llamas
(CERN)
10/14/13, 3:45 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Oral presentation to parallel session
In order to ease the management of their infrastructure, most of the WLCG sites are adopting cloud based strategies. In the case of CERN, the Tier 0 of the WLCG, is completely restructuring the resource and configuration management of their computing center under the codename Agile Infrastructure. Its goal is to manage 15,000 Virtual Machines by means of an OpenStack middleware in order to...
Olof Barring
(CERN)
10/14/13, 3:45 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Oral presentation to parallel session
In May 2012 CERN signed a contract with the Wigner Data Centre in Budapest for an extension to the CERN’s central computing facility beyond its current boundaries set by electrical power and cooling available for computing. The centre is operated as a remote co-location site providing rack-space, electrical power and cooling for server, storage and networking equipment acquired by CERN. The...
Prof.
Ivan Kisel
(GSI, Gesellschaft fuer Schwerionenforschung mbH)
10/14/13, 3:45 PM
Event Processing, Simulation and Analysis
Oral presentation to parallel session
The CBM (Compressed Baryonic Matter) experiment is an experiment being prepared to operate at the future Facility for Anti-Proton and Ion Research (FAIR, Darmstadt, Germany). Its main focus is the measurement of very rare probes, which requires interaction rates of up to 10 MHz. Together with the high multiplicity of charged tracks produced in heavy-ion collisions, this leads to huge data...
Mr
Arnim Balzer
(DESY, University Potsdam)
10/14/13, 3:45 PM
Data acquisition, trigger and controls
Oral presentation to parallel session
The High Energy Stereoscopic System (H.E.S.S.) is a system of five Imaging Atmospheric Cherenkov Telescopes (IACTs) located in the Khomas Highland in Namibia. It measures cosmic gamma-rays with very high energies (VHE; > 100 GeV) using the Earth’s atmosphere as a calorimeter. The H.E.S.S. array has entered Phase II in September 2012 with the inauguration of a fifth telescope that is larger and...
Ben Jones
(CERN),
Gavin Mccance
(CERN),
Nacho Barrientos Arias,
Steve Traylen
(CERN)
10/14/13, 4:05 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Oral presentation to parallel session
For over a decade CERN's fabric management system has been based on home-grown solutions. Those solutions are not dynamic enough for CERN to face its new challenges such as significantly scaling out, multi-site management and the Cloud Computing model, without any additional staff. This presentation will illustrate the motivations for CERN to move to a new tool-set in the context of the Agile...
Jaroslav Zalesak
(Acad. of Sciences of the Czech Rep. (CZ))
10/14/13, 4:05 PM
Data acquisition, trigger and controls
Oral presentation to parallel session
The NOvA experiment has developed a data acquisition system that is able to continuously digitize and produce a zero bias streaming readout for the more than 368,000 detectors cells that constitute the 14 kTon far detector. The NOvA DAQ system combines custom built frontend readout and data aggregation hardware, with advances in enterprise class networking to continuously deliver data to...
Jim Kowalkowski
(Fermilab)
10/14/13, 4:07 PM
Software Engineering, Parallelism & Multi-Core
Oral presentation to parallel session
For nearly two decades, the C++ programming language has been the
dominant programming language for experimental HEP. The publication of
ISO/IEC 14882:2011, the current version of the international standard
for the C++ programming language, makes available a variety of language
and library facilities for improving the robustness, expressiveness, and
computational efficiency of C++ code....
Ian Fisk
(Fermi National Accelerator Lab. (US)),
Jacob Thomas Linacre
(Fermi National Accelerator Lab. (US))
10/14/13, 4:07 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Oral presentation to parallel session
During Spring 2013, CMS processed 1 Billion RAW data events at the San Diego Super Computing Center (SDSC) that was nearly the size of half the CMS dedicated Tier-1 processing resources. This facility has none of the permanent CMS services, service level agreements, or support normally associated with a Tier-1, and was assembled with a few weeks notice to process only a few workflows. The size...
Maxim Potekhin
(Brookhaven National Laboratory (US))
10/14/13, 4:07 PM
Distributed Processing and Data Handling B: Experiment Data Processing, Data Handling and Computing Models
Oral presentation to parallel session
The ATLAS Production System is the top level workflow manager which translates physicists' needs for production level processing into actual workflows executed across about a hundred processing sites used globally by ATLAS. As the production workload increased in volume and complexity in recent years (the ATLAS production tasks count is above one million, with each task containing hundreds or...
Dr
Jakub Moscicki
(CERN)
10/14/13, 4:08 PM
Data Stores, Data Bases, and Storage Systems
Oral presentation to parallel session
Individual users at CERN are attracted by external file hosting services such as Dropbox. This trend may lead to what is know as the "Dropbox Problem": sensitive organization data stored on servers outside of corporate control, outside of established policies, outside of enforceable SLAs and in unknown geographical locations. Mitigating this risk also provides a good incentive to rethink how...
Dr
Mohammad Al-Turany
(GSI)
10/14/13, 4:10 PM
Event Processing, Simulation and Analysis
Oral presentation to parallel session
The FairRoot framework is the standard framework for simulation, reconstruction and data analysis for the FAIR experiments. The framework, is designed to optimize the accessibility for beginners and developers, to be flexible and to cope with future developments. FairRoot enhances the synergy between the different physics experiments within the FAIR project. Moreover, the framework is...
Belmiro Daniel Rodrigues Moreira
(LIP Laboratorio de Instrumentacao e Fisica Experimental (LIP)-Un)
10/14/13, 4:25 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Oral presentation to parallel session
CERN's Infrastructure as a Service cloud is being deployed in production across the two data centres in Geneva and Budapest.
This talk will describe the experiences of the first six months of production, the different uses within the organisation and the outlook for expansion to over 15,000 hypervisors based on OpenStack by 2015.
The open source toolchain used, accounting and scheduling...
Dr
Christopher Jones
(Fermi National Accelerator Lab. (US))
10/14/13, 4:25 PM
Data acquisition, trigger and controls
Oral presentation to parallel session
The DarkSide-50 dark matter experiment has recently been constructed and commissioned at the Laboratori Nazionali del Gran Sasso (LNGS). The data acquisition system for the experiment was jointly constructed by members of the LNGS Research Division and the Fermilab Scientific Computing Division, and it makes use of commercial, off-the-shelf hardware components and the artdaq DAQ software...
Dr
Friederike Nowak
(DESY)
10/14/13, 4:29 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Oral presentation to parallel session
In 2007, the National Analysis Facility (NAF) was set up within the framework of the Helmholtz Alliance "Physics at the Terascale", and is located at DESY. Its purpose was the provision of an analysis infrastructure for up-to-date research in Germany, complementing the Grid by offering a interactive access to the data. It has been well received within the physics community, and has proven to...
Tadashi Maeno
(Brookhaven National Laboratory (US))
10/14/13, 4:29 PM
Distributed Processing and Data Handling B: Experiment Data Processing, Data Handling and Computing Models
Oral presentation to parallel session
An important foundation underlying the impressive success of data processing and analysis in the ATLAS experiment at the LHC is the Production and Distributed Analysis (PanDA) workload management system. PanDA was designed specifically for ATLAS and proved to be highly successful in meeting all the distributed computing needs of the experiment. However, the core design of PanDA is not...
Stefan Lohn
(CERN)
10/14/13, 4:29 PM
Software Engineering, Parallelism & Multi-Core
Oral presentation to parallel session
Software optimization is a complex process, where the intended improvements have different effects on different platforms, with multiple operating systems and an ongoing introduction of new hardware. In addition several compilers produce differing object-code as result of different internal optimization procedures. To trace back the impact of the optimizations is going to become more...
Semen Lebedev
(Justus-Liebig-Universitaet Giessen (DE))
10/14/13, 4:30 PM
Event Processing, Simulation and Analysis
Oral presentation to parallel session
Development of fast and efficient event reconstruction algorithms is an important and challenging task in the Compressed Baryonic Matter (CBM) experiment at the future FAIR facility. The event reconstruction algorithms have to process terabytes of input data produced in particle collisions. In this contribution, several event reconstruction algorithms, which use available features of modern...
Dr
Wang Lu
(Institute of High Energy Physics,CAS)
10/14/13, 4:31 PM
Data Stores, Data Bases, and Storage Systems
Oral presentation to parallel session
Object storage systems based on Amazon’s Simple Storage Service (S3) have substantially developed in the last few years. The scalability, durability and elasticity characteristics of those systems make them well suited for a range of use cases where data is written, seldom updated and frequently read. Storage of images, static web sites and backup systems are some of the use cases where S3...
Pedro Andrade
(CERN)
10/14/13, 4:45 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Oral presentation to parallel session
At the present time computing centres are facing a massive rise in virtualization and cloud computing. The Agile Infrastructure (AI) project is working to deliver new solutions to ease the management of CERN Computing Centres. Part of the solution consists in a new common monitoring infrastructure which collects and manages monitoring data of all computing centre servers and associated...
Dr
Radoslaw Karabowicz
(GSI)
10/14/13, 4:45 PM
Data acquisition, trigger and controls
Oral presentation to parallel session
The PANDA experiment will be running up to 2.10^7 antiproton-proton
collisions per second at energies reaching 15 GeV.
The lack of simple features distinguishing the interesting events from background,
as well as strong pileup of events' data streams make the use of a
hardware trigger impossible. As a consequence the whole data stream of
about 300 GB/s has to be analyzed online, i.e:...
Wim Lavrijsen
(Lawrence Berkeley National Lab. (US))
10/14/13, 4:50 PM
Software Engineering, Parallelism & Multi-Core
Oral presentation to parallel session
The Python programming language brings a dynamic, interactive environment to physics analysis. With PyPy high performance can be delivered as well, when making use of its tracing just in time compiler (JIT) and cppyy for C++ bindings, as cppyy is able to exploit common HEP coding patterns. For example, ROOT I/O with cppyy runs at speeds equal to that of optimized, hand-tuned C++.
Python does...
Dr
Antonio Limosani
(University of Melbourne (AU))
10/14/13, 4:51 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Oral presentation to parallel session
The Australian Government is making a $AUD 100 million investment in
Compute and Storage for the academic community. The Compute facilities
are provided in the form of 24,000 CPU cores located at 8 nodes around
Australia in a distributed virtualized Infrastructure as a Service
facility based on OpenStack. The storage will eventually consist of
over 100 petabytes located at 6 nodes. All...
Dr
Michael Kirby
(Fermi National Accelerator Laboratory)
10/14/13, 4:51 PM
Distributed Processing and Data Handling B: Experiment Data Processing, Data Handling and Computing Models
Oral presentation to parallel session
The Fabric for Frontier Experiments (FIFE) project is a new far-reaching, major-impact initiative within the Fermilab Scientific Computing Division to drive the future of computing services for Fermilab Experiments. It is a collaborative effort between computing professionals and experiment scientists to produce an end-to-end, fully integrated set of services for computing on the grid and...
Seppo Sakari Heikkila
(CERN)
10/14/13, 4:54 PM
Data Stores, Data Bases, and Storage Systems
Oral presentation to parallel session
Cloud storage is an emerging architecture aiming to provide increased scalability and access performance, compared to more traditional solutions. CERN is evaluating this promise using Huawei UDS and OpenStack storage deployments, focusing on the needs of high-energy physics. Both deployed setups implement S3, one of the protocols that are emerging as standard in the cloud storage market. A set...
Stefanie Lewis
10/14/13, 4:55 PM
Event Processing, Simulation and Analysis
Oral presentation to parallel session
At the mass of a proton, the strong force is not well understood. Various quark models exist, but it is important to determine which quark model(s) are most accurate. Experimentally, finding resonances predicted by some models and not others would give valuable insight into this fundamental interaction. Several labs around the world use photoproduction experiments to find these missing...
Dr
Salvatore Tupputi
(Universita e INFN (IT))
10/14/13, 5:25 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Oral presentation to parallel session
The automation of ATLAS Distributed Computing (ADC) operations is essential to reduce manpower costs and allow performance-enhancing actions which improve the reliability of the system. In this perspective a crucial case is the automatic exclusion/recovery of ATLAS computing sites storage resources, which are continuously exploited at the edge of their capabilities.
It is challenging to...
Dr
Jamie Shiers
(CERN)
10/14/13, 5:25 PM
Data Stores, Data Bases, and Storage Systems
Oral presentation to parallel session
The international study group on data preservation in high energy physics, DPHEP, achieved a milestone in 2012 with the publication of its eagerly anticipated large scale report, which contains a description of data preservation activities from all major high energy physics collider-based experiments and laboratories. A central message of the report is that data preservation in HEP is not...
Michail Salichos
(CERN)
10/14/13, 5:25 PM
Distributed Processing and Data Handling B: Experiment Data Processing, Data Handling and Computing Models
Oral presentation to parallel session
FTS is the service responsible for distributing the majority of LHC data across the WLCG infrastructure. From the experiences of the last decade supporting and monitoring FTS, reliability, robustness and
high-performance data transfers has proved to be of high importance in the Data Management world. We are going to present the current status and features of the new File Transfer Service...
Pascal Costanza
(ExaScience Lab, Intel, Belgium)
10/14/13, 5:25 PM
Software Engineering, Parallelism & Multi-Core
Oral presentation to parallel session
Using Intel's SIMD architecture (SSE, AVX) to speed up operations on containers of complex class and structure objects is challenging, because it requires that the same data members of the different objects within a container have to be laid out next to each other, in a structure of arrays (SOA) fashion. Currently, programming languages do not provide automatic ways for arranging containers as...
Elizabeth Sexton-Kennedy
(Fermi National Accelerator Lab. (US))
10/14/13, 5:25 PM
Event Processing, Simulation and Analysis
Oral presentation to parallel session
Modern computing hardware is transitioning from using a single high frequency complicated computing core to many lower frequency simpler cores. As part of that transition, hardware manufacturers are urging developers to exploit concurrency in their programs via operating system threads. We will present CMS' effort to evolve our single threaded framework into a highly concurrent framework. We...
Dr
Jose Antonio Coarasa Perez
(CERN)
10/14/13, 5:25 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Oral presentation to parallel session
The CMS online cluster consists of more than 3000 computers. It has been exclusively used for the Data Acquisition of the CMS experiment at CERN, archiving around 20Tbytes of data per day. An openstack cloud layer has been deployed on part of the cluster (totalling more than 13000 cores) as a minimal overlay so as to leave the primary role of the computers untouched while allowing an...
Dr
Igor Oya
(Humboldt University)
10/14/13, 5:25 PM
Data acquisition, trigger and controls
Oral presentation to parallel session
The Cherenkov Telescope Array (CTA) is one of the major ground-based
astronomy projects being pursued and will be the largest facility for
ground-based gamma-ray observations ever built. CTA will consist of
two arrays: one in the Northern hemisphere composed of about 20
telescopes, and the other one in the Southern hemisphere composed of
about 100 telescopes, both arrays containing...
Peter Kreuzer
(Rheinisch-Westfaelische Tech. Hoch. (DE))
10/14/13, 5:45 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Oral presentation to parallel session
CMS is using a tiered setup of dedicated computing resources provided by sites distributed over the world and organized in WLCG. These sites pledge resources to CMS and are preparing them specially for CMS to run the experiment's applications. But there are more resources available opportunistically both on the GRID and in local university and research clusters which can be used for CMS...
Evan Niner
(Indiana University)
10/14/13, 5:45 PM
Data acquisition, trigger and controls
Oral presentation to parallel session
The NOvA detector utilizes not only a high speed streaming readout system which capable of reading out the waveforms of over 368,000 detector cells, but a distributed timing system that is able drive and program the frontend clock systems of each of these readout to allow each hit in the detector to be time stamped with a universal wall clock time. This system is used to perform an absolute...
Andrei Gheata
(CERN)
10/14/13, 5:46 PM
Software Engineering, Parallelism & Multi-Core
Oral presentation to parallel session
Among the components contributing to particle transport, geometry navigation is an important consumer of CPU cycles. The tasks performed to get answers to "basic" queries like locating a point within a geometry hierarchy or computing accurately the distance to the next boundary can become very computing intensive for complex detector setups. Among several optimization methods already in use by...
Brian Van Klaveren
(SLAC)
10/14/13, 5:47 PM
Distributed Processing and Data Handling B: Experiment Data Processing, Data Handling and Computing Models
Oral presentation to parallel session
The SLAC Computing Applications group (SCA) has developed a general
purpose data catalog framework, initially for use by the Fermi Gamma-Ray
Space Telescope, and now in use by several other experiments. The main
features of the data catalog system are:
* Ability to organize datasets in a virtual hierarchy without regard to
physical location or access protocol
* Ability to catalog...
Iban Jose Cabrillo Bartolome
(Universidad de Cantabria (ES))
10/14/13, 5:47 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Oral presentation to parallel session
The Altamira supercomputer at the Institute of Physics of Cantabria (IFCA) entered in operation in summer 2012.
Its last generation FDR Infiniband network used for message passing in parallel jobs, also supports the connection to General Parallel File System (GPFS) servers, enabling an efficient processing of multiple data demanding jobs at the same time.
Sharing a common GPFS system with...
Mike Hildreth
(University of Notre Dame (US))
10/14/13, 5:48 PM
Data Stores, Data Bases, and Storage Systems
Oral presentation to parallel session
Data and Software Preservation for Open Science (DASPOS), represents a first attempt to establish a formal collaboration tying together physicists from the CMS and ATLAS experiments at the LHC and the Tevatron experiments with experts in digital curation, heterogeneous high-throughput storage systems, large-scale computing systems, and grid access and infrastructure. Recently funded by the...
Benedikt Hegner
(CERN)
10/14/13, 5:50 PM
Event Processing, Simulation and Analysis
Oral presentation to parallel session
In the past, the increasing demands for HEP processing resources could be fulfilled by distributing the work to more and more physical machines. Limitations in power consumption of both CPUs and entire data centers are bringing an end to this era of easy scalability. To get the most CPU performance per Watt, future hardware will be characterised by less and less memory per processor, as well...
Jim Kowalkowski
(Fermilab)
10/15/13, 9:00 AM
Developments in concurrency (massive multi-core, GPU, and architectures such as ARM) are changing the physics computing landscape. In this talk dr Jim kowalkowski of Fermilab will describe on the use of GPU, massive multi-core, and the changes that result from massive parallelization and how this impacts data processing and models.
Mr
Philippe Canal
(Fermi National Accelerator Lab. (US))
10/15/13, 9:45 AM
Developments in many of our key software packages, such as Root 6 and the next generation Geant, will have a significant impact on the way analysis is done. Dr. Philippe Canal will present the birds-eye view on where these developments can lead us, on the way next generation ROOT and Geant can be combined, and on how for example the increased use of concurrency in these key software packages...
Dr
Torre Wenaus
(Brookhaven National Laboratory (US))
10/15/13, 11:00 AM
The computing for the LHC experiments has resulted in spectacular physics during the first few years of running. Now, the long shutdown offers the possibility to re-think some of the underlying concepts, look back to the lessons learned from this first run, and at the same work on revised models for the next after LS1. Dr Torre Wenaus of Brookhaven National Lab will talk about the revisions...
Stefano Spataro
(University of Turin)
10/15/13, 11:45 AM
For many experiments, e.g. those at the LHC, design choices made a very long time ago for the compute and trigger model are still used today. The incoming experiments have the opportunity to make new choices based on the current state of computing technology and novel ways to design the reconstruction frameworks, using the experience from previous experiments as well as already existing...
Martin Philipp Hellmich
(University of Edinburgh (GB))
10/15/13, 1:30 PM
Data Stores, Data Bases, and Storage Systems
Oral presentation to parallel session
Recent developments, including low power devices, cluster file systems and cloud storage, represent an explosion in the possibilities for deploying and managing grid storage. In this paper we present how different technologies can be leveraged to build a storage service with differing cost, power, performance, scalability and reliability profiles, using the popular DPM/dmlite storage solution...
Dr
Peter Elmer
(Princeton University (US))
10/15/13, 1:30 PM
Software Engineering, Parallelism & Multi-Core
Oral presentation to parallel session
In the last decade power limitations led to the introduction of
multicore CPU's. The cores on the processors were however not
dramatically different from the processors just before the
multicore-era. In some sense, this was merely a tactical choice to
maximize compatibility and buy time. The same scaling problems that
led to the power limit are likely to push processors in the...
Mario Ubeda Garcia
(CERN),
Victor Mendez Munoz
(PIC)
10/15/13, 1:30 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Oral presentation to parallel session
This contribution describes how Cloud resources have been integrated in the LHCb Distributed Computing. LHCb is using Dirac and its LHCb-specific extension LHCbDirac as an interware for its Distributed Computing. So far it was seamlessly integrating Grid resources and Computer clusters. The cloud extension of Dirac (VMDIRAC) extends it to the integration of Cloud computing...
Alessandro Lonardo
(INFN, Roma I (IT))
10/15/13, 1:30 PM
Data acquisition, trigger and controls
Oral presentation to parallel session
The integration of GPUs in trigger and data acquisition systems is currently being investigated in several HEP experiments.
At higher trigger levels, when the efficient many-core parallelization of event reconstruction algorithms is possible, the benefit of reducing significantly the number of the farm computing nodes is evident.
At lower levels, where tipically severe real-time...
Dr
Andrea Sciaba
(CERN)
10/15/13, 1:30 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Oral presentation to parallel session
The Wordwide LHC Computing Grid project (WLCG) provides the computing and storage resources required by the LHC collaborations to store, process and analyse their data. It includes almost 200,000 CPU cores, 200 PB of disk storage and 200 PB of tape storage distributed among more than 150 sites. The WLCG operations team is responsible for several essential tasks, such as the coordination of...
Christopher Jung
(KIT - Karlsruhe Institute of Technology (DE))
10/15/13, 1:30 PM
Distributed Processing and Data Handling B: Experiment Data Processing, Data Handling and Computing Models
Oral presentation to parallel session
Data play a central role in most fields of Science. In recent years, the amount of data from experiment, observation, and simulation has increased rapidly and the data complexity has grown. Also, communities and shared storage have become geographically more distributed. Therefore, methods and techniques applied for scientific data need to be revised and partially be replaced, while keeping...
Zachary Louis Marshall
(Lawrence Berkeley National Lab. (US))
10/15/13, 1:30 PM
Event Processing, Simulation and Analysis
Oral presentation to parallel session
In the 2011/12 data the LHC provided substantial multiple proton-proton collisions within each filled bunch-crossing and also multiple filled bunch-crossings within the sensitive time window of the ATLAS detector. This will increase in the near future during the run beginning in 2015. Including these effects in Monte Carlo simulation poses significant computing challenges. We present a...
Petr Zejdl
(CERN)
10/15/13, 1:50 PM
Data acquisition, trigger and controls
Oral presentation to parallel session
The CMS data acquisition (DAQ) infrastructure collects data from more than 600 custom detector Front End Drivers (FEDs). In the current implementation data is transferred from the FEDs via 3.2 Gbs electrical links to custom interface boards, which transfer the data to a commercial Myrinet network based on 2.5 Gbps optical links. During 2013 and 2014 the CMS DAQ system will undergo a major...
Ramon Medrano Llamas
(CERN)
10/15/13, 1:50 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Oral presentation to parallel session
HammerCloud was designed and born under the needs of the grid community to test the resources and automate operations from a user perspective. The recent developments in the IT space propose a shift to the software defined data centers, in which every layer of the infrastructure can be offered as a service.
Testing and monitoring is an integral part of the development, validation and...
Mrs
Tanya Levshina
(FERMILAB)
10/15/13, 1:52 PM
Distributed Processing and Data Handling B: Experiment Data Processing, Data Handling and Computing Models
Oral presentation to parallel session
The Open Science Grid (OSG) Public Storage project is focused on improving and simplifying the management of OSG Storage. Currently, OSG doesn’t provide efficient means to manage public storage offered by participating sites. A Virtual Organization (VO) that relies on opportunistic storage has difficulties finding appropriate storage, verifying its availability, and monitoring its...
Dag Larsen
(University of Silesia (PL))
10/15/13, 1:52 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Oral presentation to parallel session
Currently, the NA61/SHINE data production is performed on the CERN shared batch system, an approach inherited from its predecessor NA49. New data productions are initiated by manually submitting jobs to the batch system. An effort is now under way to migrate the data production to an automatic system, on top of a fully virtualised platform based on CernVM. There are several motivations for...
Sebastiano Schifano
(U)
10/15/13, 1:53 PM
Software Engineering, Parallelism & Multi-Core
Oral presentation to parallel session
An interesting evolution in scientific computing is represented by the
streamline introduction of co-processor boards that were originally built to
accelerate graphics rendering and that are now being used to perform
general computing tasks. A peculiarity of these boards (GPGPU, or
General Purpose Graphic Processing Units, and many-core boards like
the Intel Xeon Phi) is that they...
Dr
Paul Millar
(Deutsches Elektronen-Synchrotron (DE))
10/15/13, 1:53 PM
Data Stores, Data Bases, and Storage Systems
Oral presentation to parallel session
Storage is a continually evolving environment, with new solutions to both existing problems and new challenges. With over ten years in production use, dCache is also evolving to match this changing landscape. In this paper, we present three areas in which dCache is matching demand and driving innovation.
Providing efficient access to data that maximises both streaming and random-access...
Mike Hildreth
(University of Notre Dame (US))
10/15/13, 1:55 PM
Event Processing, Simulation and Analysis
Oral presentation to parallel session
Within the last year, design studies for LHC detector upgrades have begun to reach a level of detail that requires the simulation of physics processes with simulation performance at the level provided by Geant4. Full detector geometries for potential upgrades have been designed and incorporated into the CMS software. However, the extreme luminosities expected during the lifetimes of the...
Rainer Schwemmer
(CERN)
10/15/13, 2:10 PM
Data acquisition, trigger and controls
Oral presentation to parallel session
The architecture of the data acquisition for the LHCb upgrade is designed to allow for data transmission from the front-end electronics directly to the readout boards synchronously with the bunch crossing at the rate of 40 MHz. To connect the front-end electronics to the readout boards the upgraded detector will require order of 12000 GBT based (3.2 Gb/s radiation hard CERN serializers)...
Adriana Telesca
(CERN)
10/15/13, 2:10 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Oral presentation to parallel session
ALICE (A Large Ion Collider Experiment) is a heavy-ion detector studying the physics of strongly interacting matter and the quark-gluon plasma at the CERN LHC (Large Hadron Collider). The ALICE Data-AcQuisition (DAQ) system handles the data flow from the sub-detector electronics to the permanent data storage in the CERN computing center. The DAQ farm consists of about 1000 devices of many...
Dr
Robert Illingworth
(Fermilab)
10/15/13, 2:14 PM
Distributed Processing and Data Handling B: Experiment Data Processing, Data Handling and Computing Models
Oral presentation to parallel session
Fermilab Intensity Frontier experiments such as Minerva, NOvA, and MicroBooNE are now using an improved version of the Fermilab SAM data handling system. SAM was originally used by the CDF and D0 experiments for Run II of the Fermilab Tevatron to provide file metadata and location cataloguing, uploading of new files to tape storage, dataset management, file transfers between global processing...
Dr
David Colling
(Imperial College Sci., Tech. & Med. (GB))
10/15/13, 2:14 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Oral presentation to parallel session
The Higher Level Trigger (HLT) farm in CMS is a more than ten thousand core processor farm that is heavily used during data acquisition and largely unused when the detector is off. In this presentation we will cover the work done in CMS to utilize this large processing resource with cloud resource provisioning techniques. This resource when configured with Open Stack and Agile Infrastructure...
Robert Johannes Langenberg
(Technische Universitaet Muenchen (DE))
10/15/13, 2:15 PM
Event Processing, Simulation and Analysis
Oral presentation to parallel session
The track reconstruction algorithms of the ATLAS experiment have demonstrated excellent performance in all of the data delivered so far by the LHC. The expected large increase in the number of interactions per bunch crossing in the future introduce new challenges both in the computational aspects and physics performance of the algorithms. With the aim of taking advantage of modern CPU design...
Daniel Funke
(KIT - Karlsruhe Institute of Technology (DE))
10/15/13, 2:16 PM
Software Engineering, Parallelism & Multi-Core
Oral presentation to parallel session
The Compact Muon Solenoid (CMS) experiment at the Large Hadron Collider (LHC) at CERN near Geneva/Switzerland is a general-purpose particle detector which led, among many other results, to the discovery of a Higgs-like particle in 2012. It comprises the largest silicon-based tracking system built to date with 75 million individual readout channels and a total surface area of 205 m^2.
The...
Giacinto Donvito
(Universita e INFN (IT))
10/15/13, 2:16 PM
Data Stores, Data Bases, and Storage Systems
Oral presentation to parallel session
In this work we will show the testing activity carried on several distributed file-system in order to check the capability of supporting the HEP data analysis
In particular, we focused our attention and our test on HadoopFS, CEPH, and GlusterFS.
All are Open Source software.
HadoopFS is an Apache foundation software and is part of a more general framework, that contains: task...
Pawel Szostek
(CERN)
10/15/13, 2:30 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Oral presentation to parallel session
As Moore’s Law continues to deliver more and more transistors, the mainstream processor industry is preparing to expand its investments in areas other than simple core count. These new interests include deep integration of on-chip components, advanced vector units, memory, cache and interconnect technologies. We examine these moving trends with parallelized and vectorized High Energy Physics...
Josef Novy
(Czech Technical University (CZ))
10/15/13, 2:30 PM
Data acquisition, trigger and controls
Oral presentation to parallel session
The COMPASS is a fixed target experiment, situated at the Super Proton Synchrotron (SPS) accelerator in the north area of the CERN laboratory, in Geneva, Switzerland. The experiment was commissioned during 2001, data-taking started in 2002. The data acquisition system of the experiment is based on the DATE soft-ware package, originally developed for the ALICE experiment. In 2011, after the...
Dr
Dario Menasce
(INFN Milano-Bicocca)
10/15/13, 2:36 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Oral presentation to parallel session
Radiation detectors usually require complex calibration procedures in order to provide reliable activity measurements. The Milano-Bicocca group has developed, over the years, a complex simulation tool, based on GEANT4, that provide the functionality required to compute the correction factors necessary for such calibrations in a broad range of use-cases, considering various radioactive source...
Dr
Adam Lyon
(Fermilab)
10/15/13, 2:36 PM
Distributed Processing and Data Handling B: Experiment Data Processing, Data Handling and Computing Models
Oral presentation to parallel session
IFDH (Intensity Fronter Data Handling), is a suite of tools for data movement tasks for Fermilab experiments and is an important part of the FIFE (Fabric for Frontier Experiments) initiative described at this conference. IFDH encompasses moving input data from caches or storage elements to compute nodes (the "last mile" of data movement) and moving output data potentially to those caches as...
Rolf Edward Andreassen
(University of Cincinnati (US))
10/15/13, 2:38 PM
Software Engineering, Parallelism & Multi-Core
Oral presentation to parallel session
We present a general framework for maximum-likelihood fitting, in which GPUs are used to massively parallelise the per-event probability calculation. For realistic physics fits we achieve speedups, relative to executing the same algorithm on a single CPU, of several hundred.
Andreas Petzold
(KIT - Karlsruhe Institute of Technology (DE))
10/15/13, 2:39 PM
Data Stores, Data Bases, and Storage Systems
Oral presentation to parallel session
The need for storage continues to grow at a dazzling pace and science and society have become dependent on access to digital data. First sites storing an exabyte of data will be reality in a few years. The common storage technology in small and large computer centers continues to be magnetic disks because of their very good price performance ratio. Storage class memory and solid state disk...
Roberto Castello
(Universite Catholique de Louvain (BE))
10/15/13, 2:40 PM
Event Processing, Simulation and Analysis
Oral presentation to parallel session
Fast and efficient methods for the calibration and the alignment of the detector play a key role in ensuring reliable physics performance to an HEP experiment. CMS has set up a solid framework for alignment and calibration purpose, in close contact with the detector and physics needs. The about 200 types of calibration and alignment existing for the various sub-detectors are collected by...
Dr
Simon Patton
(LAWRENCE BERKELEY NATIONAL LABORATORY)
10/15/13, 3:45 PM
Distributed Processing and Data Handling B: Experiment Data Processing, Data Handling and Computing Models
Oral presentation to parallel session
In March 2012 the Dayabay Neutrino Experiment published the first measurement of the theta_13 mixing angle. The publication of this result occurred 20 days after the last data that appeared in the paper was taken, during which time normal data taking and processing was continuing. This achievement used over forty thousand 'core hours' of CPU and handled eighteen thousand files totaling 16 TBs....
Markus Frank
(CERN)
10/15/13, 3:45 PM
Data acquisition, trigger and controls
Oral presentation to parallel session
The LHCb experiment at the LHC accelerator at CERN collects collisions of particle bunches at 40 MHz. After a first level of hardware trigger with output of 1 MHz, the physically interesting collisions are selected by running dedicated trigger algorithms in the High Level Trigger (HLT) computing farm.
This farm consists of up to roughly 25000 CPU cores in roughly 1600 physical nodes each...
Lucien Boland
(University of Melbourne)
10/15/13, 3:45 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Oral presentation to parallel session
The Nectar national research cloud provides compute resources to Australian researchers using OpenStack. CoEPP, a WLCG Tier2 member, wants to use Nectar’s cloud resources for Tier 2 and Tier 3 processing for ATLAS and other experiments including Belle, as well as theoretical computation. CoEPP would prefer to use the Torque job management system in the cloud because they have extensive...
Vardan Gyurjyan
(Jefferson Lab)
10/15/13, 3:45 PM
Software Engineering, Parallelism & Multi-Core
Oral presentation to parallel session
The majority of developed physics data processing applications (PDP) are single, sequential processes that start at a point in time, and advance one step at a time until they are finished. In the current era of cloud computing and multi-core hardware architectures this approach has noticeable limitations.
In this paper we present a detailed evaluation of the FBP-based Clas12 event...
Phillip Urquijo
(Universitaet Bonn (DE))
10/15/13, 3:45 PM
Event Processing, Simulation and Analysis
Oral presentation to parallel session
The Belle II experiment is a future flavour factory experiment at the intensity frontier SuperKEKB e+e- collider, KEK Japan. Belle II is expected to go online in 2015, and collect a total of 50 ab-1 of data by 2022. The data will be used to study rare flavour phenomena in the decays of B- and D- mesons and tau-leptons, as well as heavy meson spectroscopy. Owing to the record breaking...
Dr
Tony Wildish
(Princeton University (US))
10/15/13, 3:45 PM
Data Stores, Data Bases, and Storage Systems
Oral presentation to parallel session
The data management elements in CMS are scalable, modular, and designed to work together. The main components are PhEDEx, the data transfer and location system; the Dataset Booking System (DBS), a metadata catalogue; and the Data Aggregation Service (DAS), designed to aggregate views and provide them to users and services. Tens of thousands of samples have been cataloged and petabytes of data...
Jason Alexander Smith
(Brookhaven National Laboratory (US))
10/15/13, 3:45 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Oral presentation to parallel session
Solid state drives (SSDs) provide significant improvements in random I/O performance over traditional rotating SATA and SAS drives. While the cost of SSDs has been steadily declining over the past few years, high density SSDs continue to remain prohibitively expensive when compared to traditional drives. Currently, 1TB SSDs generally cost more than USD $1,000, while 1TB SATA drives typically...
Shawn Mc Kee
(University of Michigan (US)),
Simone Campana
(CERN)
10/15/13, 4:05 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Oral presentation to parallel session
The WLCG infrastructure moved from a very rigid network topology, based on the MONARC model, to a more relaxed system, where data movement between regions or countries does not necessarily need to involve T1 centers. While this evolution brought obvious advantages, especially in terms of flexibility for the LHC experiment’s data management systems, it also opened the question of how to monitor...
Tomasz Bold
(AGH Univ. of Science and Technology, Krakow)
10/15/13, 4:05 PM
Data acquisition, trigger and controls
Oral presentation to parallel session
The high level trigger (HLT) of the ATLAS experiment at the LHC selects interesting proton-proton and heavy ion collision events for the wide ranging ATLAS physics program. The HLT examines events selected by the level-1 hardware trigger using a combination of specially designed software algorithms and offline reconstruction algorithms. The flexible design of the entire trigger system was...
Georgios Lestaris
(CERN)
10/15/13, 4:07 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Oral presentation to parallel session
In a virtualized environment, contextualization is the process of configuring a VM instance for the needs of various deployment use cases. Contextualization in CernVM can be done by passing a handwritten context to the “user data” field of cloud APIs, when running CernVM on the cloud, or by using CernVM web interface when running the VM locally. CernVM online is a publicly accessible web...
Niko Neufeld
(CERN)
10/15/13, 4:07 PM
Software Engineering, Parallelism & Multi-Core
Oral presentation to parallel session
The ARM architecture is a power-efficient design that is used in most processors in mobile devices all around the world today since they provide reasonable compute performance per watt. The current LHCb software stack is designed (and expected) to build and run on machines with the x86/x86_64 architecture. This paper outlines the process of measuring the performance of the LHCb software stack...
Juan Carlos Diaz Velez
(University of Wisconsin-Madison)
10/15/13, 4:07 PM
Distributed Processing and Data Handling B: Experiment Data Processing, Data Handling and Computing Models
Oral presentation to parallel session
IceProd is a data processing and management framework developed by IceCube Neutrino Observatory for processing of Monte Carlo simulations and data. IceProd runs as a separate layer on top of middleware and can take advantage of a variety of computing resources including grids and batch systems such as GLite, Condor, NorduGrid, PBS and SGE. This is accomplished by a set of dedicated daemons...
Vincent Garonne
(CERN)
10/15/13, 4:08 PM
Data Stores, Data Bases, and Storage Systems
Oral presentation to parallel session
Rucio is the next-generation Distributed Data Management (DDM) system benefiting from recent advances in cloud and "Big Data" computing to address HEP experiments scaling requirements. Rucio is an evolution of the ATLAS DDM system Don Quijote 2 (DQ2), which has demonstrated very large scale data management capabilities with more than 140 petabytes spread worldwide across 130 sites, and...
Mathias Michel
(Helmholtz-Institut Mainz)
10/15/13, 4:10 PM
Event Processing, Simulation and Analysis
Oral presentation to parallel session
A large part of the physics program of the PANDA experiment at FAIR
deals with the search for new conventional and exotic hadronic states
like e.g. hybrids and glueballs. In a majority of analyses PANDA will
need a Partial Wave Analsis (PWA) to identify possible candidates and
for the classification of known states. Therefore, a new, agile and
efficient PWA-Framework will be...
Dr
Gabriele Garzoglio
(FERMI NATIONAL ACCELERATOR LABORATORY)
10/15/13, 4:25 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Oral presentation to parallel session
As the need for Big Data in science becomes ever more relevant, networks around the world are upgrading their infrastructure to support high-speed interconnections. To support its mission, the high-energy physics community as a pioneer in Big Data has always been relying on the Fermi National Accelerator Laboratory to be at the forefront of storage and data movement. This need was reiterated...
Dr
Remi Mommsen
10/15/13, 4:25 PM
Data acquisition, trigger and controls
Oral presentation to parallel session
The DAQ system of the CMS experiment at the LHC is redesigned during the accelerator shutdown in 2013/14. To reduce the interdependency of the DAQ system and the high-level trigger (HLT), we investigate the feasibility of using a file-system-based HLT. Events of ~1 MB size are built at the level-1 trigger rate of 100 kHz. The events are assembled by ~50 builder units (BUs). Each BU writes the...
Graeme Andrew Stewart
(CERN)
10/15/13, 4:29 PM
Distributed Processing and Data Handling B: Experiment Data Processing, Data Handling and Computing Models
Oral presentation to parallel session
The need to run complex workflows for a high energy physics experiment such as ATLAS has always been present. However, as computing resources have become even more constrained, compared to the wealth of data generated by the LHC, the need to use resources efficiently and manage complex workflows within a single grid job have increased.
In ATLAS, a new Job Transform framework has been...
Mr
Davide Salomoni
(INFN CNAF), Dr
Elisabetta Ronchieri
(INFN CNAF), Mr
Marco Canaparo
(INFN CNAF), Mr
Vincenzo Ciaschini
(INFN CNAF)
10/15/13, 4:29 PM
Software Engineering, Parallelism & Multi-Core
Oral presentation to parallel session
Software packages in our scientific environment are constantly growing in size, and are written by any number of developers. This implies a strong churn on the code itself, and an associated
risk of bugs and stability problems. This risk is unavoidable as long as the software undergoes active evolution, as it always happens with software that is still in use. However, the necessity of having...
Jakob Blomer
(CERN)
10/15/13, 4:29 PM
Distributed Processing and Data Handling A: Infrastructure, Sites, and Virtualization
Oral presentation to parallel session
The traditional virtual machine building and and deployment process is centered around the virtual machine hard disk image. The packages comprising the VM operating system are carefully selected, hard disk images are built for a variety of different hypervisors, and images have to be distributed and decompressed in order to instantiate a virtual machine. Within the HEP community, the CernVM...
Dr
Tom Whyntie
(Queen Mary, University of London/The Langton Star Centre)
10/15/13, 4:30 PM
Event Processing, Simulation and Analysis
Oral presentation to parallel session
The Langton Ultimate Cosmic ray Intensity Detector (LUCID) experiment [1] is a satellite-based device that uses five Timepix hybrid silicon pixel detectors [2] to make measurements of the radiation environment at an altitude of approximately 660km, i.e. in Low Earth Orbit (LEO). The experiment is due to launch aboard Surrey Satellite Technology Limited's (SSTL's) TechDemoSat-1 in Q3 of 2013....
Ilija Vukotic
(University of Chicago (US))
10/15/13, 4:31 PM
Data Stores, Data Bases, and Storage Systems
Oral presentation to parallel session
In the past year the ATLAS Collaboration has accelerated its program to federate data storage resources using an architecture based on XRootD with its attendant redirection and storage integration services. The main goal of the federation is an improvement in the data access experience for the end user while allowing for more efficient and intelligent use of computing resources by monitoring...
David Gutierrez Rueda
(CERN)
10/15/13, 4:45 PM
Facilities, Production Infrastructures, Networking and Collaborative Tools
Oral presentation to parallel session
The network infrastructure at CERN has evolved with the increasing service and bandwidth demands of the scientific community. Analysing the massive amounts of data gathered by the experiments requires more computational power and faster networks to carry the data. The new Data Centre in Wigner and the adoption of 100Gbps in the core of the network are the latest answers to these demands. In...
Mr
Pierre Vande Vyvre
(CERN)
10/15/13, 4:45 PM
Data acquisition, trigger and controls
Oral presentation to parallel session
for the ALICE O2 Collaboration
ALICE (A Large Ion Collider Experiment) is a heavy-ion detector
studying the physics of strongly interacting matter and the
quark-gluon plasma at the CERN LHC (Large Hadron Collider).
After the second long shutdown of the LHC, the ALICE detector will
be upgraded in order to make high precision measurements of rare
probes at low pT, which cannot be...