Help us make Indico better by taking this survey! Aidez-nous à améliorer Indico en répondant à ce sondage !

HEPiX Fall 2014 Workshop

America/Chicago
Nebraska Union - Regency Suite (University of Nebraska - Lincoln)

Nebraska Union - Regency Suite

University of Nebraska - Lincoln

<b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
Brian Paul Bockelman (University of Nebraska (US)), Helge Meinhard (CERN), Sandy Philpott (JLAB)
Description

HEPiX Fall 2014 at University of Nebraska-Lincoln

The HEPiX forum brings together worldwide Information Technology staff, including system administrators, system engineers, and managers from the High Energy Physics and Nuclear Physics laboratories and institutes, to foster a learning and sharing experience between sites facing scientific computing and data challenges.
Participating sites include BNL, CERN, DESY, FNAL, IN2P3, INFN, IRFU, JLAB, NIKHEF, PIC, RAL, SLAC, TRIUMF and many others.
 

HEPiX Fall 2014 is proudly sponsored by Western Digital at the platinum level and DDN at the silver level.

 

Participants
  • Aaron Dominguez
  • Adam Caprez
  • Ahmed Khoudir
  • Ajit Mohapatra
  • Alexander Pattison
  • Alexandr Zaytsev
  • Alvaro Simon
  • Andrea Chierici
  • Andrea Manzi
  • Andreas Haupt
  • Andreas Petzold
  • Anthony Tiradani
  • Aresh Vedaee
  • Aris Angelogiannopoulos
  • Arne Wiebalck
  • Ben Jones
  • Benjeman Meekhof
  • Bjorn Barrefors
  • Bob Cowles
  • Bockjoo Kim
  • Borja Aparicio Cotarelo
  • Brian Behlendorf
  • Brian Bockelman
  • Carl Lundstedt
  • Chi Woong Kim
  • Constance Sieh
  • David Kelsey
  • David Sanford
  • David Swanson
  • Dennis van Dok
  • Derek Weitzel
  • Dimitri Bourilkov
  • Domokos Szabo
  • Edgar Fajardo Hernandez
  • Edward Karavakis
  • Elizabeth Bautista
  • Emelie Harstad
  • Erik Gough
  • Fazhi QI
  • Frederic Schaer
  • Frederique Chollet
  • Garhan Attebury
  • Helge Meinhard
  • Ian Peter Collier
  • Jaime Frey
  • James Borden
  • James Botts
  • James Pryor
  • Jan Trautmann
  • Jean-Claude DESVIGNES
  • Jeff Dost
  • jin kim
  • Jingyan Shi
  • Joe Metzger
  • Josh Samuelson
  • Jérôme Belleman
  • Keith Chadwick
  • Kenneth Bloom
  • Lincoln Bryant
  • Lisa Giacchetti
  • Liviu Valsan
  • Luca Mascetti
  • Lueders Stefan
  • Majid Arabgol
  • Marian Zvada
  • Martin Bly
  • Matthew Baxa
  • Mattias Wadenstein
  • MICHAIL SALICHOS
  • Michel Jouvin
  • Michele Michelotto
  • Pat Riehecky
  • Paul Kuipers
  • Pete Gronbech
  • Ray Spence
  • Rob Quick
  • Roberto Gomezel
  • Ruben Domingo Gaspar Aparicio
  • Samir Siqueira
  • Sandra Philpott
  • Sean Crosby
  • Shawn McKee
  • Szabolcs Hernath
  • Sébastien Gadrat
  • Terrence Martin
  • Thomas Oulevey
  • Tim Skirvin
  • Todd Tannenbaum
  • Tony Quan
  • Tony Wong
  • Ulf Tigerstedt
  • Walter Schoen
  • Wayne Salter
    • 08:30
      Registration and Coffee Georgian Suite

      Georgian Suite

      University of Nebraska - Lincoln

    • HEPiX Business Nebraska Union - Georgian Suite

      Nebraska Union - Georgian Suite

      slides
      • 1
        Local Organizer Info
        Speaker: Brian Paul Bockelman (University of Nebraska (US))
        Slides
    • Site Reports Nebraska Union - Georgian Suite

      Nebraska Union - Georgian Suite

      • 2
        Australia Site Report
        An update on the ATLAS Tier 2 and distributed Tier 3 of HEP groups in Australia. Will talk about our integration of Cloud resources, Ceph filesystems and integration of 3rd party storage into our setup
        Speaker: Sean Crosby (University of Melbourne (AU))
        Slides
      • 3
        Oxford Particle Physics Computing update
        Site report from the University of Oxford Physics department.
        Speaker: Peter Gronbech (University of Oxford (GB))
        Slides
      • 4
        Site report: NDGF-T1
        Site report for NDGF-T1, mainly focusing on dCache.
        Speakers: Erik Mattias Wadenstein (University of Umeå (SE)), Ulf Tigerstedt (CSC Oy)
        Slides
      • 5
        Updates from Jefferson Lab HPC and Scientific Computing
        An overview since our spring meeting on JLab's latest developments for 12 GeV physics computing and storage, Lustre update, openZFS plan, load balancing between HPC and data analysis, Facilities changes in the Data Center, ...
        Speaker: Sandy Philpott (JLAB)
        Slides
      • 6
        KIT Site Report
        News about GridKa Tier-1 and other KIT IT projects and infrastructure.
        Speaker: Andreas Petzold (KIT - Karlsruhe Institute of Technology (DE))
        Slides
    • 10:30
      Coffee Break Nebraska Union - Georgian Suite

      Nebraska Union - Georgian Suite

    • Site Reports Nebraska Union - Georgian Suite

      Nebraska Union - Georgian Suite

      • 7
        Fermilab Site Report - Fall 2014 HEPiX
        Fermilab Site Report - Fall 2014 HEPiX
        Speaker: Dr Keith Chadwick (Fermilab)
        Slides
      • 8
        University of Wisconsin Madison CMS T2 site report
        As a major WLCG/OSG T2 site, the University of Wisconsin Madison CMS T2 has provided very productive and reliable services for CMS MonteCarlo production/processing, and large scale global CMS physics analysis using high throughput computing (HT-Condor), highly available storage system (Hadoop), efficient data access using xrootd/AAA, and scalable distributed software systems (CVMFS). An update on the current status of and activities (since the last report at Ann Arbor meeting) at the UW Madison Tier-2 will be presented that includes efforts on the 100Gb network upgrade and IPv6 etc., among other things.
        Speakers: Ajit mohapatra (University of Wisconsin (US)), Tapas Sarangi (University of Wisconsin (US))
        Slides
      • 9
        INFN-T1 site report
        Updates on INFN Tier1 site
        Speaker: Andrea Chierici (INFN-CNAF)
        Slides
    • 12:00
      Lunch Downtown Nebraska - on your own

      Downtown Nebraska - on your own

    • IT End User and Operating Systems Nebraska Union - Georgian Suite

      Nebraska Union - Georgian Suite

      • 10
        Scientific Linux current status update
        This presentation will provide an update on the current status of Scientific Linux, descriptions for some possible future goals, and allow a chance for users to provide feedback on its direction.
        Speaker: Pat Riehecky (Fermilab)
        Slides
      • 11
        Next Linux Version at CERN.
        CERN is maintaining and deploying Scientific Linux CERN since 2004. In January 2014 CentOS and Red Hat announced joining forces in order to provide common platform for open source community project needs. CERN decided to see how CentOS 7 fits his needs and evaluate CentOS release 7 as their next version. An updated report will be provided, as agreed at HEPiX Spring 2014.
        Speaker: Thomas Oulevey (CERN)
        Slides
      • 12
        Experiences with EL 7 at T2_US_Nebraska
        Seven years have passed since the initial EL 5 release and yet it's still found in active use at many sites. The successor EL 6 is also showing age with its 4th birthday just around the corner. While both are still under support from RedHat for many years to come, it never hurts to prepare for the future. This talk will detail the experiences at T2_US_Nebraska in transitioning towards EL 7 using CentOS 7 as a base. Highlights will include the major differences between EL6 and EL7 and how they relate to our daily operations as a Tier2 CMS site.
        Speaker: Garhan Attebury (University of Nebraska (US))
        Slides
      • 13
        FTS3, large scale file transfer service with simplicity and performance at its best
        FTS3 is the service responsible for globally distributing the majority of the LHC data across the WLCG infrastructure. It is a file transfer scheduler which scales horizontally and it's easy to install and configure. In this talk we would like to bring the attention to the FTS3 features that could attract wider communities and administrators with several new friendly features. We will present both the new tools for the management of the FTS3 transfer parameters, for instance bandwidth-limits, max active transfers per endpoint and VO, banning users and endpoints, plus new Data Management operations (deletions and staging files from archive) easily accessed via REST-API. In addition we will also showcase the new captivating FTS3 Graphical Interface for end-users to manage their transfers (WebFTS) together with the new activities to extend the FTS3 transfers capabilities outside the grid boundaries ( Dropbox, S3, etc.) In this manner we demonstrate that FTS3 can cover the needs from casual users to high load services.
        Speaker: Michail Salichos (CERN)
        Slides
    • 15:30
      Coffee Break Nebraska Union - Georgian Suite

      Nebraska Union - Georgian Suite

    • IT End User and Operating Systems Nebraska Union - Georgian Suite

      Nebraska Union - Georgian Suite

      • 14
        Issue Tracking and Version Control Services status update
        The current efforts around the Issue Tracking and Version Control services at CERN will be presented. Their main design and structure will be shown giving special attention to the new requirements from the community of users in terms of collaboration and integration tools and how we address this challenge in the definition of new services based on GitLab for collaboration and Code Review and Jenkins for Continuous Integration. The presentation will also address infrastructure issues for these services, such as the back-end storage, administration portal "Cernforge" for power-users and experience using Puppet for managing the server software configuration.
        Speaker: Borja Aparicio Cotarelo (CERN)
        Slides
    • Opening Reception Schorr Center, 2nd Floor

      Schorr Center, 2nd Floor

      University of Nebraska - Lincoln

    • 08:30
      Registration and Coffee Regency Suite

      Regency Suite

      University of Nebraska - Lincoln

    • IT End User and Operating Systems Nebraska Union - Regency Suite

      Nebraska Union - Regency Suite

      University of Nebraska - Lincoln

      <b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
      • 15
        Monviso: a portal for metering and reporting CNAF resources usage
        CNAF T1 monitoring and alarming systems produce tons of data describing state, performance and usage of our resources. Collecting this kind of information centrally would benefit both resource administrators and our user community in processing information and generating reporting graphs. We built the “Monviso reporting portal” that consumes a set of key metrics, graphing them based on two main viewpoints: resource administration and experiments support. The resulting portal is a lightweight charts gallery used also by our operator-on-call.
        Speaker: Andrea Chierici (INFN-CNAF)
        Slides
    • Site Reports Nebraska Union - Regency Suite

      Nebraska Union - Regency Suite

      University of Nebraska - Lincoln

      <b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
      • 16
        CERN Site Report
        News from CERN since the Annecy meeting.
        Speaker: Dr Arne Wiebalck (CERN)
        Slides
      • 17
        IRFU site report
        In this site report, we will speak about what changed at CEA/IRFU and what has been interesting since Hepix@Annecy, 6 months ago.
        Speaker: Mr Frederic Schaer (CEA)
        Slides
      • 18
        DESY Site Report
        News from DESY since the Annecy meeting.
        Speaker: Andreas Haupt (Deutsches Elektronen-Synchrotron (DE))
        Slides
      • 19
        BNL RACF Site Report
        A summary of developments at BNL's RHIC/ATLAS Computing Facility since the last HEPiX meeting.
        Speaker: James Pryor (B)
        Slides
    • 10:30
      Coffee Break Nebraska Union - Regency Suite

      Nebraska Union - Regency Suite

      University of Nebraska - Lincoln

      <b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
    • Site Reports Nebraska Union - Regency Suite

      Nebraska Union - Regency Suite

      University of Nebraska - Lincoln

      <b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
      • 20
        AGLT2 Site Report Fall 2014
        I will present an update on our site since the last report and cover our work with dCache, perfSONAR-PS and VMWare. I will also report on our recent hardware purchases for 2014 as well as the status of our new networking configuration and 100G connection to the WAN. I conclude with a summary of what has worked and what problems we encountered and indicate directions for future work.
        Speaker: Shawn Mc Kee (University of Michigan (US))
        Slides
      • 21
        RAL Site Report
        Latest from RAL Tier1
        Speaker: Martin Bly (STFC-RAL)
        Slides
    • Grids, Clouds, Virtualisation Nebraska Union - Regency Suite

      Nebraska Union - Regency Suite

      University of Nebraska - Lincoln

      <b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
      • 22
        LHC@home status - Outlook for wider use of volunteer computing at CERN
        [LHC@home][1] was brought back to CERN-IT in 2011, with 2 projects; Sixtrack and Test4Theory, the latter using virtualization with CernVM. Thanks to this development, there is increased interest in volunteer computing at CERN, notably since native virtualization support has been added to the BOINC middleware. Pilot projects with applications from the LHC experiment collaborations running on CernVM have also been deployed, opening the perspective for wider use of BOINC also for High Energy Physics software. The presentation will address the current status of LHC@home and the evolution of the CERN BOINC service to address the needs of a wider range of applications and users. [1]: http://cern.ch/lhcathome
        Speaker: Dr Helge Meinhard (CERN)
        Slides
    • 12:00
      Lunch Nebraska Union - Regency Suite

      Nebraska Union - Regency Suite

      University of Nebraska - Lincoln

      <b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
    • Computing and Batch Systems Nebraska Union - Regency Suite

      Nebraska Union - Regency Suite

      University of Nebraska - Lincoln

      <b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
      • 23
        Benchmarking on System on Chip Archtecture and fast benchmarking
        The traditional architecture for High Energy Physics is x86-64 but in the community there is interest in processor more efficient in term of computing power per Watt. I'll show my measurement on ARM and Avoton processor. I'll conclude with some measurements on candidate for fast benchmark that are requested by the physics community, mostyl to measure the performance of machine in cloud.
        Speaker: Michele Michelotto (Universita e INFN (IT))
        Slides
      • 24
        Future of Batch Processing at CERN: a Condor Pilot Service
        The CERN Batch System comprises 4000 worker nodes, 60 queues and offers a service for various types of large user communities. In light of the developments driven by the Agile Infrastructure and the more demanding processing requirements, it will be faced with increasingly challenging scalability and flexibility needs. Last HEPiX, we presented the results of our evaluation of SLURM, Grid Engine derivatives and HTCondor. The latter being the most promising one, we started setting up an HTCondor pilot service. This talk will present the initial functions provided by this service. We will then discuss the development steps we took to implement them. Finally, we will name the next features which will ultimately be leading to the new production CERN Batch Service.
        Speaker: Jerome Belleman (CERN)
        Slides
      • 25
        Compute node benchmarks for Compact Muon Solenoid workflows
        Hardware benchmarks are often relative to the target application. In CMS sites, new technologies, mostly processors, need to be evaluated on an yearly basis. A framework was developed at the Caltech CMS Tier-2 to benchmark compute nodes with one of the most CPU-intensive CMS workflows - The Tier-0 Reconstruction. The benchmark is a CMS job that reports the results to a central database based on CPU model and makes them available to real-time monitoring web interfaces. The goal is to provide to the collaboration a reference for CPU performance, which can also be used in automated systems through an API. The jobs run in parallel to normal Grid activity and could have their submission and reporting automated.
        Speaker: Samir Cury Siqueira (California Institute of Technology (US))
        Slides
      • 26
        HTCondor and HEP Partnership and Activities
        The goal of the HTCondor team is to to develop, implement, deploy, and evaluate mechanisms and policies that support High Throughput Computing (HTC) on large collections of distributively owned computing resources. Increasingly, the work performed by the HTCondor developers is being driven by its partnership with the High Energy Physics (HEP) community. This presentation will provide an overview of how the HTCondor developers and the HEP community currently work together to advance the capabilities of batch computing, and it will also expose some of the new HTCondor functionality currently under development.
        Speaker: Todd Tannenbaum (Univ of Wisconsin-Madison, Wisconsin, USA)
        Slides
      • 27
        Releasing the HTCondor-CE into the Wild
        One of the most critical components delivered by the Open Science Grid (OSG) software team is the compute element, or the OSG-CE. At the core of the CE itself is the gatekeeper software for translating grid pilot jobs into local batch system jobs. OSG is in the process of migrating from the Globus gatekeeper to the HTCondor-CE, supported by the HTCondor team. The HTCondor-CE provides an alternate view on how grid gatekeepers can offer provisioning services, with a model that significantly differs from other gatekeepers such as GRAM or CREAM. Further, the HTCondor-CE is not a standalone product in itself but a specialized configuration of the familiar HTCondor software. Basing the HTCondor-CE on the much larger HTCondor product will allow a rich set of new features in the future (with low development costs!). In this presentation, I'll highlight some of the biggest technical similarities and differences between the HTCondor-CE and other gatekeepers. Further, I'll discuss some of the non-technical considerations (documentation, operations, rollout, etc) we had to take into account in managing such a large-scale software transition in the OSG Production Grid environment.
        Speaker: Brian Paul Bockelman (University of Nebraska (US))
        Slides
    • 15:40
      Coffee Break Nebraska Union - Regency Suite

      Nebraska Union - Regency Suite

      University of Nebraska - Lincoln

      <b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
    • Computing and Batch Systems Nebraska Union - Regency Suite

      Nebraska Union - Regency Suite

      University of Nebraska - Lincoln

      <b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
      • 28
        HTCondor on the Grid and in the Cloud
        An important use of HTCondor is as a scalable, reliable interface for jobs destined for other scheduling systems. These include Grid intefaces to batch systems (Globus, CREAM, ARC) and Cloud services (EC2, OpenStack, GCE). The High Energy Physics community has been a major user of this functionality and has driven its development. This talk will provide an overview of HTCondor's Grid and Cloud capabilities, how they're being employed, and our plans for future functionality.
        Speaker: James Frey
        Slides
    • Networking and Security Nebraska Union - Regency Suite

      Nebraska Union - Regency Suite

      University of Nebraska - Lincoln

      <b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
      • 29
        Evaluating Infiniband Based Networking Solutions for HEP/NP Data Processing Applications
        The Infiniband networking technology is a long established and rapidly developing technology which is currently dominating the field of low-latency, high-throughput interconnects for HPC systems in general and those included in the TOP-500 list in particular. Over the last 4 years a successful use of Infiniband networking technology combined with additional IP-over-IB protocol and Infiniband to Ethernet bridging layers was demonstrated well beyond the realm of HPC, covering various high throughput computing (HTC) systems, including data processing farms and private clouds devoted to HEP/NP data processing. With the recent advances of Mellanox VPI technology in 2013-2014 the 4X FDR IB now stands as the most versatile networking solution available for existing and future data centers that need to support both HTC and HPC oriented activities that can be seamlessly integrated into the existing Ethernet based infrastructure. Furthermore, it can be done completely transparently for the end users of these facilities, though certain modifications of the end user’s activity-patterns are needed in order to utilize the full potential of the Infiniband based networking infrastructure. This contribution contains a detailed report on the series of tests and evaluation activities performed within the RACF over the last year in order to evaluate a Mellanox 4X FDR Infiniband based networking architecture (provided with an oversubscribed tree topology) as a potential alternative networking solution for both the RHIC and ATLAS data processing farms of the RACF, as well as the existing dCache and future Ceph based storage systems associated with them. Results of the price/performance comparison of such a networking system with a competing solution based on the 10 GbE technology (provided with non-blocking fabric topology) for a HEP/NP data processing farm consisting of 1500 compute nodes are presented. Job placement optimizations in Condor for the offline data processing farm of the PHENIX experiment were implemented in order to demonstrate a sample user activity-pattern that more optimally utilizes the Infiniband-based networking solution. The results of using those optimizations in production for the PHENIX experiment over the last 9 months are presented.
        Speaker: Mr Alexandr Zaytsev (Brookhaven National Laboratory (US))
        Slides
    • HEPiX Business: HEPiX Board Meeting Nebraska Union - Regency Suite

      Nebraska Union - Regency Suite

      University of Nebraska - Lincoln

      <b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
      slides
    • 08:30
      Morning Coffee Regency Suite

      Regency Suite

      University of Nebraska - Lincoln

    • Networking and Security Nebraska Union - Regency Suite

      Nebraska Union - Regency Suite

      University of Nebraska - Lincoln

      <b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
      • 30
        Plans for Dual Stack IPv4/IPv6 services on WLCG - an update from the HEPiX IPv6 Working Group
        This talk will present an update on the recent activities of the HEPiX IPv6 Working Group including our plans for moving to dual-stack services on WLCG.
        Speaker: Dave Kelsey (STFC - Rutherford Appleton Lab. (GB))
        Slides
      • 31
        EEX: ESnet Extension to Europe and ESnet support for the LHC Community
        The ESnet Extension to Europe (EEX) project is building out the ESnet backbone in to Europe. The goal of the project is to provide dedicated transatlantic network services that support U.S. DOE funded science. The EEX physical infrastructure build will be substantially completed before the end of December. Initial services will be provided to BNL, FERMI and CERN while the infrastructure is being built out and tested. EEX services will be expanded, following the build, to serve all current ESnet sites and the U.S. LHC community including some computing centers at U.S. Universities. This talk will cover the EEX architecture, schedule, and services, and include initial thoughts about what universities will need to do to take advantage of this opportunity.
        Speaker: Joe Metzger (LBL)
        Slides
      • 32
        Do You Need to Know Your Users?
        After several years investigation of trends in Identity Management (IdM), the eXtreme Scale Identity Management (XSIM) project has concluded there is little reason for resource providers to provide IdM functions for research collaborations or even for many groups within the institution. An improved user experience and decreased cost can be achieved with "a small amount of programming."
        Speaker: Bob Cowles (BrightLite Information Security)
        Slides
    • 10:30
      Coffee Break Nebraska Union - Regency Suite

      Nebraska Union - Regency Suite

      University of Nebraska - Lincoln

      <b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
    • Networking and Security Nebraska Union - Regency Suite

      Nebraska Union - Regency Suite

      University of Nebraska - Lincoln

      <b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
      • 33
        OSG IPv6 Software and Operations Preparations
        OSG Operations and Software will soon be configuring our operational infrastructure and middleware components with an IPv6 network stack capabilities in addition to its existing IPv4 stack. For OSG services this means network interfaces will thus have at least one IPv6 address on which it listens, in addition to whatever IPv4 addresses it is already listening on. For middleware components we will test dual stacked servers versus dual-stacked clients and IPv4 only clients.
        Speaker: Robert Quick (Indiana University)
        Slides
      • 34
        Situational Awareness: Computer Security
        Computer security is important as ever outside the HEP community, but also within. This presentation will give the usual overview on recent issues being reported or made public since the last HEPix workshop (like the ripples of "Heartbleed"). It will discuss trends (identity federation and virtualisation) and potential mitigations to new security threats.
        Speaker: Dr Stefan Lueders (CERN)
        Slides
    • 12:00
      Lunch Nebraska Union - Regency Suite

      Nebraska Union - Regency Suite

      University of Nebraska - Lincoln

      <b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
    • IT Facilities and Business Continuity Nebraska Union - Regency Suite

      Nebraska Union - Regency Suite

      University of Nebraska - Lincoln

      <b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
      • 35
        UPS Monitoring with Sensaphone-A cost-effective solution
        We describe a cost-effective indirect UPS monitoring system that was implemented recently in parts of its RACF complex. This solution was needed to address a lack of centralized monitoring solution, and it is integrated with an event notification mechanism and overall facility management.
        Speaker: Dr Tony Wong (Brookhaven National Laboratory)
        Slides
      • 36
        First Experience with the Wigner Data Centre
        After a tender for a CERN remote Tier0 centre issued at the end of 2011, and awarded to the Wigner Data Centre in May 2012, operations commenced at the beginning of 2013. This talk will give a brief introduction to the history of this project and it scope. It will then summarise the initial experience that has been gained to-date and highlight a number of issues that have been encountered; some maybe expected but others not.
        Speaker: Wayne Salter (CERN)
        Slides
    • Storage and Filesystems Nebraska Union - Regency Suite

      Nebraska Union - Regency Suite

      University of Nebraska - Lincoln

      <b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
      • 37
        The Lustre Filesystem for Petabyte Storage at the Florida HPC Center
        Design, performance, scalability, operational experience, monitoring, different modes of access and expansion plans for the Lustre filesystems, deployed for high performance computing at the University of Florida, are described. Currently we are running storage systems of 1.7 petabytes for the CMS Tier2 center and 2.0 petabytes for the university-wide HPC center.
        Speaker: Dr Dimitri Bourilkov (University of Florida (US))
        Slides
      • 38
        EOS across 1000 km
        In this contribution we report our experience in operating EOS, the CERN-IT high-performance disk-only solution, in multiple Computer Centres. EOS is one of the first production services exploiting the CERN's new facility located in Budapest, using his stochastic geo-location of data replicas. Currently EOS holds more than 100PB of raw disk space for the four big experiments (ALICE, ATLAS, CMS, LHCb) and for our general purpose instance, of which 40PB are installed 1000 km away from Geneva.
        Speaker: Luca Mascetti (CERN)
        Slides
    • 15:30
      Coffee Break Nebraska Union - Regency Suite

      Nebraska Union - Regency Suite

      University of Nebraska - Lincoln

      <b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
    • Storage and Filesystems Nebraska Union - Regency Suite

      Nebraska Union - Regency Suite

      University of Nebraska - Lincoln

      <b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
      • 39
        Using XRootD to Minimize Hadoop Replication
        We have developed an XRootD extension to Hadoop at UCSD that allows a site to significantly free local storage space by taking advantage of the file redundancy already provided by the XRootD Federation. Rather than failing when a corrupt portion of a file is accessed, the hdfs-xrootd-fallback system retrieves the segment from another site using XRootD, thus serving the original file to the end user seamlessly. These XRootD-fetched blocks are then cached locally, so subsequent accesses to the same segment do not require wide area network access. A second process is responsible for comparing the fetched blocks with corrupt blocks in Hadoop, and injects the cached blocks back into the cluster. This on-demand healing allows a site admin to relax the file replication number, commonly required to ensure availability. The system has been put into production at the UCSDT2 since March of 2014, and we finished implementing the healing portion in September. The added resiliency of the hdfs-xrootd-fallback system has allowed us to free 236 TB in our local storage facility.
        Speaker: Jeffrey Dost (UCSD)
        Slides
      • 40
        Cernbox + EOS: Cloud Storage for Science
        Cernbox is a cloud synchronization service for end-users: it allows to sync and share files on all major platforms (Linux, Windows, MacOSX, Android, iOS). The very successful beta phase of the service demonstrated high demand in the community for such easily accessible cloud storage solution. Integration of Cernbox service with the EOS storage backend is the next step towards providing sync and share capabilities for scientific and engineering use-cases. In this report we will present lessons learnt from the beta phase of the Cernbox service, key technical aspects of Cernbox/EOS integration and new, emerging usage possibilities. The latter include the ongoing integration of sync and share capabilities with the LHC data analysis tools and transfer services.
        Speaker: Luca Mascetti (CERN)
        Slides
    • Grids, Clouds, Virtualisation Nebraska Union - Regency Suite

      Nebraska Union - Regency Suite

      University of Nebraska - Lincoln

      <b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
      • 41
        Addressing the VM IO bottleneck
        This is summary of our efforts to address the issue of providing sufficient IO capacity to VMs running in our OpenStack cloud.
        Speaker: Dr Arne Wiebalck (CERN)
        Slides
    • Conference Dinner Sheldon Museum of Art

      Sheldon Museum of Art

    • 08:30
      Morning Coffee Regency Suite

      Regency Suite

      University of Nebraska - Lincoln

    • Storage and Filesystems Nebraska Union - Regency Suite

      Nebraska Union - Regency Suite

      University of Nebraska - Lincoln

      <b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
      • 42
        OpenZFS on Linux
        OpenZFS is a storage platform that encompasses the functionality of a traditional filesystem and volume manager. It's highly scalable, provides robust data protection, supports advanced features like snapshots and clones, and is easy to administer. These features make it an appealing choice for HPC sites like LLNL which uses it for all production Lustre filesystems. This contribution will discuss the state of OpenZFS on Linux including its goals and challenges. It will review the core features which make OpenZFS an excellent choice for managing large amounts of storage. Several new features will be discussed along our current plans for future improvements. I'll report on LLNL's use of OpenZFS on Linux over the last year to manage 100PB of production storage, including our experiences regarding stability, performance, and administration.
        Speaker: Brian Behlendorf (LLNL)
        Slides
      • 43
        SSD benchmarking at CERN
        Flash storage is slowly becoming more and more prevalent in the High Energy Physics community. When deploying Solid State Drives (SSDs) it's important to understand their capabilities and limitations, allowing to choose the best adapted product for the use case at hand. Benchmarking results from synthetic and real-world workloads on a wide array of Solid State Drives will be presented. The new NVM Express SSD interface specification will be detailed with a look towards the future of enterprise flash storage. The presentation will end by touching on endurance concerns often associated with the use of flash storage.
        Speaker: Liviu Valsan (CERN)
        Slides
      • 44
        Ceph Based Storage Systems for RACF
        Ceph based storage solutions are becoming increasingly popular within the HEP/NP community over the last few years. With the current status of Ceph project, both object storage and block storage layers are production ready on a large scale, and the Ceph file system storage layer (CephFS) is rapidly getting to that state as well. This contribution contains a thorough review of various functionality, performance and stability tests performed with all three (object storage, block storage and file system) levels of Ceph by using the RACF computing resources in 2012-2014 on various hardware platforms (including HP Moonshot) and with different networking solutions (10/40 GbE and IPoIB/4X FDR Infiniband based). We also report the status of commissioning a large scale (1 PB of usable capacity, 4.0k HDDs behind the RAID arrays by design) Ceph based object storage system provided with AMZ/S3 complaint RadosGW interfaces which is currently being finalized within RACF, and the early performance results obtained with it.
        Speaker: Mr Alexandr Zaytsev (Brookhaven National Laboratory (US))
        Slides
    • 10:30
      Group Photo (and Coffee Break) Nebraska Union - Regency Suite

      Nebraska Union - Regency Suite

      University of Nebraska - Lincoln

      <b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
    • Storage and Filesystems Nebraska Union - Regency Suite

      Nebraska Union - Regency Suite

      University of Nebraska - Lincoln

      <b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
      • 45
        DPM performance tuning hints for HTTP/WebDAV and Xrootd
        In this contribution we give a set of hints for the performance tuning of the upcoming DPM releases, and we show what one can achieve by looking at different graphs taken from the DPM nightly performance tests. Our focus is on the HTTP/WebDAV and Xrootd protocols and the newer "dmlite" software framework, and some of these hints may give some benefit also to older, legacy protocol implementations. Our goal is to make sure that single-VO and multi-VO DPM sites can join HEP and non-HEP computing models and HTTP and Xrootd federations, while giving the needed level of performance and the best system administration experience.
        Speaker: Andrea Manzi (CERN)
        Slides
      • 46
        Experience in running relational databases on clustered storage
        CERN IT-DB group is migrating its storage platform, mainly NetApp NAS’s running on 7-mode but also SAN arrays, to a set of NetApp C-mode clusters. The largest one is made of 14 controllers and it will hold a range of critical databases from administration to accelerators control or experiment control databases. This talk shows our setup: network, monitoring, use of features like transparent movement of file systems, flash pools (SSD + HDD storage pools), snapshots, etc. It will also show how these features are used on our infrastructure to support backup & recovery solutions with different database solutions: Oracle (11g and 12c multi tenancy), MySQL or PostgreSQL. Performance benchmarks and experience collected while running services on this platform will be also shared. It will be also covered the use of the cluster to provide iSCSI (block device) access for OpenStack windows virtual machines.
        Speaker: Ruben Domingo Gaspar Aparicio (CERN)
        Slides
      • 47
        New High Availability Storage at PDSF
        The PDSF Cluster at NERSC has been providing a data-intensive computing resource for experimental high energy particle and nuclear physics experiments (currently Alice, ATLAS, STAR, ICECUBE, MAJORANA) since 1996. Storage is implemented as a GPFS cluster built out of a variety of commodity hardware (Dell, Raidinc, Supermicro storage and servers). Recently we increased its capacity by 500TB by adding two file systems using NetApp E5600 series storage directly SAS attached to a pair of servers in a high availability configuration. Data IO routes to the cluster through dual 10Gb Ethernet. A 1Gb private network is used for monitoring and management. We will describe the configuration, share observations from the deployment process and provide the initial performance results. One of the new file systems was used to replace the back-end of the Tier3 ATLAS Storage Element (Bestman in gateway mode). We will share our experiences related to that move.
        Speaker: Tony Quan (LBL)
        Slides
    • 12:00
      Lunch Nebraska Union - Regency Suite

      Nebraska Union - Regency Suite

      University of Nebraska - Lincoln

      <b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
    • Basic IT Services Nebraska Union - Regency Suite

      Nebraska Union - Regency Suite

      University of Nebraska - Lincoln

      <b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
      • 48
        Developing Nagios code to suspend checks during planned outages
        Lawrence Berkeley National Laboratory/NERSC Division Developing Nagios code to suspend checks during planned outages. Raymond E. Spence NERSC currently supports more than 13,000 computation nodes spread over six supercomputing or clustered systems. These systems access cumulatively more than 13.5PB of disk space via thousands of network interfaces. This environment enables scientists from anywhere on the planet to login, run code and thereby to conduct science at elite levels. Scientists depend on NERSC for 24x7 availability and NERSC personnel in turn depend on industrial-strength system administration tools for our support efforts. Since monitoring everything from our largest system to the last network uplink is a chief concern at NERSC we chose several years ago to employ Nagios for our monitoring solution. Nagios is a mature product with a great degree of flexibility. Although NERSC has found the free, open source Nagios version sufficient in many ways we had eventually tired of one specific hole in this tool’s arsenal. The hole NERSC found in Nagios’ configuration involves planned downtime. Any Nagios user eventually comes to know where to point and click to acknowledge alerts and twiddle other Nagios switches. However, when it comes to running large systems with multiple monitored services per node, point and click solutions do not scale. Like any supercomputing center NERSC has many planned downtimes of varying size throughout the year. Unfortunately we found no obvious path to configure Nagios to temporarily turn off checks on a to-be downed resource. NERSC then began writing code to communicate directly with Nagios to suspend these checks. Over the past year NERSC has produced scripts which configure Nagios to respectively obey a planned downtime, remove a planned downtime and to list scheduled downtimes. Further, each downtime can cover any number of services running on any number of nodes. We used our dedicated Physics cluster, PDSF, as our test bed and first production system for the scripts. Managing planned outages on PDSF aided debugging the code and how to avoid misuse of its various configuration options. Today NERSC system managers can use our Nagios downtime scripts to quickly and easily accommodate downtime for anything Nagios monitors. Our downtime tool has saved a mountain of both point and click tasks and avoided the risky last resort of manually disabling Nagios checks. NERSC wishes to present these Nagios downtime scripts and describe more fully how this code has aided our support efforts.
        Speaker: Ray Spence (u)
        Slides
      • 49
        Configuration Services at CERN: update
        A status of the Puppet-based Configuration Service at CERN will presented giving a general update and discussing our current plans for the next 6 months. The presentation will also highlight the work being done to secure the Puppet infrastructure making it appropriate for use by a large number of administratively distinct user-groups.
        Speaker: Ben Jones (CERN)
        Slides
      • 50
        CFEngine Application at AGLT2
        CFEngine is a highly flexible configuration management framework. It also has a very high learning curve which can sometimes make decisions about how to deploy and use it difficult. At AGLT2 we manage a variety of different systems with CFEngine. We also have an effective version-controlled workflow for developing, testing, and deploying changes to our configuration. The talk will demonstrate what a practical and useful CFEngine infrastructure might look like. It will also include examples of how we organize our policy and effectively use the CFEngine policy language.
        Speaker: Mr Ben Meekhof (University of Michigan)
        Slides
      • 51
        Puppet at USCMS-T1 and FermiLab - Year 2
        USCMS-T1's work to globally deploy Puppet as our configuration management tool is well into the "long tail" phase, and has changed in fairly significant ways since its inception. This talk will discuss what has worked, how the Puppet tool itself has changed over the project, and our first thoughts as to what we expect to be doing in the next year (hint: starting again is rather likely!).
        Speaker: Timothy Michael Skirvin (Fermi National Accelerator Lab. (US))
        Slides
      • 52
        DataBase on Demand : insight how to build your on DBaaS
        Inspired on different database as a service, DBaas, providers, the database group at CERN has developed a platform to allow CERN user community to run a database instance with database administrator privileges providing a full toolkit that allows the instance owner to perform backup/ point in time recoveries, monitoring specific database metrics, start/stop of the instance and uploading/downloading specific logging or configuration files. With about 150 instances Oracle (11g and 12c), MySQL and PostgreSQL the platform has been designed and proofed to be flexible to run different RDBMS vendors and to scale up. Initially running on virtual machines, OracleVM, the instances are represented as objects in the management database toolset, making it independent of its physical representation. Nowadays instances run on physical servers together with virtual machines. A high availability solution has been implemented using Oracle cluster ware. This talk explains how we have built this platform, different technologies involved, actual user interface, command execution based on a database queue, backups based on snapshots, and possible future evolution (Linux containers, storage replication, OpenStack, Puppet,…).
        Speaker: Ruben Domingo Gaspar Aparicio (CERN)
        Slides
    • 15:30
      Coffee Break Nebraska Union - Regency Suite

      Nebraska Union - Regency Suite

      University of Nebraska - Lincoln

      <b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
    • IT Facilities and Business Continuity Nebraska Union - Regency Suite

      Nebraska Union - Regency Suite

      University of Nebraska - Lincoln

      <b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
      • 53
        Joint procurement of IT equipment and services
        The presentation describes options for joint activities around procurement of equipment and services by public labs, possibly with funding by the European Commission. The presentation is intended to inform the community and check whether there is interest.
        Speaker: Wayne Salter (CERN)
        Slides
    • Basic IT Services Nebraska Union - Regency Suite

      Nebraska Union - Regency Suite

      University of Nebraska - Lincoln

      <b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
      • 54
        Configuration Management, Change Management, and Culture Management
        In 2010, the RACF at BNL began investigating Agile/DevOps practices and methodologies to be able to do more in less time or effort. We choose Puppet in 2010 and by Spring of 2011 we had converted about half our of configuration shell scripts into Puppet code on a handful of machines. Today we have scaled Puppet 3.x to support our entire facility and and host a common Puppet code base that is now shared and used upstream by the Physics and IT departments.
        Speaker: James Pryor (B)
        Slides
      • 55
        Ermis service for DNS Load Balancer configuration
        This presentation describes the implementation and use cases of the Ermis Service. Ermis is a RESTful service to manage the configuration of DNS load balancers. It enables direct creation and deletion of DNS delegated zones using a SOAP interface provided by the Network group thus simplifying the procedure needed for supporting new services. It is written in Python as a Django Application. This is quite generic and can be easily adapted to other types of Load Balancers. Ermis is being integrated with Openstack. It uses the Openstack Keystone API as a means of authentication and authorization as well as Kerberos and e-groups. The ultimate aim of the project is to provide Load Balancing as a Service (LBaaS) to the end users of the cloud.
        Speaker: Aris Angelogiannopoulos (Ministere des affaires etrangeres et europeennes (FR))
        Slides
    • 08:30
      Morning Coffee Regency Suite

      Regency Suite

      University of Nebraska - Lincoln

    • Grids, Clouds, Virtualisation Nebraska Union - Regency Suite

      Nebraska Union - Regency Suite

      University of Nebraska - Lincoln

      <b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
      • 56
        RAL Tier 1 Cloud Computing Developments
        Update on the RAL Tier 1 cloud deployment and cloud computing activities.
        Speaker: Ian Peter Collier (STFC - Rutherford Appleton Lab. (GB))
        Slides
      • 57
        The Adoption of Cloud Technologies within the LHC Experiments
        The adoption of cloud technologies by the LHC experiments is currently focused on IaaS, more specifically the ability to dynamically create virtual machines on demand. This talk provides an overview of how this alternative approach for resource provision fits into the existing workflows used by the experiments. It shows that in order to fully exploit this approach, solutions are required in the areas of image management, capacity management, monitoring, accounting, pilot job frameworks and supporting services. Each of those areas is covered in more detail to explain the various architectural choices and the rational for the decisions made. Finally, a general overview of the state of adoption within each experiment is given that describes the ongoing integration with their individual frameworks.
        Speaker: Laurence Field (CERN)
        Slides
      • 58
        Evolution of WLCG monitoring
        The WLCG monitoring system provides a solid and reliable solution that has supported LHC computing activities and WLCG operations during the first years of LHC data-taking. The current challenge consists of ensuring that the WLCG monitoring infrastructure copes with the constant increase of monitoring data volume and complexity (new data-transfer protocols, new dynamic types of resource providers - cloud computing). At the same time, simplification of the monitoring system is desirable in order to reduce maintenance and operational costs. The current evolution of the system aims to achieve these two goals: decrease the complexity of the system and ensure its scalability and performance with the steady increase of monitoring information. The presentation will describe the new WLCG monitoring platform including the new technology stack for large-scale data analytics.
        Speaker: Dr Edward Karavakis (CERN)
        Slides
      • 59
        CERN Cloud Report
        This is a report on the current status of CERN's OpenStack-based Cloud Infrastructure.
        Speaker: Dr Arne Wiebalck (CERN)
        Slides
    • 11:00
      Coffee Break Nebraska Union - Regency Suite

      Nebraska Union - Regency Suite

      University of Nebraska - Lincoln

      <b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
    • HEPiX Business Nebraska Union - Regency Suite

      Nebraska Union - Regency Suite

      University of Nebraska - Lincoln

      <b>GPS coordinates :</b> <a href="https://www.google.com/maps/place/Nebraska+Union/@40.8177945,-96.7004067,17z"> N 40.81 E -96.70 </a>
      slides