HEPiX Spring 2019 Workshop

America/Los_Angeles
E-B 212 (SDSC Auditorium)

E-B 212

SDSC Auditorium

10100 Hopkins Drive La Jolla, CA 92093-0505
Helge Meinhard (CERN), Tony Wong (Brookhaven National Laboratory)
Description

HEPiX Spring 2019 at San Diego Supercomputing Center (SDSC) / University of California in San Diego

The HEPiX forum brings together worldwide Information Technology staff, including system administrators, system engineers, and managers from High Energy Physics and Nuclear Physics laboratories and institutes, to foster a learning and sharing experience between sites facing scientific computing and data challenges.

Participating sites include BNL, CERN, DESY, FNAL, IHEP, IN2P3, INFN, IRFU, JLAB, KEK, LBNL, NDGF, NIKHEF, PIC, RAL, SLAC, TRIUMF, many other research labs and numerous universities from all over the world.

The workshop was held at the San Diego Supercomputing Center at the University of California in San Diego, and was proudly sponsored by

Platinum

Gold

  Teradactyl

Silver

Bronze

 

Entons

Surveys
Endpoint user device and app provisioning models
Videoconference Rooms
HEPiX_Workshop
Name
HEPiX_Workshop
Description
Vidyo virtual room for the HEPiX workshops
Extension
10637013
Owner
Dennis Van Dok
Auto-join URL
Useful links
Phone numbers
    • 08:00 09:00
      Logistics & announcements: Registration E-B 212

      E-B 212

      SDSC Auditorium

      10100 Hopkins Drive La Jolla, CA 92093-0505
    • 09:00 09:20
      Welcome to SDSC and UCSD E-B 212

      E-B 212

      SDSC Auditorium

      10100 Hopkins Drive La Jolla, CA 92093-0505
    • 09:20 09:30
      Logistics & announcements E-B 212

      E-B 212

      SDSC Auditorium

      10100 Hopkins Drive La Jolla, CA 92093-0505
    • 09:30 10:30
      Site Reports E-B 212

      E-B 212

      SDSC Auditorium

      10100 Hopkins Drive La Jolla, CA 92093-0505
      • 09:30
        PDSF Site Report 15m

        PDSF, the Parallel Distributed Systems Facility has been in continuous operation since 1996, serving high energy physics research. The cluster is a tier-1 site for Star, a tier-2 site for Alice and a tier-3 site for Atlas.

        We'll give a status report of the PDSF cluster and the migration into Cori, the primary computing resource at NERSC. We'll go into how we tried to ease the process by providing a stepping stone environment as intermediary between a commodity cluster and a supercomputer. Updates on NERSC systems will be given as well.

        Speaker: Georg Rath (Lawrence Berkeley National Laboratory)
      • 09:45
        BNL Site Report 15m

        News and updates from BNL activities since the Barcelona meeting

        Speaker: Ofer Rind (Brookhaven National Laboratory)
      • 10:00
        Update of Canadian T1 / T2 15m

        I will present recent developments of the Canadian T1 and T2.

        Speaker: Rolf Seuster (University of Victoria (CA))
      • 10:15
        AGLT2 Site Report 15m

        We will present an update on AGLT2, focusing on the changes since the Fall 2018 report.
        The primary topics to cover include the update on VMware, update of dCache, status of new purchased hardware, encountered problems and solutions on improving the CPU utilization of our HTCondor system.

        Speaker: Dr Wenjing Wu (University of Michigan)
    • 10:30 11:00
      Coffee Break 30m
    • 11:00 12:30
      Site Reports E-B 212

      E-B 212

      SDSC Auditorium

      10100 Hopkins Drive La Jolla, CA 92093-0505
      • 11:00
        University of Wisconsin-Madison CMS T2 site report 15m

        As a major WLCG/OSG T2 site, the University of Wisconsin-Madison CMS T2 has consistently been delivering highly reliable and productive services towards large scale CMS MC production/processing, data storage, and physics analysis for last 13 years. The site utilizes high throughput computing (HTCondor), highly available storage system (Hadoop), scalable distributed software systems (CVMFS), and provides efficient data access using xrootd/AAA. The site fully supports IPv6 networking, and is a member of the LHCONE community with 100Gb WAN connectivity. An update on the activities and developments at the T2 facility over the past 1.5 years (since the KEK meeting) will be presented.

        Speaker: Ajit Kumar Mohapatra (University of Wisconsin Madison (US))
      • 11:15
        University of Nebraska CMS Tier2 Site Report 15m

        Updates on the activities at T2_US_Nebraska over the past year. Topics will cover the site configuration and tools we use, troubles we face in daily operation, and contemplation of what the future might hold for sites like ours.

        Speaker: Garhan Attebury (University of Nebraska Lincoln (US))
      • 11:30
        KEK Site Report 15m

        We would like to report an update of the computing research center at KEK including the Grid system from the last HEPiX Fall 2018 for the data taking period of SuperKEKB and J-PARC experiments in 2019. The network connectivity of KEK site has been improved by the replacement of network equipment and security devices in September 2018. The situation of the international network for Japan will also be introduced. In addition to the status report, we will present on the preparation for procurement of the next system.

        Speaker: Tomoaki Nakamura (High Energy Accelerator Research Organization (JP))
      • 11:45
        Tokyo Tier-2 Site Report 15m

        The Tokyo Tier-2 center, which is located in the International Center for Elementary Particle Physics (ICEPP) at the University of Tokyo, is providing computing resources for the ATLAS experiment in the WLCG. Almost all hardware devices of the center are supplied by a lease, and are upgraded in every three years. This hardware upgrade was performed in December 2018. In this presentation, experiences of the system upgrade will be reported. The configuration of the new system will also be shown.

        Speaker: Tomoe Kishimoto (University of Tokyo (JP))
      • 12:00
        UW CENPA site report 15m

        At CENPA at the University of Washington we have a heterogeneous rocks 7 cluster of
        about 135 nodes containing 1250 cores. We will present the current status and issues.

        Speaker: Dr Duncan Prindle (University of Washington)
      • 12:15
        JLab Site Report 15m

        Updates from JLab since the Autumn 2018 meeting at PIC in Barcelona.

        Speaker: Sandra Philpott
    • 12:30 14:00
      Lunch Break 1h 30m
    • 14:00 15:15
      End-User IT Services & Operating Systems E-B 212

      E-B 212

      SDSC Auditorium

      10100 Hopkins Drive La Jolla, CA 92093-0505
      • 14:00
        Text Classification via Supervised Machine Learning for an Issue Tracking System 25m

        Comet is SDSC’s newest supercomputer. The result of a $27M National Science Foundation (NSF) award, Comet deliverers over 2.7 petaFLOPS of computing power to scientists, engineers, and researchers all around the world. In fact, within its first 18 months of operation, Comet served over 10,000 unique users across a range of scientific disciplines, becoming one of the most widely used supercomputers in NSF’s Extreme Science and Engineering Discover Environment (XSEDE) program ever.

        The High-Performance Computing (HPC) User Services Group at SDSC helps manage user support for Comet. This includes, but is not limited to, managing user accounts, answering general user inquires, debugging technical problems reported by users, and making best practice recommendations on how users can achieve high-performance when running their scientific workloads on Comet. These interactions between Comet’s user community and the User Service Group are largely managed through email exchanges tracked by XSEDE’s internal issue tracking system. However, while Comet is expected to maintain a 24x7x365 uptime, user support is generally only provided during normal business hours. With such a large user community spread across nearly every timezone, the result is a number of user support tickets submitted during non-business hours waiting between 12 hours to several days for responses from the User Services Group.

        The aim of this research project is to use supervised machine learning techniques to perform text classification on Comet’s user support tickets. If an efficient classification scheme can be developed, the User Services Group may eventually be able to provide automated email responses to some of the more common user issues reported during non-business hours.

        Speaker: Martin Kandes (Univ. of California San Diego (US))
      • 14:25
        Endpoint user device and app provisioning models 25m

        Over the last years, there has been a number of trends related to how devices are provisioned and managed within organizations, such as BYOD - "Bring Your Own Device" or COPE: "Company Owned, Personally Enabled". In response, a new category of products called "Enterprise Mobility Management Suites", which includes MDM - "Mobile Device Management" and MAM - "Mobile Application Management" emerged on the market. Vendors like VMWare, MobileIron, Microsoft and Citrix now all provide more or less comprehensive systems in this category. But how do these commercial systems correspond to the needs of the Scientific Community?

        This talk will summarize current status of device management practices and the strategy for provisioning of devices and applications at CERN. It will also attempt to initiate wider discussion within the community.

        Speaker: Michal Kwiatek (CERN)
      • 14:50
        How to make your Cluster look like a Supercomputer (for Fun and Profit) 25m

        During the last two years, the computational systems group at NERSC, in partnership with Cray, has been developing SMWFlow, a tool that makes managing system state as simple as switching branches in git. This solution is the cornerstone of collaborative systems management at NERSC and enables code-review, automated testing and reproducibility.
        Besides supercomputers, NERSC hosts Mendel, a commodity meta-system containing multiple clusters, among them PDSF, used by the HEP community, and Denovo, used by the Joint Genome Institute, which uses a custom management stack built on top of xCAT and cfengine.
        To merge efforts, provide a consistent user experience, and to leverage the work done on SMWFlow, we will talk about how we adapted the Cray imaging and provisioning system to work on a system on an architecture like Mendel and therefore reap the benefits of a modern systems management approach.

        Speaker: Georg Rath (Lawrence Berkeley National Laboratory)
    • 15:15 15:45
      Coffee Break 30m
    • 15:45 17:00
      Storage & Filesystems E-B 212

      E-B 212

      SDSC Auditorium

      10100 Hopkins Drive La Jolla, CA 92093-0505
      • 15:45
        OpenAFS Release Team report 25m

        A report from the OpenAFS Release Team on recent OpenAFS releases and development branch updates. Topics include acknowledgement of contributors, descriptions of issues fixed, updates for new versions of Linux and Solaris, changes currently under review, and an update on the new RXGK security class for improved security.

        Speaker: Mr Michael Meffie (Sine Nomine)
      • 16:10
        IntegratingHadoop Distributed File System to Logistical Storage 25m

        Logistical Storage (LStore) provides a flexible logistical networking storage framework for distributed and scalable access to data in both an HPC and WAN environment. LStore uses commodity hard drives to provide unlimited storage with user controllable fault tolerance and reliability. In this talk, we will briefly discuss LStore's features and discuss the newly developed native LStore plugin for the Apache Hadoop ecosystem. The Hadoop Distributed File System (HDFS) will directly access LStore using this plugin allowing users to create Hadoop clusters on the fly in an HPC environment. The primary benefit of the plugin is that it avoids the need for data redundancy across a traditional Hadoop and HPC cluster. Moreover, the on the fly Hadoop clusters created in the HPC environment can be scaled as needed and tune the hardware requirements to the analysis - large memory needs, GPU, etc.

        We will show several empirical results using the plugin in both a traditional HPC environment and utilizing a high-latency WAN connection. The proposed plugin is compared with two current LStore interfaces: LStore command line interface and LStore FUSE mounted client interface.

        Speaker: Dr Shunxing Bao (Vanderbilt University)
      • 16:35
        Developments in disk and tape storage at the RAL Tier 1 25m

        RAL's Ceph-based Echo storage system is now the primary disk storage system running at the Tier 1, replacing a legacy CASTOR system that will be retained for tape. This talk will give an update on Echo's recent development, in particular the adaptations needed to support the ALICE experiment and the challenges of scaling an erasure-coded Ceph cluster past the 30PB mark. These include the smoothing of data distribution, managing disk errors, and dealing with a very full cluster.

        In addition, I will discuss the completed project to remodel RAL's CASTOR service from a combined disk and tape endpoint to a low-maintenance system only providing access to tape.

        Speaker: Rob Appleyard (STFC)
    • 17:00 17:15
      Site Reports E-B 212

      E-B 212

      SDSC Auditorium

      10100 Hopkins Drive La Jolla, CA 92093-0505
      • 17:00
        BEIJING Site Report 15m

        News and updates from IHEP since the last HEPiX Workshop. In this talk we would like to present the status of IHEP site including computing farm, HPC, IHEPCloud, Grid, data storage ,network and so on.

        Speaker: Dr Qiulan Huang (Institute of High Energy Physics, Chinese Academy of Science)
    • 18:00 20:00
      Welcome reception 2h 15th Floor (The Village)

      15th Floor

      The Village

      Scholars Drive N La Jolla, CA 92093-0505
    • 09:00 10:15
      Site Reports E-B 212

      E-B 212

      SDSC Auditorium

      10100 Hopkins Drive La Jolla, CA 92093-0505
    • 10:15 11:00
      Coffee Break 45m
    • 11:00 11:45
      Site Reports E-B 212

      E-B 212

      SDSC Auditorium

      10100 Hopkins Drive La Jolla, CA 92093-0505
      • 11:00
        Prague Site Report 15m

        We will give an overview of the site including our recent network redesign. We will dedicate a part of the talk to disk servers: report on the newest additions as well as upgraded old hardware. We will also share experience with our distributed HT-Condor batch system.

        Speakers: Jiri Chudoba (Acad. of Sciences of the Czech Rep. (CZ)), Martin Adam (Acad. of Sciences of the Czech Rep. (CZ))
      • 11:15
        GSI Site Report 15m

        Ongoing developments at GSI/FAIR: diggers, Lustres, procurements, relocations, operating systems

        Speaker: Thomas Roth (GSI)
      • 11:30
        Diamond Light Source Site Report 15m

        Diamond Light Source is an X-ray synchrotron light source co-located with STFC RAL in the UK. This is the first site report from Diamond at HEPiX since 2015. The talk will discuss recent changes, current status and future plans as well as the odd disaster story thrown in for good measure.

        Diamond has a new data centre, new storage and new compute as well as new staff and a few forays into various cloud providers.

        Speaker: James Thorne (Diamond Light Source)
    • 11:45 12:35
      Grid, Cloud and Virtualization E-B 212

      E-B 212

      SDSC Auditorium

      10100 Hopkins Drive La Jolla, CA 92093-0505
      • 11:45
        Addressing the Challenges of Executing Massive Computational Clusters in the Cloud 25m

        This talk will discuss how we worked with Dr. Amy Apon, Brandon Posey, AWS and the Clemson DICE lab team dynamically provisioned a large scale computational cluster of more than one million cores utilizing Amazon Web Services (AWS). We discuss the trade-offs, challenges, and solutions associated with creating such a large scale cluster with commercial cloud resources. We utilize our large scale cluster to study a parameter sweep workflow composed of message-passing parallel topic modeling jobs on multiple datasets.

        At peak, we achieve a simultaneous core count of 1,119,196 vCPUs across nearly 50,000 instances, and are able to execute almost half a million jobs within two hours utilizing AWS Spot Instances in a single AWS region.

        Additionally we will discuss a follow on project that the DICE Lab is currently working on in the Google Cloud Platform (GCP) that will enable a Computer Vision analytics system to concurrently processes hundreds of thousands of hours of highway traffic video providing statistics on congestions, vehicle trajectories and neural net pre-annotation. We will discuss how this project will differ from the previous one and how additional boundaries are being pushed.

        Relevant Papers:
        https://ieeexplore.ieee.org/abstract/document/8411029

        https://tigerprints.clemson.edu/computing_pubs/38/

        Speaker: Boyd Wilson (Omnibond)
      • 12:10
        CloudScheduler version2 25m

        I present the recent developments for our cloudschdeduler, which we use to run HEP workloads on various clouds in North America and Europe. We are working on a complete re-write utilizing modern software technologies and practices.

        Speaker: Rolf Seuster (University of Victoria (CA))
    • 12:35 14:00
      Lunch Break 1h 25m
    • 14:00 15:40
      Networking & Security E-B 212

      E-B 212

      SDSC Auditorium

      10100 Hopkins Drive La Jolla, CA 92093-0505
      • 14:00
        CERN DNS and DHCP service improvement plans 25m

        The configuration of the CERN IT central DNS servers, based on ISC BIND, is generated automatically from scratch every 10 minutes using a software developed at CERN several years ago. This in-house set of Perl scripts has evolved and is reaching its limits in terms of maintainability and architecture. CERN is in the process of reimplementing the software with a modern language and is taking the opportunity to redefine the DNS service architecture by introducing a redundant solution for the master DNS. Meanwhile, Anycast is being evaluated in order to increase the DNS service robustness and scalability. Finally, CERN is considering the possibility of moving from a static to a dynamic zone for the cern.ch domain to allow immediate commissioning while controlling the update process.
        Concerning the DHCP services, ISC DHCP has been the software of choice to support dynamic host configuration for almost 20 years. However system provisioning has massively scaled in the last years and DHCP software shortcomings have lead ISC to develop Kea. CERN intends to modernize the service replacing ISC DHCP with Kea, which will allow the implementation of a highly available and geographically dispersed DHCP service, as well as a fast provisioning so that changes in the network database are immediately propagated to the DHCP servers.

        Speaker: Quentin Barrand (CERN)
      • 14:25
        Computer Security Update 25m

        This presentation provides an update on the global security landscape since the last HEPiX meeting.It describes the main vectors of risks to and compromises in the academic community including lessons learnt, presents interesting recent attacks while providing recommendations on how to best protect ourselves. It also covers security risks management in general, as well as the security aspects of the current hot topics in computing and around computer security.

        This talk is based on contributions and input from the CERN Computer Security Team.

        Speaker: Stefan Lueders (CERN)
      • 14:50
        BNL activities on federated access and Single Sign-On 25m

        Various High energy and nuclear physics experiments already benefit from using the different components of Federated architecture to access storage and infrastructure services. BNL moved to Identity management (Redhat IPA) in late 2018 which will serve as the foundation to move to Federated authentication and authorization. IPA provides central authentication via Kerberos or LDAP, simplifies administration,
        has a rich CLI and a web based user interface. This presentation describes how federated authn/authz will be enabled in the near future at the level of individual applications like Globus online, Invenio, BNLbox, Indico, Web services and Jupyter.

        Speaker: Tejas Rao (Brookhaven National Laboratory)
      • 15:15
        The difference in network equipment in the 25/100/400G era and how to test/break them 25m

        The network market has changed a lot compared with a decade ago. Every hardware vendor sells their own switches and routers. Most of the switches and routers are based on the same merchant silicon that is available on the market.
        Therefore the amount of real choices is limited because what is inside is the same for most of them.
        This talk will tell about the differences that are still there and what are the risks for choosing certain solutions.
        What will vendors really allow you to do in their "Open Networking" strategy?
        Why is knowing how many packets per second more important than how much bandwidth a network device can process?
        How do you test this and what type of effects can you expect when reaching the limits of the equipment?
        Why aren’t commercial network testers the best way of testing network equipment?
        Is building your own network test machine expensive?

        Speaker: Tristan Suerink (Nikhef National institute for subatomic physics (NL))
    • 15:40 16:10
      Coffee Break 30m
    • 16:10 17:25
      Networking & Security E-B 212

      E-B 212

      SDSC Auditorium

      10100 Hopkins Drive La Jolla, CA 92093-0505
      • 16:10
        WLCG/OSG Network Activities, Status and Plans 25m

        WLCG relies on the network as a critical part of its infrastructure and therefore needs to guarantee effective network usage and prompt detection and resolution of any network issues, including connection failures, congestion and traffic routing. The OSG Networking Area is a partner of the WLCG effort and is focused on being the primary source of networking information for its partners and constituents. We will report on the changes and updates that have occurred since the last HEPiX meeting.

        The primary areas to cover include the status of and plans for the WLCG/OSG perfSONAR infrastructure, the WLCG Throughput Working Group and the activities in the IRIS-HEP and SAND projects.

        Speaker: Shawn Mc Kee (University of Michigan (US))
      • 16:35
        IPv6 & WLCG - an update from the HEPiX IPv6 Working Group 25m

        The transition of WLCG storage services to dual-stack IPv4/IPv6 is progressing well, aimed at enabling the use of IPv6-only CPU resources as agreed by the WLCG Management Board and presented by us at previous HEPiX meetings.

        The working group, driven by the requirements of the LHC VOs to be able to use IPv6-only opportunistic resources, continues to encourage wider deployment of dual-stack services and has been monitoring the transition. During recent months we have also started to investigate in more detail the reasons for various edge cases where the fraction of data transferred over IPv6 is lower than expected.

        This talk will present the current status of the transition to IPv6 together with some of the common reasons for sites that have not yet been able to move to dual-stack operations. Some issues related to unexpected monitoring results for IPv6 versus IPv4 will also be discussed.

        Speaker: Andrea Sciabà (CERN)
      • 17:00
        Network Functions Virtualisation Working Group Update 25m

        High Energy Physics (HEP) experiments have greatly benefited from a strong relationship with Research and Education (R&E) network providers and thanks to the projects such as LHCOPN/LHCONE and REN contributions, have enjoyed significant capacities and high performance networks for some time. RENs have been able to continually expand their capacities to over-provision the networks relative to the experiments needs and were thus able to cope with the recent rapid growth of the traffic between sites, both in terms of achievable peak transfer rates as well as in total amount of data transferred. For some HEP experiments this has lead to designs that favour remote data access where network is considered an appliance with almost infinite capacity. There are reasons to believe that the network situation will change due to both technological and non-technological reasons starting already in the next few years. Various non-technological factors that are in play are for example anticipated growth of the non-HEP network usage with other large data volume sciences coming online; introduction of the cloud and commercial networking and their respective impact on usage policies and securities as well as technological limitations of the optical interfaces and switching equipment.

        As the scale and complexity of the current HEP network grows rapidly, new technologies and platforms are being introduced that greatly extend the capabilities of today’s networks. With many of these technologies becoming available, it’s important to understand how we can design, test and develop systems that could enter existing production workflows while at the same time changing something as fundamental as the network that all sites and experiments rely upon. In this talk we’ll give an update on the working group's recent activities, updates from sites and R&E network providers as well as plans for the near-term future.

        Speaker: Shawn Mc Kee (University of Michigan (US))
    • 17:30 19:00
      Board Meeting
    • 09:00 10:40
      Storage & Filesystems E-B 212

      E-B 212

      SDSC Auditorium

      10100 Hopkins Drive La Jolla, CA 92093-0505
      • 09:00
        Using the dynafed data federation as site storage element 25m

        We describe our experience and use of the Dynafed data federator with cloud and traditional Grid computing resources as an substitute for a traditional Grid SE.
        This is an update of the report given at the Fall HEPiX meeting of 2017 where we introduced our use case for such federation and described our initial experience with it.
        We used Dynafed in production for Belle-II since late 2017 and also in testing mode for Atlas. We will report on changes we made since then to our setup and also report on changes in Dynafed itself that makes it more suitable as a site SE. We will also report on a new monitoring system we developed for such data federation and report also on a way to use such data federation by anyone who uses distributed compute and storage but with the need to read/write from a local file system.

        Speaker: Marcus Ebert (University of Victoria)
      • 09:25
        OSiRIS: Open Storage Research Infrastructure 25m

        OSiRIS is a pilot project funded by the NSF to evaluate a
        software-defined storage infrastructure for our primary Michigan
        research universities and beyond. In the HEP world OSiRIS is involved
        with ATLAS as a provider of Event Service storage via the S3 protocol
        as well as experimenting with dCache backend storage for AGLT2. We
        are also in the very early stages of working with IceCube and the
        nationwide Open Storage Network. Our talk will cover current status
        on these projects and the latest details of how we use Ceph, HAproxy,
        NFSv4, LDAP, COmanage, Puppet and other tools to provision, manage,
        and monitor storage services to federated users.

        Speaker: Benjeman Jay Meekhof (University of Michigan (US))
      • 09:50
        Virtualization for Online Storage Clusters 25m

        The computing center GridKa is serving the ALICE, ATLAS, CMS and
        LHCb experiments as Tier-1 center with compute and storage resources.
        It is operated by the Steinbuch Centre for Computing at Karlsruhe Institute
        of Technology in Germany. In its current stage of expansion GridKa
        offers the HEP experiments a capacity of 35 Petabytes of online storage.
        The storage system is based on Spectrum Scale as software-defined-storage
        layer. Its storage servers are inter-connected via two redundant
        infiniband fabrics and have ethernet uplinks to the GridKa backbone network.
        In this presentation we discuss the use of virtualization technologies
        in the context of the described storage system, including hardware
        virtualization of the infiniband and ethernet interfaces.

        Speaker: Jan Erik Sundermann (Karlsruhe Institute of Technology (KIT))
      • 10:15
        DPM DOME 25m

        DPM (Disk Pool Manager) is mutli-protocol distributed storage system that can be easily used within grid environment and it is still popular for medium size sites. Currently DPM can be configured to run in legacy or DOME mode, but official support for the legacy flavour ends this summer and sites using DPM storage should think about their upgrade strategy or coordinate with WLCG DPM Upgrade task force.
        We are going to present our almost a year long experience with DPM running in DOME mode on our production storage hosting several petabytes of data for different VOs. Our experience can help others to avoid common problems and also choose right protocols to get best performance from DPM storage. DOME provides support for SRM-less site configuration, but SRM can be still used if necessary and we'll show advantages and/or disadvantages that comes from such configuration.
        New DPM features are developed only for DOME mode. We would like to summarize which improvements are available in DOME and these features will never be available for legacy flavour. Running DOME brings greatly improved support for non-GridFTP protocols like full support for transfer checksums, storage resource reporting and most importantly third-party-copy (TPC). We are going to describe DPM TPC support including various credentials delegation mechanisms for XRootD and WebDAV protocols and interoperability with other storage implementations.

        Speaker: Petr Vokac (Czech Technical University (CZ))
    • 10:40 11:10
      Coffee Break 30m
    • 11:10 12:00
      Storage & Filesystems E-B 212

      E-B 212

      SDSC Auditorium

      10100 Hopkins Drive La Jolla, CA 92093-0505
      • 11:10
        Storage services at CERN 25m

        The Storage group of the CERN IT department is responsible for the development and the operation of petabyte-scale services needed to accommodate the diverse requirements for storing physics data generated by LHC and non-LHC experiments as well as supporting users of the laboratory in their day-by-day activities.

        This contribution presents the current operational status of the main storage services at CERN, summarizes our experience in operating largely distributed systems and highlights the ongoing efforts for the evolution of the storage infrastructure.

        It reports about EOS, the high-performance distributed filesystem developed at CERN designed to store all the physics data and to operate at the high rates demanded by experiments data taking. EOS is also used as the storage backend for CERNBox, the cloud storage synchronization and sharing service for users’ personal files. CERNBox provides uniform access to storage from all modern devices and represents the data hub for integration with various applications ranging from office suites (Microsoft Office 365, OnlyOffice, Draw.io) to specialized tools for data analysis (SWAN).

        Besides storage for physics data and personal files, the Storage group runs multiple large Ceph clusters to provide the storage backbones for the OpenStack infrastructure and the HPC facility, and to offer an S3 service and a CephFS/Manila shares for other internal IT services. Also, the Storage group operates the release managers, replica servers and caches of CVMFS (a fundamental WLCG service used for software distribution) in collaboration with the SoFTware Development for Experiments (CERN EP-SFT) department.

        Speaker: Enrico Bocchi (CERN)
      • 11:35
        Storage management in a large scale at BNL 25m

        Brookhaven National Laboratory stores and processes large amounts of data from the following: PHENIX,STAR,ATLAS,Belle II, Simons, as well as smaller local projects. This data is stored long term in tape libraries but one working data is stored in disk arrays. Hardware raid devices from companies such as Hitachi Ventara are very convenient and require minimal administrative intervention. However, they are very expensive relative the alternatives. BNL is making a move toward JBOD (Just a Bunch of Disk) arrays with Linux based software raid. The performance is comparable and sometimes better than the hardware cousins but the cost is less than half. However, the construction and administration is more complex. This requires more hours of skilled manpower from staff to install and maintain. I am developing software at BNL to automate these processes to the level of hardware raid in order to reduce this burden while allowing cost savings.

        Speaker: Robert Hancock (Brookhaven National Laboratory)
    • 12:00 12:25
      Miscellaneous E-B 212

      E-B 212

      SDSC Auditorium

      10100 Hopkins Drive La Jolla, CA 92093-0505
    • 12:25 14:00
      Lunch Break 1h 35m
    • 14:00 15:40
      Computing & Batch Systems E-B 212

      E-B 212

      SDSC Auditorium

      10100 Hopkins Drive La Jolla, CA 92093-0505
      • 14:00
        Benchmarking Worrking Group - Status Report 25m

        The Benchmarking Working Group has been very active in the last months. The group observed that SPEC CPU 2017 is not very different from SPEC CPU 2006. On the worker node available the two benchmark are higly correlated. Analysis with Trident shows that the hardware counters usage is rather different from the HEP applications. So the group started to investigate the usage of real applications running inside docker. The result are very promising. The current efforts are in the directions of having a suite very simple that can be distributed and runs everywhere without any knowledge of the applications, so that is can be given to a WLCG data center, a supercomputer center or a vendor for procurement procedure.

        Speaker: Michele Michelotto (Università e INFN, Padova (IT))
      • 14:25
        How Fair is my Fair-Sharing? Exposing Some Hidden Behavior Through Workload Analysis 25m

        Monitoring and analyzing how a workload is processed by a job and resource management system is at the core of the operation of data centers. It allows operators to verify that the operational objectives are satisfied, detect any unexpected and unwanted behavior, and react accordingly to such events. However, the scale and complexity of large workloads composed of millions of jobs executed each month on several thousands of cores, often limit the depth of such analysis. This may lead to overlook some phenomena that, while they are not harmful at the global scale of the system, can be detrimental to a specific class of users.

        In this talk, we illustrate such a situation by analyzing the large High Throughput Computing (HTC) workload trace coming from the Computing Center of the National Institute of Nuclear Physics and Particle Physics~(CC-IN2P3) which is one of the largest academic computing centers in France. The batch scheduler implements the classical Fair-Share algorithm which ensures that all user groups are fairly provided with an amount of computing resources commensurate to their expressed needs for the year. However, the deeper we analyze this workload's scheduling, especially the waiting times of jobs, the clearer we see a certain degree of unfairness between user groups.We identify some of the root causes of this unfairness and propose a drastic reconfiguration of the quotas and scheduling queues managed by the job and resource management system. This modification aims at being more suited to the characteristics of the workload and at improving the balance across user groups in terms of waiting. We evaluate the impact of this modification through detailed simulations. The obtained results show that it still guarantees the satisfaction of the main operational objectives while significantly improving the quality of service experienced by the formerly unfavored users.

        Speaker: Frederic Suter (CNRS / CC-IN2P3)
      • 14:50
        Evolution of interactive data analysis for HEP at CERN – SWAN, Kubernetes, Apache Spark and RDataFrame 25m

        This talk is focused on recent experiences and developments in providing data analytics platform SWAN based on Apache Spark for High Energy Physics at CERN.

        The Hadoop Service expands its user base for analysts who want to perform analysis with big data technologies - namely Apache Spark – with main users from accelerator operations and infrastructure monitoring. Hadoop Service integration with SWAN Service offers scalable interactive data analysis and visualizations using Jupyter notebooks, with computations being offloaded to compute clusters - on-premise YARN clusters and more recently to cloud-native Kubernetes clusters. The ROOT framework is most widely used tool for high-energy physics analysis. Its integration with SWAN allows physicists to perform web-based interactive analysis using standard tools and libraries, in the cloud.

        The first part of presentation will focus on integration of Spark on Kubernetes into SWAN service, which allows to offload computations to elastic, virtualized and container-based infrastructure in the private or public clouds, compared to complex to manage and operate on-premise Hadoop clusters.

        The second part will focus on evolutions in exploiting analytics infrastructure - namely new developments in ROOT framework – Distributed RDataFrame - which would allow interactive, parallel and distributed analysis on large physics datasets by transparently exploiting dynamically pluggable resources in SWAN, e.g. Hadoop or Kubernetes clusters.

        Speaker: Piotr Mrowczynski (CERN)
      • 15:15
        Jupyter at SDCC 25m

        ...Placeholder...

        Speaker: William Strecker-Kellogg (Brookhaven National Lab)
    • 15:40 16:10
      Coffee Break 30m
    • 16:10 17:25
      Computing & Batch Systems E-B 212

      E-B 212

      SDSC Auditorium

      10100 Hopkins Drive La Jolla, CA 92093-0505
      • 16:10
        Computing/Storage/Networking for next generation photon science experiments @DESY 25m

        We will briefly show the current onsite accelerator infrastructure and their resulting computing and storage usage and future requirements. The second section will discuss the plans and work done regarding the hardware infrastructure, the system level middleware (i.e. container, storage connection, networks) and the higher level middleware (under development) covering low latency data access and selection (including metadata queries) directly connection the DAQ at detector level to the data processing code (mostly developed by experimentators running on multiple nodes in parallel. The last section will shortly discuss initial result (technical and non-technical) and possible collaborations with other sites with similar challenges.

        Speaker: Mr Martin Gasthuber (DESY)
      • 16:35
        Swiss HPC Tier-2 Computing @ CSCS 25m

        For the past 10 years, CSCS has been running compute capability in the WLCG Tier-2 for ATLAS, CMS and LHCb on standard commodity hardware (a cluster named Phoenix). Three years ago, CSCS began providing this service on the flagship High Performance Computing (HPC) system, Piz Daint (a Cray XC40/50 system). Piz Daint is a world-class HPC system with over 1800 dual-processor multicore nodes and more than 5700 hybrid compute nodes with GPU accelerators. Piz Daint currently holds the 5th position on the Top500 List and is the most powerful HPC system in Europe.

        In preparation for future challenges that the HL-LHC will impose on the computing sites, CSCS is in the process of decommissioning the Phoenix cluster and fully consolidating the Tier-2 compute load onto Piz Daint.

        In this presentation, the critical milestones to achieve on the road to a successful migration to Piz Daint will be explained.

        Speaker: Mr Dino Conciatore (CSCS (Swiss National Supercomputing Centre))
      • 17:00
        Deep learning in a container, experience and best practices 25m

        Deep Learning techniques are gaining interest in the High Energy Physics, following a new and efficient approach to solve different problems. These techniques leverage the specific features of GPU accelerators and rely on a set of software packages allowing users to compute on GPUs and program Deep Learning algorithms. However, the rapid pace at which both the hardware and the low and high level libraries are evolving poses several operational issues to computing centers such as the IN2P3 Computing Center (CC-IN2P3 -- http://cc.in2p3.fr).

        In this talk we present how we addressed these operational challenges thanks to the use of container technologies. We show that the flexibility offered by containers comes with no overhead while allowing users to benefit of the better performance of compiled from sources versions of popular deep learning frameworks. Finally, we detail the best practices proposed to the users of the CC-IN2P3 to prepare and submit their deep learning oriented jobs on the available GPU resources.

        Speaker: Frederic Suter
    • 18:00 21:00
      Conference Dinner 3h Draft Republic

      Draft Republic

      282 Esplanade Ct San Diego, CA 92122
    • 09:00 10:40
      IT Facilities & Business Continuity E-B 212

      E-B 212

      SDSC Auditorium

      10100 Hopkins Drive La Jolla, CA 92093-0505
      • 09:00
        Cost and system performance modelling in WLCG and HSF: an update 25m

        The HSF/WLCG cost and performance modeling working group was established in November 2017 and has since then achieved considerable progress in our understanding of the performance factors of the LHC applications, the estimation of the computing and storage resources and the cost of the infrastructure and its evolution for the WLCG sites. This contribution provides an update on the recent developments of the working group activities, with a special focus on the implications for computing sites.

        Speakers: Jose Flix Molina (Centro de Investigaciones Energéti cas Medioambientales y Tecno), Dr Andrea Sciabà (CERN)
      • 09:25
        RACF/SDCC Datacenter Transformation within the Scope of BNL CFR Project and Beyond 25m

        The BNL Computing Facility Revitalization (CFR) project aimed at repurposing the former National Synchrotron Light Source (NSLS-I) building (B725) located on BNL site as the new datacenter for BNL Computational Science Initiative (CSI) and RACF/SDCC Facility in particular. The CFR project is currently wrapping up the design phase and expected to enter the construction phase in the first half of 2019. The new B725 data center is to become available in early 2021 for ATLAS compute, disk storage and tape storage equipment, and later during the year of 2021 - for all other collaborations supported by the RACF/SDCC Facility, including but not limited to: STAR and PHENIX experiments at RHIC collider at BNL, Belle II Experiment at KEK (Japan), and BNL CSI HPC clusters. Migration of the majority of IT payload from B515 based datacenter to the B725 datacenter is expected to begin even earlier, as the central networking systems and first BNL ATLAS Tier-1 Site tape robot are to be deployed in B725 starting from early FY21, and expected to continue throughout the period of FY21-23, leaving the B515 datacenter physically reduced down to a subset of areas it is currently occupying, and drastically reducing its power profile. In this talk I am going to highlight the main design features of the new RACF/SDCC datacenter, summarize the preparation activities already underway in our existing datacenter since FY18 needed to ensure a smooth transition B515 and B725 datacenters inter-operation period in FY21, discuss the planned sequence of equipment migration between these two datacenters in FY21 and gradual equipment replacement in FY21-24, and also show the expected state of occupancy and infrastructure utilization for both datacenters in FY25.

        Speaker: Alexandr Zaytsev (Brookhaven National Laboratory (US))
      • 09:50
        Omni, N9 and the Superfacility 25m

        I will be presenting how we are using our data collection framework (Omni) to help facilitate the installation of N9 (our new system) and how this all ties together with the Superfacility concept which mentioned in the fall.

        Speaker: Cary Whitney (LBNL)
      • 10:15
        HEPiX Technology Watch working group 25m

        A short report on what has happened, how we have organised ourselves, how we intend to present results etc. Note that the findings themselves will be discussed in other contributions - this is about how the group works.

        Speaker: Helge Meinhard (CERN)
    • 10:40 11:10
      Coffee Break 30m
    • 11:10 12:25
      IT Facilities & Business Continuity E-B 212

      E-B 212

      SDSC Auditorium

      10100 Hopkins Drive La Jolla, CA 92093-0505
    • 12:25 14:00
      Lunch Break 1h 35m
    • 14:00 14:25
      Storage & Filesystems E-B 212

      E-B 212

      SDSC Auditorium

      10100 Hopkins Drive La Jolla, CA 92093-0505
      • 14:00
        Keeping Pace with Science: How a Modern Filesystem Can Accelerate Discovery 25m

        In November 2018, running on a mere half-rack of ordinary SuperMicro servers, WekaIO's Matrix Filesystem outperformed 40 racks of specialty hardware on Oak Ridge National Labs' Summit system, yielding the #1 ranked result for the IO-500 10-Node Challenge. How can that even be possible?

        This level of performance becomes important for modern use cases whether they involve GPU-accelerated servers for artificial intelligence and deep learning or traditional CPU-based servers at massive scale. Teams of researchers and data scientists should be free to focus on their work and not lose precious time waiting for results caused by IO bottlenecks. An example use case within HEP where this technology may be most useful is the production of pre-mixing libraries in experiments like CMS. CMS uses at present a 600TB “library” to simulate overlapping proton proton collisions during its simulation campaigns. The production of this library is an IO limited workflow on any filesystem in use within the experiment today.

        In this tech-talk, the architecture of the Matrix filesystem will be put under the microscope, explored and discussed. This talk will include real-world examples of data intensive workloads along with a variety of benchmark results that show the filesystem's versatility and ability to scale.

        Speaker: Mr Andy Watson (WekaIO)
    • 14:25 15:15
      Basic IT Services E-B 212

      E-B 212

      SDSC Auditorium

      10100 Hopkins Drive La Jolla, CA 92093-0505
      • 14:25
        Config Management and Deployment Setup at KIT 25m

        For several years, the GridKa Tier-1 center, the Large Scale Data Facility and other infrastructures at KIT have been using Puppet and Foreman for configuration management and machine deployment.
        We will present our experiences, the workflows that are used and our current efforts to establish a completely integrated system for all our infrastructures based on Katello.

        Speaker: Andreas Petzold (KIT - Karlsruhe Institute of Technology (DE))
      • 14:50
        Token Renewal Service (TRS) at SLAC 25m

        The token renewal service (TRS) has been used at SLAC National Accelerator
        Laboratory since the late 1990s. In 2018 it was found to be lacking in some
        critical areas (encryption types used and other basic mechanism would no
        longer be available for post Red Hat 6 systems.)

        1-to-1 replacement areas:

        Running Batch Jobs:
        Our local solution to batch jobs (LSF): The need for TRS was already resolved
        with the movement to a new mechanism used for re-authorizing in IBM's LSF
        version ?.??.

        Our remote solution for batch (like) jobs: OSGrid and CVMFS services have no requirement for a TRS type solution. In the area of using remote (full or partial-stack) computing resources from the industry titans, (e.g. Azure, AWS and Google), SLAC-OCIO does not actively use those resources as of March 2019.

        Users that run TRScron jobs: a distributed cron service which leverages
        the token renewal service, the need remained.

        Questions:

        What adaptations should be pondered in the face of future computing workloads?

        How does a new vision of TRS compare to past and present mechanisms used to
        provide renewable secure access to a distributed service. What are we missing? Comments, questions, and concerns?

        https://confluence.slac.stanford.edu/display/TRS/TRS+Home

        Speaker: Andrew May (SLAC National ACCELERATOR LABORATORY)
    • 15:15 15:45
      Coffee Break 30m
    • 15:45 16:10
      Grid, Cloud and Virtualization E-B 212

      E-B 212

      SDSC Auditorium

      10100 Hopkins Drive La Jolla, CA 92093-0505
      • 15:45
        The glideinWMS system: recent developments 25m

        GlideinWMS is a workload management and provisioning system that lets
        you share computing resources distributed over independent sites. A
        dynamically sized pool of resources is created by GlideinWMS pilot
        Factories, based on the requests made by GlideinWMS Frontends. More
        than 400 computing elements are currently serving  more than 10
        virtual organizations through glideinWMS. This contribution will give
        an overview of the glideinWMS setup, and will present the recent
        developments in the project, including the addition of the singularity
        support, and the improvements to minimize resource wastages. Future
        enhancements for automatizing the generation of facotry configurations
        will also be outlined.

        Speaker: Marco Mascheroni (Univ. of California San Diego (US))
    • 16:10 17:00
      Birds of a Feather (BoF) E-B 212

      E-B 212

      SDSC Auditorium

      10100 Hopkins Drive La Jolla, CA 92093-0505
    • 09:00 10:15
      Grid, Cloud and Virtualization E-B 212

      E-B 212

      SDSC Auditorium

      10100 Hopkins Drive La Jolla, CA 92093-0505
      • 09:00
        The Experience and Challenge in Grid Computing at KEK 25m

        The KEK Central Computer System (KEKCC) is a service, which provides large-scale computer resources, Grid and Cloud computing, as well as common IT services. The KEKCC is entirely replaced every four or five years according to Japanese government procurement policy for the computer system. Current KEKCC has been in operation since September 2016 and decommissioning will start in early 2020.

        In this talk, we would like to share our experiences and challenges for the security, operation, and some applications dedicated to each experiment. In particular, we report several improvements on the Grid computing system for the Belle II experiment based on the nearly three years operational performance of the KEKCC. We also introduce a prospect for the next KEKCC which is planned to be launched in September 2020.

        Speaker: Go Iwai (KEK)
      • 09:25
        Creating an opportunistic OSG site inside the PRP Kubernetes cluster 25m

        The Pacific Research Platform (PRP) is operating a Kubernetes cluster that manages over 2.5k CPU cores and 250 GPUs. Most of the resources are being used by local users interactively starting directly Kubernetes Pods.

        To fully utilize the available resources, we have deployed an opportunistic HTCondor pool as a Kubernetes deployment, with worker nodes environment being fully OSG compliant. This includes both the OSG client software and CVMFS. A OSG HTCondor-CE is available for OSG users to access the resources as any other OSG site. The first user of the new site is the IceCube collaboration, which is using the available GPUs.

        In this presentation we will describe the steps (and challenges) involved in creating the opportunistic OSG site in the Kubernetes cluster and the experience of running GPU jobs of the IceCube collaboration.

        Speaker: Igor Sfiligoi (UCSD)
      • 09:50
        Public cloud for high throughput computing 25m

        The vast breadth and configuration possibilities of the public cloud offer intriguing opportunities for loosely coupled computing tasks. One such class of tasks is simply statistical in nature requiring many independent trials over the targeted phase space in order to converge on robust, fault tolerant and optimized designs. Our single threaded target application (50-200 MB) solves a stochastic non-linear integro-differential equation relevant for read/write simulations of heat assisted magnetic recording (HAMR) for high areal density hard disk drives (HDD). Here, the phase space is multi-dimensional in physical parameters and potential recording schemes. Furthermore, for any one such point in phase space, 100s of simulations must be repeated due to the stochastic nature of the physical simulation.

        In this talk, we show that a simple abstraction layer between the target application and cloud vendor provided batch systems can be easily constructed thus avoiding changes to the underlying simulation and workflow. With some planning, this abstraction layer is portable between three available cloud providers: Amazon Web Services, Microsoft Azure and Google Cloud. This abstraction layer is required to be light weight and not introduce significant overhead and was implemented as simple Bash scripts. To reduce cost, it was critical to test the application under multiple configurations (e.g. instance types and compiling options), avoid local block storage and minimize network traffic. Fleets of 100,000 concurrent simulations are easily achieved with over 99.99% of the cost just for compute (versus storage or network). By implementing a third party grid engine, 1,000,000 concurrent simulations were achieved with no modifications to the abstraction layer.

        Best practices and design principles for HTC in public cloud will be discussed with emphasis on robustness, cost and horizontal scale and unique challenges encountered in this migration.

        Speaker: Dr Gregory Parker (Entonos)
    • 10:15 10:45
      Coffee Break 30m
    • 10:45 12:00
      Grid, Cloud and Virtualization E-B 212

      E-B 212

      SDSC Auditorium

      10100 Hopkins Drive La Jolla, CA 92093-0505
      • 10:45
        6 years of CERN Cloud - From 0 to 300k cores 25m

        CERN, the European Laboratory for Particle Physics, is running OpenStack for its private Could Infrastructure among other leading open source tools that helps thousands of scientists around the world to uncover the mysteries of the Universe.
        In 2012, CERN started the deployment of its private Cloud Infrastructure using OpenStack. Since then we moved from few hundred cores to a multi-cell deployment spread between two data centres.
        After 6 years deploying and managing OpenStack at scale, we now look back and discuss the challenges of building a massive scale infrastructure from 0 to +300K cores.
        With this talk we will dive into the history, architecture, tools and technical decisions behind the CERN Cloud Infrastructure.

        Speaker: Belmiro Moreira (CERN)
      • 11:10
        Developing for a Services Layer At The Edge (SLATE) 25m

        Modern software development workflow patterns often involve the use of a developer’s local machine as the first platform for testing code. SLATE mimics this paradigm with an implementation of a light-weight version, called MiniSLATE, that runs completely contained on the developer local machine (laptop, virtual machine, or another physical server). MiniSLATE resolves many development environment issues by providing an isolated and local configuration for the developer. Application developers are able to download MiniSLATE which provides a fully orchestrated set of containers on top of a production SLATE platform, complete with central information service, API server, and a local Kubernetes cluster. This approach mitigates the overhead of a hypervisor but still provides the requisite isolated environment. They are able to create the environment, iterate, destroy it, and repeat at will. A local MiniSLATE environment also allows the developer to explore the packaging of the edge service within a constrained security context in order to validate its full functionality within limited permissions. As a result, developers are able to test the functionality of their application with the complete complement of SLATE components local to their development environment without the overhead of building a cluster or virtual machine, registering a cluster, interacting with the production SLATE platform, etc.

        Speaker: Mr Ben Kulbertis (University of Utah)
      • 11:35
        Changes to OSG and how they affect US WLCG sites 25m

        In the spring of 2018, central operations services were migrated out of the Grid Operations Center of Indiana into other participating Open Science Grid institutions. This talk summarizes how the migration has affected the services provided by the OSG, and gives a summary of how central OSG services interface with US WLCG sites.

        Speaker: Jeffrey Michael Dost (Univ. of California San Diego (US))
    • 12:00 12:30
      Workshop wrap-up E-B 212

      E-B 212

      SDSC Auditorium

      10100 Hopkins Drive La Jolla, CA 92093-0505