Session

Software

18 May 2021, 10:50

Conveners

Software: Tue AM

  • Benjamin Krikler (University of Bristol (GB))
  • David Bouvet (IN2P3/CNRS (FR))

Software: Tue PM

  • Teng Jian Khoo (Humboldt University of Berlin (DE))
  • Enrico Guiraud (EP-SFT, CERN)

Software: Wed AM

  • Stefan Roiser (CERN)
  • Enrico Guiraud (EP-SFT, CERN)

Software: Wed PM

  • Teng Jian Khoo (Humboldt University of Berlin (DE))
  • Luisa Arrabito (LUPM IN2P3/CNRS)

Presentation materials

  1. Haolai Tian (Institute of High Energy Physics)
    18/05/2021, 10:50
    Online Computing
    Short Talk

    Daisy (Data Analysis Integrated Software System) has been designed for the analysis and visualization of the X-ray experiments. To address an extensive range of Chinese radiation facilities community’s requirements from purely algorithmic problems to scientific computing infrastructure, Daisy sets up a cloud-native platform to support on-site data analysis services with fast feedback and...

    Go to contribution page
  2. Pieter David (Universite Catholique de Louvain (UCL) (BE))
    18/05/2021, 11:03
    Offline Computing
    Short Talk

    With the LHC continuing to collect more data and experimental analyses becoming increasingly complex, tools to efficiently develop and execute
    these analyses are essential. The bamboo framework defines a domain-specific
    language, embedded in python, that allows to concisely express the analysis
    logic in a functional style. The implementation based on ROOT’s RDataFrame
    and cling C++ JIT...

    Go to contribution page
  3. Gokhan Unel (University of California Irvine (US))
    18/05/2021, 11:16
    Offline Computing
    Short Talk

    This paper presents an overview and features of an Analysis Description Language (ADL) designed for HEP data analysis. ADL is a domain-specific, declarative language that describes the physics content of an analysis in a standard and unambiguous way, independent of any computing frameworks. It also describes infrastructures that render ADL executable, namely CutLang, a direct runtime...

    Go to contribution page
  4. Anton Alkin (CERN)
    18/05/2021, 11:29
    Offline Computing
    Short Talk

    In LHC Run 3 the ALICE Collaboration will have to cope in Run 3 with an increase of lead-lead collision data of two orders of magnitude com- pared to the Run 1 and 2 data-taking periods. The Online-Offline (O$^2$) software framework has been developed to allow for distributed and efficient process- ing of this unprecedented amount of data. Its design, which is based on a message-passing back...

    Go to contribution page
  5. Vit Kucera (CERN)
    18/05/2021, 11:42
    Offline Computing
    Short Talk

    Precise measurements of heavy-flavour hadrons down to very low pT represent the core of the physics program of the upgraded ALICE experiment in Run 3.
    These physics probes are characterised by a very small signal-to-background ratio requiring very large statistics of minimum-bias events.
    In Run 3, ALICE is expected to collect up to 13 nb^{-1} of lead–lead collisions, corresponding to about...

    Go to contribution page
  6. Mason Proffitt (University of Washington (US))
    18/05/2021, 11:55
    Offline Computing
    Short Talk

    The traditional approach in HEP analysis software is to loop over every event and every object via the ROOT framework. This method follows an imperative paradigm, in which the code is tied to the storage format and steps of execution. A more desirable strategy would be to implement a declarative language, such that the storage medium and execution are not included in the abstraction model....

    Go to contribution page
  7. Vakho Tsulaia (Lawrence Berkeley National Lab. (US))
    18/05/2021, 15:00
    Offline Computing
    Short Talk

    The GeoModel class library for detector description has recently been released as an open-source package and extended with a set of tools to allow much of the detector modeling to be carried out in a lightweight development environment, outside of large and complex software frameworks. These tools include the mechanisms for creating persistent representation of the geometry, an interactive 3D...

    Go to contribution page
  8. Dr Antonio Augusto Alves Junior (Institute for Astroparticle Physics of Karlsruhe Institute of Technology)
    18/05/2021, 15:26
    Offline Computing
    Short Talk

    This document is devoted to the description of advances in the generation of high-quality random numbers for CORSIKA 8, which is being developed in modern C++17 and is designed to run on modern multi-thread processors and accelerators. CORSIKA 8 is a Monte Carlo simulation framework to model ultra-high energy secondary particle cascades in astroparticle physics. The aspects associated with...

    Go to contribution page
  9. Markus Frank (CERN)
    18/05/2021, 15:39
    Offline Computing
    Short Talk

    Consistent detector description is an integral part of all modern experiments and also the main motivation behind the creation of DD4hep, which tries to address detector description in a broad sense including: geometry and the materials used in the device, additional parameters describing e.g. the detection techniques, constants required for alignment and calibration, description of the...

    Go to contribution page
  10. Andre Sailer (CERN)
    18/05/2021, 15:52
    Offline Computing
    Short Talk

    Detector optimisation and physics performance studies are an
    integral part for the development of future collider
    experiments. The Key4hep project aims to design a common set of
    software tools for future, or even present, High Energy Physics
    projects. These proceedings describe the main components that are
    developed as part of Key4hep: the event data model EDM4hep,
    simulation interfaces...

    Go to contribution page
  11. Daniel Britzger (Max-Planck-Institut für Physik München)
    18/05/2021, 16:05
    Offline Computing
    Short Talk

    The lepton–proton collisions produced at the HERA collider represent a unique high energy physics data set. A number of years after the end of collisions, the data collected by the H1 experiment, as well as the simulated events and all software needed for reconstruction, simulation and data analysis were migrated into a preserved operational mode at DESY. A recent modernisation of the H1...

    Go to contribution page
  12. Vasyl Hafych (Max-Planck-Institut fur Physik (DE))
    19/05/2021, 10:50
    Offline Computing
    Short Talk

    We present BAT.jl 2.0, the next generation of the Bayesian Analysis Toolkit. BAT.jl is a highly efficient and easy to use software package for Bayesian Inference. It's predecessor, BAT 1.0 in C++, has been very successful over the years with a large number of citations. Our new incarnation of BAT was rewritten from scratch in Julia and we recently released the long-term stable version...

    Go to contribution page
  13. Frank Berghaus (Argonne National Laboratory (US))
    19/05/2021, 11:03
    Offline Computing
    Short Talk

    Processing and scientific analysis of the data taken by the ATLAS experiment requires reliable information describing the event data recorded by the detector or generated in software. ATLAS event processing applications store such descriptive metadata information in the output data files along with the event information.

    To better leverage the available computing resources during LHC Run3...

    Go to contribution page
  14. Anastasiia Zhadan (BINP)
    19/05/2021, 11:16
    Offline Computing
    Short Talk

    The project of Super Charm-Tau (SCT) factory --- a high-luminosity
    electron-positron collider for studying charmed hadrons and tau lepton
    --- is proposed by Budker INP. The project implies single collision point
    equipped with a universal particle detector. The Aurora software
    framework has been developed for the SCT detector. It is based on
    trusted and widely used in high energy physics...

    Go to contribution page
  15. Michal Kamil Simon (CERN)
    19/05/2021, 11:29
    Distributed Computing, Data Management and Facilities
    Short Talk

    Across the years, being the backbone of numerous data management solutions used within the WLCG collaboration, the XRootD framework and protocol became one of the most important building blocks for storage solutions in the High Energy Physics (HEP) community. The latest big milestone for the project, release 5, introduced multitude of architectural improvements and functional enhancements,...

    Go to contribution page
  16. Alexander Held (New York University (US))
    19/05/2021, 11:42
    Offline Computing
    Short Talk

    The cabinetry library provides a Python-based solution for building and steering binned template fits. It tightly integrates with the pythonic High Energy Physics ecosystem, and in particular with pyhf for statistical inference. cabinetry uses a declarative approach for building statistical models, with a JSON schema describing possible configuration choices. Model building instructions can...

    Go to contribution page
  17. Ralf Ulrich (KIT - Karlsruhe Institute of Technology (DE))
    19/05/2021, 11:55
    Offline Computing
    Short Talk

    The CORSIKA 8 project is an international collaboration of scientists working together to deliver the most modern, flexible, robust and efficient framework for the simulation of ultra-high energy secondary particle cascades in matter. The main application is for cosmic ray air shower simulations, but is not limited to that. Besides a comprehensive collection of physics models and algorithms...

    Go to contribution page
  18. Marianette Wospakrik (Fermi National Accelerator Laboratory)
    19/05/2021, 17:40
    Distributed Computing, Data Management and Facilities
    Short Talk

    High Energy Physics (HEP) experiments generally employ sophisticated statistical methods to present results in searches of new physics. In the problem of searching for sterile neutrinos, likelihood ratio tests are applied to short-baseline neutrino oscillation experiments to construct confidence intervals for the parameters of interest. The test statistics of the form $\Delta \chi^2$ is often...

    Go to contribution page
  19. Andrew Malone Melo (Vanderbilt University (US))
    19/05/2021, 17:53
    Distributed Computing, Data Management and Facilities
    Short Talk

    Apache Spark is one of the predominant frameworks in the big data space, providing a fully-functional query processing engine, vendor support for hardware accelerators, and performant integrations with scientific computing libraries. One difficulty in adopting conventional big data frameworks to HEP workflows is the lack of support for the ROOT file format in these frameworks. Laurelin...

    Go to contribution page
  20. Mr Vincenzo Eduardo Padulano (Valencia Polytechnic University (ES))
    19/05/2021, 18:06
    Distributed Computing, Data Management and Facilities
    Short Talk

    Thanks to its RDataFrame interface, ROOT now supports the execution of the same physics analysis code both on a single machine and on a cluster of distributed resources. In the latter scenario, it is common to read the input ROOT datasets over the network from remote storage systems, which often increases the time it takes for physicists to obtain their results. Storing the remote files much...

    Go to contribution page
  21. Nikolai Hartmann (Ludwig Maximilians Universitat (DE))
    19/05/2021, 18:19
    Offline Computing
    Short Talk

    Future analysis of ATLAS data will involve new small-sized analysis
    formats to cope with the increased storage needs. The smallest of
    these, named DAOD_PHYSLITE, has calibrations already applied
    to allow fast downstream analysis and avoid the need for further
    analysis-specific intermediate formats. This allows for application
    of the "columnar analysis" paradigm where operations are...

    Go to contribution page
  22. Jim Pivarski (Princeton University)
    19/05/2021, 18:32
    Offline Computing
    Short Talk

    File formats for generic data structures, such as ROOT, Avro, and Parquet, pose a problem for deserialization: it must be fast, but its code depends on the type of the data structure, not known at compile-time. Just-in-time compilation can satisfy both constraints, but we propose a more portable solution: specialized virtual machines. AwkwardForth is a Forth-driven virtual machine for...

    Go to contribution page
  23. Gordon Watts (University of Washington (US))
    19/05/2021, 18:45
    Offline Computing
    Short Talk

    Array operations are one of the most concise ways of expressing common filtering and simple aggregation operations that is the hallmark of the first step of a particle physics analysis: selection, filtering, basic vector operations, and filling histograms. The High Luminosity run of the Large Hadron Collider (HL-LHC), scheduled to start in 2026, will require physicists to regularly skim...

    Go to contribution page
Building timetable...