Conveners
Software: Tue AM
- Benjamin Krikler (University of Bristol (GB))
- David Bouvet (IN2P3/CNRS (FR))
Software: Tue PM
- Teng Jian Khoo (Humboldt University of Berlin (DE))
- Enrico Guiraud (EP-SFT, CERN)
Software: Wed AM
- Stefan Roiser (CERN)
- Enrico Guiraud (EP-SFT, CERN)
Software: Wed PM
- Teng Jian Khoo (Humboldt University of Berlin (DE))
- Luisa Arrabito (LUPM IN2P3/CNRS)
-
Haolai Tian (Institute of High Energy Physics)18/05/2021, 10:50Online ComputingShort Talk
Daisy (Data Analysis Integrated Software System) has been designed for the analysis and visualization of the X-ray experiments. To address an extensive range of Chinese radiation facilities community’s requirements from purely algorithmic problems to scientific computing infrastructure, Daisy sets up a cloud-native platform to support on-site data analysis services with fast feedback and...
Go to contribution page -
Pieter David (Universite Catholique de Louvain (UCL) (BE))18/05/2021, 11:03Offline ComputingShort Talk
With the LHC continuing to collect more data and experimental analyses becoming increasingly complex, tools to efficiently develop and execute
Go to contribution page
these analyses are essential. The bamboo framework defines a domain-specific
language, embedded in python, that allows to concisely express the analysis
logic in a functional style. The implementation based on ROOT’s RDataFrame
and cling C++ JIT... -
Gokhan Unel (University of California Irvine (US))18/05/2021, 11:16Offline ComputingShort Talk
This paper presents an overview and features of an Analysis Description Language (ADL) designed for HEP data analysis. ADL is a domain-specific, declarative language that describes the physics content of an analysis in a standard and unambiguous way, independent of any computing frameworks. It also describes infrastructures that render ADL executable, namely CutLang, a direct runtime...
Go to contribution page -
Anton Alkin (CERN)18/05/2021, 11:29Offline ComputingShort Talk
In LHC Run 3 the ALICE Collaboration will have to cope in Run 3 with an increase of lead-lead collision data of two orders of magnitude com- pared to the Run 1 and 2 data-taking periods. The Online-Offline (O$^2$) software framework has been developed to allow for distributed and efficient process- ing of this unprecedented amount of data. Its design, which is based on a message-passing back...
Go to contribution page -
Vit Kucera (CERN)18/05/2021, 11:42Offline ComputingShort Talk
Precise measurements of heavy-flavour hadrons down to very low pT represent the core of the physics program of the upgraded ALICE experiment in Run 3.
Go to contribution page
These physics probes are characterised by a very small signal-to-background ratio requiring very large statistics of minimum-bias events.
In Run 3, ALICE is expected to collect up to 13 nb^{-1} of lead–lead collisions, corresponding to about... -
Mason Proffitt (University of Washington (US))18/05/2021, 11:55Offline ComputingShort Talk
The traditional approach in HEP analysis software is to loop over every event and every object via the ROOT framework. This method follows an imperative paradigm, in which the code is tied to the storage format and steps of execution. A more desirable strategy would be to implement a declarative language, such that the storage medium and execution are not included in the abstraction model....
Go to contribution page -
Vakho Tsulaia (Lawrence Berkeley National Lab. (US))18/05/2021, 15:00Offline ComputingShort Talk
The GeoModel class library for detector description has recently been released as an open-source package and extended with a set of tools to allow much of the detector modeling to be carried out in a lightweight development environment, outside of large and complex software frameworks. These tools include the mechanisms for creating persistent representation of the geometry, an interactive 3D...
Go to contribution page -
Dr Antonio Augusto Alves Junior (Institute for Astroparticle Physics of Karlsruhe Institute of Technology)18/05/2021, 15:26Offline ComputingShort Talk
This document is devoted to the description of advances in the generation of high-quality random numbers for CORSIKA 8, which is being developed in modern C++17 and is designed to run on modern multi-thread processors and accelerators. CORSIKA 8 is a Monte Carlo simulation framework to model ultra-high energy secondary particle cascades in astroparticle physics. The aspects associated with...
Go to contribution page -
Markus Frank (CERN)18/05/2021, 15:39Offline ComputingShort Talk
Consistent detector description is an integral part of all modern experiments and also the main motivation behind the creation of DD4hep, which tries to address detector description in a broad sense including: geometry and the materials used in the device, additional parameters describing e.g. the detection techniques, constants required for alignment and calibration, description of the...
Go to contribution page -
Andre Sailer (CERN)18/05/2021, 15:52Offline ComputingShort Talk
Detector optimisation and physics performance studies are an
Go to contribution page
integral part for the development of future collider
experiments. The Key4hep project aims to design a common set of
software tools for future, or even present, High Energy Physics
projects. These proceedings describe the main components that are
developed as part of Key4hep: the event data model EDM4hep,
simulation interfaces... -
Daniel Britzger (Max-Planck-Institut für Physik München)18/05/2021, 16:05Offline ComputingShort Talk
The lepton–proton collisions produced at the HERA collider represent a unique high energy physics data set. A number of years after the end of collisions, the data collected by the H1 experiment, as well as the simulated events and all software needed for reconstruction, simulation and data analysis were migrated into a preserved operational mode at DESY. A recent modernisation of the H1...
Go to contribution page -
Vasyl Hafych (Max-Planck-Institut fur Physik (DE))19/05/2021, 10:50Offline ComputingShort Talk
We present BAT.jl 2.0, the next generation of the Bayesian Analysis Toolkit. BAT.jl is a highly efficient and easy to use software package for Bayesian Inference. It's predecessor, BAT 1.0 in C++, has been very successful over the years with a large number of citations. Our new incarnation of BAT was rewritten from scratch in Julia and we recently released the long-term stable version...
Go to contribution page -
Frank Berghaus (Argonne National Laboratory (US))19/05/2021, 11:03Offline ComputingShort Talk
Processing and scientific analysis of the data taken by the ATLAS experiment requires reliable information describing the event data recorded by the detector or generated in software. ATLAS event processing applications store such descriptive metadata information in the output data files along with the event information.
To better leverage the available computing resources during LHC Run3...
Go to contribution page -
Anastasiia Zhadan (BINP)19/05/2021, 11:16Offline ComputingShort Talk
The project of Super Charm-Tau (SCT) factory --- a high-luminosity
Go to contribution page
electron-positron collider for studying charmed hadrons and tau lepton
--- is proposed by Budker INP. The project implies single collision point
equipped with a universal particle detector. The Aurora software
framework has been developed for the SCT detector. It is based on
trusted and widely used in high energy physics... -
Michal Kamil Simon (CERN)19/05/2021, 11:29Distributed Computing, Data Management and FacilitiesShort Talk
Across the years, being the backbone of numerous data management solutions used within the WLCG collaboration, the XRootD framework and protocol became one of the most important building blocks for storage solutions in the High Energy Physics (HEP) community. The latest big milestone for the project, release 5, introduced multitude of architectural improvements and functional enhancements,...
Go to contribution page -
Alexander Held (New York University (US))19/05/2021, 11:42Offline ComputingShort Talk
The cabinetry library provides a Python-based solution for building and steering binned template fits. It tightly integrates with the pythonic High Energy Physics ecosystem, and in particular with pyhf for statistical inference. cabinetry uses a declarative approach for building statistical models, with a JSON schema describing possible configuration choices. Model building instructions can...
Go to contribution page -
Ralf Ulrich (KIT - Karlsruhe Institute of Technology (DE))19/05/2021, 11:55Offline ComputingShort Talk
The CORSIKA 8 project is an international collaboration of scientists working together to deliver the most modern, flexible, robust and efficient framework for the simulation of ultra-high energy secondary particle cascades in matter. The main application is for cosmic ray air shower simulations, but is not limited to that. Besides a comprehensive collection of physics models and algorithms...
Go to contribution page -
177. Grid-based minimization at scale: Feldman-Cousins corrections for light sterile neutrino searchMarianette Wospakrik (Fermi National Accelerator Laboratory)19/05/2021, 17:40Distributed Computing, Data Management and FacilitiesShort Talk
High Energy Physics (HEP) experiments generally employ sophisticated statistical methods to present results in searches of new physics. In the problem of searching for sterile neutrinos, likelihood ratio tests are applied to short-baseline neutrino oscillation experiments to construct confidence intervals for the parameters of interest. The test statistics of the form $\Delta \chi^2$ is often...
Go to contribution page -
Andrew Malone Melo (Vanderbilt University (US))19/05/2021, 17:53Distributed Computing, Data Management and FacilitiesShort Talk
Apache Spark is one of the predominant frameworks in the big data space, providing a fully-functional query processing engine, vendor support for hardware accelerators, and performant integrations with scientific computing libraries. One difficulty in adopting conventional big data frameworks to HEP workflows is the lack of support for the ROOT file format in these frameworks. Laurelin...
Go to contribution page -
Mr Vincenzo Eduardo Padulano (Valencia Polytechnic University (ES))19/05/2021, 18:06Distributed Computing, Data Management and FacilitiesShort Talk
Thanks to its RDataFrame interface, ROOT now supports the execution of the same physics analysis code both on a single machine and on a cluster of distributed resources. In the latter scenario, it is common to read the input ROOT datasets over the network from remote storage systems, which often increases the time it takes for physicists to obtain their results. Storing the remote files much...
Go to contribution page -
Nikolai Hartmann (Ludwig Maximilians Universitat (DE))19/05/2021, 18:19Offline ComputingShort Talk
Future analysis of ATLAS data will involve new small-sized analysis
Go to contribution page
formats to cope with the increased storage needs. The smallest of
these, named DAOD_PHYSLITE, has calibrations already applied
to allow fast downstream analysis and avoid the need for further
analysis-specific intermediate formats. This allows for application
of the "columnar analysis" paradigm where operations are... -
Jim Pivarski (Princeton University)19/05/2021, 18:32Offline ComputingShort Talk
File formats for generic data structures, such as ROOT, Avro, and Parquet, pose a problem for deserialization: it must be fast, but its code depends on the type of the data structure, not known at compile-time. Just-in-time compilation can satisfy both constraints, but we propose a more portable solution: specialized virtual machines. AwkwardForth is a Forth-driven virtual machine for...
Go to contribution page -
Gordon Watts (University of Washington (US))19/05/2021, 18:45Offline ComputingShort Talk
Array operations are one of the most concise ways of expressing common filtering and simple aggregation operations that is the hallmark of the first step of a particle physics analysis: selection, filtering, basic vector operations, and filling histograms. The High Luminosity run of the Large Hadron Collider (HL-LHC), scheduled to start in 2026, will require physicists to regularly skim...
Go to contribution page