ROOT is a unified software package for the storage, processing, and analysis of scientific data: from its acquisition to the final visualization in the form of highly customizable, publication-ready plots. Successfully used by experiments and thousands of physicists, the ROOT Project is preparing its seventh release cycle, sustained by intense R&D activities.
In this contribution, after...
One of the objectives of the EOSC (European Open Science Cloud) Future Project was to integrate diverse analysis workflows from Cosmology, Astrophysics and High Energy Physics in a common framework. This led to the inception of the Virtual Research Environment (VRE) at CERN, a prototype platform supporting the goals of Dark Matter and Extreme Universe Science Projects in compliance with FAIR...
The Geant4 simulation throughput of LHC experiments is limited by increasing detector complexity in the high-luminosity phase. As high-performance computing shifts toward heterogeneous architectures such as GPUs, GPU-accelerated particle transport simulations offer a potential way to improve performance. Currently, only electromagnetic showers can be offloaded to GPUs, making an efficient...
The existing roadmaps and computing model plans from ATLAS and CMS for the HL-LHC area are primarily focused on the centralized aspect of computing: those steps that lead up to sets of files made available to physicists for analysis. The general approaches, resources used, and software frameworks for the area of “end-user physics analysis”, which starts from those files, are much less clearly...
As the HL-LHC era approaches, the scale and complexity of data present challenges for analysis workflows within ATLAS and other HL-LHC experiments. This contribution reports on recent developments in ServiceX, a cross experiment utility, and its role as a data delivery and transformation service within the analysis ecosystem. Designed to bridge the gap between centrally produced datasets and...
This contribution will present how the EVERSE project interfaces with the European Open Science Clusters (ENVRI-FAIR for environmental sciences, Life Sciences RI, ESCAPE for Particle physics and astrophysics, PaNOSC for Photon and neutron science and SSHOC for social sciences and humanities) through use cases of software packages or infrastructures that are in current used by researchers....
Scikit-HEP is a community-driven and community-oriented project with the goal of providing an ecosystem for particle physics data analysis in Python fully integrated with the wider scientific Python ecosystem. The project provides many packages and a few “affiliated” packages for data analysis. It expands the typical Python data analysis tools for particle physicists, with packages spanning...
Marionette is a header-only C++ library that was designed to allow the description of arbitrary data structures that can work across heterogeneous compute devices and on the host, providing complete interoperability and convenient interfaces with no impact on runtime performance. This is achieved by decoupling the description of the data to be held from the way in which data will be stored,...
AUDITOR (Accounting Data Handling Toolbox for Opportunistic Resources) is a flexible and extensible accounting system designed to support a wide range of use cases and infrastructures. Its integration with APEL enables it to function as a generic component within the WLCG accounting infrastructure, tracking the usage of various types of site computing resources. Several WLCG sites have...
The EVERSE project aims to collect, enhance and curate training resources aligned with domain-specific practices, create a long-term training activity supported by community services and platforms and establish a framework for recognizing Trainers and RSEs.
This contribution will describe how EVERSE plans to collect and provide training, guidance and education to researchers, software...
Since their discovery in 2015, gravitational waves have become a hot topic in physics research.
Gravitational-wave data produced by the LVK Collaboration, formed by the LIGO, Virgo and KAGRA collaborations, become fully public after a grace period; combined with the relative simplicity of the data themselves (one time series of the main signal channel per each interferometer, plus some simple...
The HSF Training group has built a fruitful learning environment within the high-energy and nuclear physics community through the organization of numerous training events. This talk will share practical insights gained from years of experience in planning and executing training events. We have learnt that organizing effective training requires careful planning, continuous adaptation, and the...
Talk to set the scene and summarise the actions from the WLCG environmental sustainability workshop in December 2024: https://indico.cern.ch/event/1450885/timetable/
In this contribution, we will describe the efforts within the ATLAS experiment to evaluate and mitigate various aspects of the environmental impact of ATLAS computing sites, such as building awareness in the experiment community, adjusting aspects of the computing policy, and modifications of data center configurations, either in ways that take advantage of particular features of ATLAS work or...
In this talk, we will describe the studies undertaken at the University of Manchester to estimate and improve the energy efficiency of computing hardware and software used by students and researchers.
The goal of these studies is to build an understanding of the environmental impact of paticle physics research focusing on two fronts:
1) the carbon cost of the hardware uses for high power...