We'll give an overview of current activities around the GridKa Tier-1 including updates on the compute, disk and tape resources and the internal network setup. In addition there are news on the HPC systems and the Large Scale Data Facility at KIT.
News from CERN since the last HEPiX workshop. This talk gives a general update from services in the CERN IT department.
Site report from INFN-T1
The International Center for Elementary Particle Physics (ICEPP) operates a WLCG Tier-2 site that provides essential computing resources for the ATLAS experiment.
This talk will present the current operational status of the site, covering hardware specifications, global network connectivity, recent operational developments, and ongoing R&D activities.
The KEK Central Computer System (KEKCC) is KEK's largest-scale computer system, providing essential services such as Grid and Cloud computing for the High Energy and Nuclear Physics community.
Following the procurement policy for large-scale computer systems, the KEKCC is replaced entirely every four years. The new system entered production in September 2024 and has successfully completed...
A large distributed computing environment with different groups of interest (funding agencies, resource providers, users), such as the WLCG, requires a score to define the amount of needed and provided computing resources.
The HEPiX Benchmarking Working Group has developed and provided benchmarks for high-energy physics (HEP) for years.
In addition to the classic CPU benchmark, we are...
The effective processing of experimental and simulated HEP data at scale on restrictive High Performance Computing (HPC) systems poses a known challenge to the physics communities. A significant aspect of such challenge is of technical nature. Many of the barriers have been successfully overcome or mitigated thanks to several integration projects over the last decade. Containerisation...
We built a coordinated scheduling layer across our local cluster and grid resources. Using glideins, idle grid capacity is elastically federated into our local HTCondor pool, while busy periods trigger proactive glidein removal to promptly return resources—achieving cross-domain elastic scale-out/in. On the local side, we introduced performance-aware labeling: nodes are tiered...
The BESIII experiment has been operating since 2009 and received several upgrades, to study the τ -charm physics utilizing the BEPCII accelerator. The offfine software system is the fundamental tool for physics analysis. It is developed to process the raw data from BESIII detector and Monte Carlo data from simulation,produce the reconstructed data which contains various physics information of...
AUDITOR is a flexible framework designed to address the challenges of managing large volumes of accounting data across diverse computing environments, including institute clusters, Grid sites, and Kubernetes clusters. Due to its modular design, it is possible to gather job information from various batch systems. AUDITOR effectively stores this data in a database for easy access and processing....
The Technology Watch Working Group, established in 2018 to take a close look at the evolution of the technology relevant to HEP computing, has resumed its activities after a long pause. In this report, we provide an overview of the hardware technology landscape and some recent developments, highlighting the impact on the HEP computing community.
Achieving higher precision of particle physics observables from large part depends on ever-increasing amount of recorded collisions. Together with improved sensitivity and increasing granularity of detectors, which leads into online processing of O(TBit/s) data with output as O(GBit/s). Such a conditions are leading into strict computing requirements for current and future HEP...
During the operation and scientific research of large-scale scientific facilities such as synchrotron light sources (BSRF, HEPS) and spallation neutron sources (CSNS), the massive, complex, and heterogeneous data are generated continuously. Enabling systematic management and efficient use of facility data; experimental data (diffraction, scattering, imaging, spectroscopy, etc.); simulation...
INFN-CNAF is the National Center of INFN (National Institute for Nuclear Physics) dedicated to research and development in the field of information technologies applied to subnuclear, nuclear and astro particle physics. CNAF hosts the largest INFN data center, which also includes a WLCG Tier1 site.
For more than 15 years, the Grid Enabled Mass Storage System (GEMSS), an in-house solution,...
2025 is widely recognized as the Year of the AI Agent. Large language models have moved beyond conversational interfaces to become callable tools that boost productivity—evident in the rapid adoption of systems like Manus, Claude-Code, and Cursor. AI Agent technologies are also increasingly being applied in scientific research to assist in data analysis and literature exploration, as...
This talk provides a quick overview of Echo, the largest disk storage cluster operated at STFC's Scientific Computing Department. Echo is the storage cluster supporting the UK's WLCG Tier 1.
Echo has been running continuously for over eight years and has scaled to 137PiB.
We will share key lessons learned from managing a cluster of this scale, including best practices, operational...
CERN, the birthplace of the World Wide Web, continues to evolve its web infrastructure to follow technology evolution and meet users’ needs. This talk will provide an overview of the web landscape at CERN, from web governance to web hosting services, including recent and ongoing changes. We will look at the change of the Content Management System, and the transition from Drupal to a new...
The Data Stores section of the CERN IT Database and Analytics group provides foundational services to store, process, and analyze scientific and operational data across the CERN community. These services currently rely on technologies such as Hadoop, OpenSearch, and NetApp. As data volumes and usage patterns continue to increase, the team is actively evolving the service portfolio to meet new...
Computing projects rely on a wide range of resources and services: databases to store data, containers to run simulations, websites to publish results, and the infrastructure needed to test and deploy software. High Energy Physics is no exception, and access to these resources must be regulated — for instance, only members of a given experiment may use a specific database cluster, or only...
Ink (Interactive Analysis Workbench) is an self-developed software tool by the Computing Center of IHPE It provides seamless access to cluster computing and storage resources through flexible API interfaces. Since its initial release and deployment on IHEP computing platform in March this year, Ink has been well-received by its users.
This report will focus on the introduction of several...
In the research and education environment, cybersecurity threats are significant and growing. We must collaborate as a community protect our environment.
Effective protection requires the use of detailed, timely and accurate threat intelligence alongside fine-grained monitoring.
We illustrate the current capabilities of the SOC. Covering how we collect, enrich, analyse and use security relevant data.
This presentation reports on the analysis of network incidents in WLCG infrastructures utilizing data from perfSONAR. We examine a few case studies of reported incidents in the past three years, which involved routing path changes and performance impacts across site pairs, highlighting IP-level anomalies and their correlation with throughput and latency metrics. Through graph-based anomaly...
The HEPiX-IPv6 working group will present the status report. The current still ongoing GGus ticket campaign of the dual-stack protocol deployment at the wlcg tier-1/2 Worker node farms will be presented.
Even while the major data exchange of LHCOPN is over IPv6, some of the still remaining IPv4 flows will be shown and examples how to mitigate them.
The partial remove at end of September 2025...
This presentation aims to give an update on the global security landscape from the past months. The global political situation has introduced a novel challenge for security teams everywhere. What's more, the worrying trend of data leaks, password dumps, ransomware attacks and new security vulnerabilities does not seem to slow down.
We present some interesting cases that CERN and the wider...
China Environment for Network Innovations (CENI) is China's first national major science and technology infrastructure in the field of communications and information. It is an open, user-friendly, and sustainable large-scale general test facility, which can provide a simple, efficient, and low-cost test and verification environment for researching the innovative architecture of future...
We will present the latest updates in WLCG site network monitoring through ps-dash, a web-based dashboard designed to visualize and analyze network performance data collected by perfSONAR. The tool presents alarms derived from alerting data measurements to help identify site-specific network issues. It reports on routing anomalies, packet loss, variations in bandwidth, and...
As part of the SwissAI initiative, CSCS resources were used to train APERTUS, an open-source multilingual language model whose entire development process including its architecture, model weights, training data, and recipes is openly accessible and fully documented.
In this presentation, I will describe the new services that CSCS is developing to deploy this model for inference using...
China’s High-Energy Photon Source (HEPS)—the country’s first national high-energy synchrotron radiation light source—is currently in the design and construction phase. The HEPS Computing Center serves as the core provider of high-performance computing (HPC) resources, data resources, and technical services for HEPS scientific experiments. The overarching mission of the HEPS scientific...
Managing Apple devices at scale in a large organization presents unique challenges and opportunities. Our environment includes thousands of macOS and iOS devices used by a diverse workforce with varied needs. To support this, we have developed a management strategy built around automation, security, and user experience.
This session will explore how we deploy and manage Macs and iOS devices...
Track reconstruction is one of the most important and challenging tasks in the offline data processing of collider experiments. The BESIII and Super Tau-Charm Facility (STCF) is the current and a next-generation electron-positron collider running in the tau-charm energy region proposed in China, where conventional track reconstruction methods face challenges from the higher background...
The growing complexity and scale of modern scientific computing infrastructures, such as the Port d’Informació Científica (PIC), a Tier-1 center within the Worldwide LHC Computing Grid (WLCG), require continuous optimization to maintain performance, reliability, and energy efficiency. Artificial Intelligence (AI) and Machine Learning (ML) techniques provide powerful means to tackle these...
The data processing and analyzing is one of the main challenges at HEP experiments. To accelerate the physics analysis and drive new physics discovery, the rapidly developing Large Language Model (LLM) is the most promising approach, it have demonstrated astonishing capabilities in recognition and generation of text while most parts of physics analysis can be benefitted. In this talk we will...