LLM/AI tool use in particle physics

Europe/Zurich
Matthew Feickert (University of Wisconsin (US)), Robert Currier Tuck (Princeton University (US))
Description

Talks focusing on personal experiences using and developing tool applications to assist in their physics.

YouTube:  https://youtu.be/nbKjafrmReY

    • 17:00 17:05
      Introduction 5m
      Speaker: Matthew Feickert (University of Wisconsin (US))
    • 17:05 17:25
      Vibe Plotting - Using LLM's to make physics analysis plots 20m
      Speaker: Gordon Watts (University of Washington (US))
    • 17:25 17:45
      Building the HEP Agent Stack, MCP Servers, CLI Tools, and Skill-Based Workflows for Scientific Analysis 20m

      This talk presents a practical agentic AI stack for high-energy physics through three projects: root-mcp, cerngitlab-mcp, and inspirehep-mcp. These tools expose ROOT data analysis, CERN GitLab software discovery, and INSPIRE-HEP literature workflows through MCP servers, command-line tools, and lightweight SKILL.md files. The talk focuses on the shared-backend design behind these tools, the tradeoffs between MCP and token-efficient SKILL + CLI workflows, and how this approach can support reusable, reproducible HEP research infrastructure.

      Speaker: Mohamed Elashri (University of Cincinnati)
    • 17:45 18:05
      MITRA: An AI assistant for Knowledge Retrieval in Physics collaborations 20m

      We present MITRA, a Retrieval-Augmented Generation (RAG) system developed to assist particle physicists in navigating collaboration documentation. The system facilitates the extraction of relevant information from analyses and shift documentation and provides answers to user queries with direct citations to the original sources. While the current implementation uses CMS analysis documentation, the underlying workflow is designed to be experiment-agnostic. The pipeline: including document ingestion, embedding generation, and retrieval, is modular, allowing for adaptation to the documentation of other collaborations.

      To meet the privacy requirements of HEP experiments, MITRA operates entirely on local collaboration infrastructure using self-hosted, language models. This ensures that internal documents such as unpublished results remain within approved servers and are not exposed to external APIs. This also reduces cost, allowing the system to scale across the user size of HEP collaborations and across time. We describe the motivation for building MITRA, performance of the retrieval pipeline, and outline ongoing work toward more complex, multi-step reasoning workflows.

      https://arxiv.org/abs/2603.09800

      Speaker: Abhishikth Mallampalli (University of Wisconsin Madison (US))