Speaker
Description
We are developing a new Machine Learning (ML) service at CERN to support the use of Large Language Models (LLMs) and Agentic AI. Our goal is to provide a reliable and secure foundation for researchers and developers. In this presentation, we will describe the architecture and plans for this new service. It will include several key components: an LLM Proxy that works with OpenAI-compatible APIs, a Model Catalog for on-premises and cloud-based models, and an Agent Hosting and Orchestration platform. We will show some initial applications that will be used with this service.
These examples will demonstrate the potential benefits of the service, such as improved productivity and faster research workflows. We will also discuss our future plans, including a shared AI Agent evaluation service and incorporating new use cases from the CERN community.
Our aim is to create a scalable and sustainable ecosystem for LLMs and Agentic AI at CERN, balancing innovation with security and operational needs.
| Desired slot length | 20 minutes |
|---|---|
| Speaker release | Yes |