2-9 September 2007
Victoria, Canada
Europe/Zurich timezone
Please book accomodation as soon as possible.

ATLAS Analysis Model

3 Sep 2007, 14:00
20m
Lecture (Victoria, Canada)

Lecture

Victoria, Canada

oral presentation Software components, tools and databases Software components, tools and databases

Speaker

Dr Amir Farbin (European Organization for Nuclear Research (CERN))

Description

As we near the collection of the first data from the Large Hadron Collider, the ATLAS collaboration is preparing the software and computing infrastructure to allow quick analysis of the first data and support of the long-term steady-state ATLAS physics program. As part of this effort considerable attention has been payed to the "Analysis Model", a vision of the interplay of the software design, computing constraints, and various physics requirements. An important input to this activity has been the experience of Tevatron and B-Factory experiments, one topic which was explored discussed in the ATLAS October 2006 Analysis Model workshop. Recently, much of the Analysis Model has focused on ensuring the ATLAS software framework supports the required manipulations of event data; the event data design and content is consistent with foreseen calibration and physics analysis tasks; the event data is optimized in size, access speed, and is accessible both inside and outside the software framework; and that the analysis software may be developed collaboratively.

Summary

In this talk, I will overview the lessons learned from
various High Energy Experiments and the resulting actions within the ATLAS
software.

Submitted on behalf of Collaboration (ex, BaBar, ATLAS) ATLAS Offline Computing

Primary author

Dr Amir Farbin (European Organization for Nuclear Research (CERN))

Co-authors

Dr Ketevi Assamagan (Brookhaven National Lab) Dr Kyle Cranmer (Brookhaven National Lab)

Presentation Materials