Speakers
Dr
Gabriele Compostella
(CNAF INFN)Dr
Manoj Kumar Jha
(INFN Bologna)
Description
Being a large international collaboration established well before the
full development of the Grid as the main computing tool for High
Energy Physics, CDF has recently changed and improved its computing model, decentralizing some parts of it in order to be able to exploit the rising number of distributed resources available nowadays.
Despite those efforts, while the large majority of CDF Monte Carlo
production has moved to the Grid, data processing is still mainly
performed in dedicated farms hosted at FNAL, requiring a centralized
management of data and Monte Carlo samples needed for physics
analysis.
This rises the question on how to manage the transfer of produced
Monte Carlo samples from remote Grid sites to FNAL in an efficient
way; up to now CDF has relied on a non scalable centralized solution
based on dedicated data servers accessed through rcp protocol, which
has proven to be unsatisfactory.
A new data transfer model has been designed that uses SRMs as local
caches for remote Monte Carlo production sites, interfaces them with
SAM, the experiment data catalog, and finally realizes the
file movement exploiting the features provided by the data catalog
transfer layer.
We describe here the model and its integration within the current CDF
computing architecture. We discuss the performance gain and the
benefits of the new framework in comparison with the old approach.
Presentation type (oral | poster) | oral |
---|
Authors
Dr
Donatella Lucchesi
(University and INFN Padova)
Dr
Douglas Benjamin
(Duke University)
Dr
Gabriele Compostella
(CNAF INFN)
Dr
Manoj Kumar Jha
(INFN Bologna)
Dr
Simone Pagan Griso
(University and INFN Padova)