Speaker
Dr
Douglas Benjamin
(Duke University)
Description
The CDF experiment at Fermilab produces Monte Carlo data files using computing
resources on both the Open
Science Grid (OSG) and LHC Computing Grid (LCG) grids. This data produced must be
brought back to Fermilab
for archival storage. In the past CDF produced Monte Carlo data on dedicated
computer farms through out
the world. The data files were copied directly from the worker nodes to a few file
servers located at FNAL
using rcp and Kerberos authentication. As the experiment has moved from dedicated
resources to shared
resources on the grid, this technique has proven to be problematic. We plan to show
how changing our
data delivery model to one of concentrating the data in a few sites located in
different regions
throughout the world and shipping it back in a coordinated method back to FNAL
improved our use of
opportunistic CPU cycles. The details of the evaluation process and implemented
solution including
metrics demonstrating our performance will also be presented.
Submitted on behalf of Collaboration (ex, BaBar, ATLAS) | CDF collaboration |
---|
Author
Dr
Douglas Benjamin
(Duke University)