Rapolas Kaselis (Vilnius University (LT))
The goal for CMS computing is to maximise the throughput of simulated event generation while also processing the real data events as quickly and reliably as possible. To maintain this achievement as the quantity of events increases, since the beginning of 2011 CMS computing has migrated at the Tier 1 level from its old production framework, ProdAgent, to a new one, WMAgent. The WMAgent framework offers improved processing efficiency and increased resource usage as well as a reduction in manpower. In addition to the challenges encountered during the design of the WMAgent framework, several operational issues have arisen during its commissioning. The largest operational challenges were in the usage and monitoring of resources, mainly a result of a change in the way work is allocated. Instead of work being assigned to operators, all work is centrally injected and managed in the Request Manager system and the task of the operators has changed from running individual workflows to monitoring the global workload. In this report we present how we tackled some of the operational challenges, and how we benefitted from the lessons learned in the commissioning of the WMAgent framework at the Tier 2 level in late 2011. As case studies, we will show how the WMAgent system performed during some of the large data reprocessing and Monte Carlo simulation campaigns.
Ajit Kumar Mohapatra (University of Wisconsin (US)) Dr Dave Evans (Fermi National Accelerator Lab. (US)) Markus Klute (Massachusettes Institute of Technology) Mr Matthew Norman (University of California at San Diego) Rapolas Kaselis (Vilnius University (LT)) Simon Metson (University of Bristol (GB)) Stephen Foulkes (Fermi National Accelerator Lab. (Fermilab)) Valentina Dutta (Massachusetts Inst. of Technology (US)) Vincenzo Spinoso (Universita e INFN (IT)) Zdenek Maxa (California Institute of Technology (US))