Help us make Indico better by taking this survey! Aidez-nous à améliorer Indico en répondant à ce sondage !

CCRC'08 planning conference call

Europe/Zurich
salle b (CERN)

salle b

CERN

Description
Mailing list: wlcg-ccrc08@cern.ch

Web archive: here

To join the call, do one of the following:

  • Dial +41227676000 and enter access code 0121632, or
  • To have the system call you, click here
(Leader code is 0111659).
    • 17:00 17:05
      Minutes of previous meeting and action items 5m
      Notes of the CCRC'08 planning conference call of 12.11.07 Present: WLCG Service Coordination: J.Shiers (chair JS) H.Renshall (notes HR) GSSD: F.Donno Tier0: M.Coelhos (MC) WLCG/LHC-Ops link: absent MB: A.Aimar (AA) ATLAS: absent CMS: absent ALICE: L.Betev (LB) LHCb: absent ASGC: absent BNL: M.Ernst (MB) CNAF: L.dell'Agnello (LA) FNAL: absent FZK: absent IN2P3: F.Hernandez (FH) NDGF: absent NL-T1: absent PIC: G.Merino (GM) RAL: A.Sansum (AS) TRIUMF: absent Notes: JS set the scene of this and the next phone conference as primarily preparation for the December Face 2 Face meeting before moving on to the agenda points. As regards scaling factors for the February challenge we are still missing the ATLAS analysis. JS will raise this to the MB. For the site resources available for February HR said he was still waiting details from ASGC, CNAF, IN2P3, NDGF, NL-T1 and RAL. ASGC should report soon. For RAL AS said there would only be small changes, for CNAF LA said they were tendering now but would not get extra hardware for February while FH said that IN2P3 are reviewing their pledges but expect no major changes for February. HR then presented 4 draft spreadsheets (attached to the agenda) showing the per Tier 1 per experiment data rates and integrated volumes to tape and disk for the February run. This is based on some simplistic assumptions that now need refining: Although the primary purpose is to test functionality we should attempt to reach and maintain the full 2008 p-p rates for data acquisition, first pass reconstruction and data export to all Tier-1. An LHC efficiency of 55% is assumed in a model where data processing and export should aim to run at machine-on rates and use machine-off time for any catching up. The integrated volumes include this efficiency factor. HR pointed out that though the extra disk and tape storage requirements were 10 to 20% of the installed capacity at most Tier-1 the cpu required to perform full 2nd pass reconstruction would not be available at most Tier-1 and that for the moment he had not included extra storage for this activity in the tables. The assumptions used for ATLAS data are that all raw goes to tape via a disk buffer (type T1D0) and that all ESD goes to permanent disk (type T0D1). Probably all the data can be scratched as it will be pure simulated events. For CMS it is assumed that all the FEVT data exported from CERN goes to tape via a disk buffer. A fraction of this will be cosmics to be kept. For ALICE the raw data goes to tape via a disk buffer and the reconstructed data will go to permanent disk. They estimate some 30% will be cosmics to be kept and LB made the point that storage for this should already be included in the 1Q2008 site resources. LHCb had given their own detailed breakdown. JS then asked if any more information should be included under this heading. The general concensus was not. FH thought that the number of jobs per day per site and experiment would be good. GM would like to see the expected Tier-1 to Tier-1 and Tier-2 data rates and HR agreed to send round a pointer to the latest version of the 'megatable' of such rates. The next item was blocks of functional tests where MC suggested blocks of a few hours of sending the combined mock and cosmics data then demonstrating removal of the mock data. He volunteered to coordinate this. HR added that the end points and/or directory structures associated with this should be published as well as site recipes for cleaning tapes. The next item was the state of SRMv2.2 preparations where there was good input from LHCb (attached to the agenda). FD announced that ATLAS had sent her a quite detailed plan. They have finished updating their framework to use SRMv2 but not yet started testing. K.Boss is organising a meeting to coordinate these activities. LB said that as long as FTS works properly ALICE had no other requirements on SRMv2. FD said we would like to stress the use of space tokens and that if sites had switched them on they must be specified to get the correctly managed space. LB asked if this only applied to dcache and why the T0D1 and T1D0 types may need to be split under different space tokens. FD replied that space tokens organise pools of disks and tapes of any type and that different data may want to go to the same type of storage but be on separate physical storage. She agreed to sit down with LB to look at the ALICE requirements. FD reminded us that it was agreed at the GSSD to use the ccrc'08 phone conferences to coordinate the SRMv1 to v2 transitions. Next Monday is an LHCC LCG review and the week after is the service reliability workshop so JS will choose another day for the next phone conference (it is Wednesday 21st at 17.00). JS finally asked if we are preparing well enough for the December F2F and ME felt uncomfortable without more experiment participation. All agreed this was important and JS will encourage them to participate at the next meeting. Resulting Action Items: HR to send round a pointer to the latest version of the 'megatable' of Tier-1 to Tier-1 and Tier-2 rates. FD and LB to look at the ALICE mock and cosmics data separation requirements. JS to encourage experiment participation in these meetings.
    • 17:05 17:10
      Conclusion on Scaling Factors for February challenge 5m

      Scaling factors now also for ATLAS

      Slides
    • 17:10 17:20
      SRM v2.2 preparation - timetable for experiments to use SRM v2.2 features 10m
      Details from LHCb:

      • to move our current tools to use SRM 2.2 in a mixed environment we estimate we need 2-4 weeks (including testing) to do that. This would not include any new functionality at this stage. This assumes the latest lcg utils (and the necessary gLite s/w) is deployed at the sites. Need to clarify on the use of srm-get-metadata within the centralised stager agent. On the WN we will need lcg-cp & lcg-gt.
      • to make full use of the new functionality with SRM 2.2 will not come until we release DIRAC3 which is expected at the end of December/beginning of January.

      If experiments are to use SRM v2.2 (as SRM v2.2) in February, then sites need to have the necessary storage configured early January. This matches the timetable for the experiments to have made the necessary modifications (end December / early January). Thus, in December at the F2F we need to clarify exactly what is needed by VO and by site so that it can be prepared for January 2008.

      agenda
      Slides
    • 17:20 17:30
      CDR challenge in December 10m
      From discussion it’s not obvious that all experiments know how to write mock plus real data in a way that mock data can be dropped by the end of the exercise. What measures should be taken to ensure this will work? Should we have a small test right now in December? 24h CDR run at nominal rate with a mix of real and mock data? The important step is not the test but obviously the preparation that needs to go into it. Miguel
    • 17:30 17:45
      Actions in preparation for December F2F meeting 15m
      • Data volumes and rates - information required by sites to prepare
        See questions from Tier0, for example:

        • how much tape space is required for the test in February ?
        • how much of the space will be kept after the test (i.e. can we delete it all) ?
        • how much tape space is required for the test in April ?
        • how much of the space will be kept after the test ?

        Are there any indications of the data rate required

        • to tape (i.e. writing) ?
        • from tape (i.e. processing of data which has not cached at the disk layer) ?
      • Preparation of walk-throughs of services required for each 15m
        Detail of experiment, WLCG services involved in each block of the challenge with appropriate "scaling factors":

        Possible scenario: blocks of functional tests, Try to reach 2008 scale for tests at…

        1. CERN: data recording, processing, CAF, data export
        2. Tier-1’s: data handling (import, mass-storage, export), processing, analysis
        3. Tier-2’s: Data Analysis, Monte Carlo, data import and export

        This is perhaps just combining the existing presentations on the plans, e.g. from the November MB F2F meeting, with the Critical Services that will be used in each step (and the appropriate "scaling factors")

      • Monitoring the Challenge
        Do we understand what tools are required / will be used to monitor the challenge / progress?

        Are all of these tools in place with the necessary functionality?

        See Wednesday session at WLCG Service Reliability workshop

    • 17:45 17:55
      Agenda for December pre-GDB 10m
      agenda
    • 17:55 18:05