GSSD face-to-face meeting

Europe/Zurich
IT Auditorium (CERN)

IT Auditorium

CERN

Flavia Donno (CERN)
Description
Discussions about SRM 2.2 deployment, Grid Storage Services, Data Management clients. VRVS available: Mountain virtual room

Minutes of the GGSD meeting - 12 April 2007

Status of SRM 2.2 (Flavia)

A status update of the SRM 2.2 implementations was given (please, check slides). Experiments are invited to let Flavia know about possible use cases to test. All implementations pass basic tests. Still working on some details with use-cases. Copy tests need attention for the endpoints where copy is implemented. 4 implementations provide deployable releases: dCache (v1.8Beta), DPM (v1.6.4), StoRM (v1.3.15), BeStMan (v2.2.0.0)
  • It was asked to include S2 based tests for lcg-utils and GFAL. This will be done.
  • It was asked to document use-case tests. This is work in progress.
  • BNL asked what is needed to "participate" to the S2 test. As a minimum you need to:
    1. Subscribe to the srm-tester mailing list: srmtester@lbl.gov
    2. Register the following DN: /C=IT/O=INFN/OU=Personal Certificate/L=Pisa/CN=Flavia Donno/Email=flavia.donno@pi.infn.it
    3. Send the full endpoint (srm://<full hostname>:<port>/srm/managerv2?SFN=<dteam path> to srmtester@lbl.gov

    lcg-utils status (Jean-Philippe)

    Please, check slides. lcg-utils v1.5.1-1 and GFAL 1.9.0-2 available in production. SRM v1 is the default and SRM v2 used only if no SRM v1 published or space token is specified. This behavior is not configurable.
  • No ReserveSpace functionality offered through lcg-util/GFAL.
  • BringOnline not possible with lcg-utils but it is with GFAL. If needed BringOnline can be made available via lcg-utils.
  • Pinning file not available since it is not supported by all implementations. In particular, the CASTOR garbage collector can be configured per VO, depending on the needs. VOs need to contact the CASTOR support team.
  • Releasing files is possible with both lcg-utils and GFAL.
  • lcg-cp now takes also SURL as destination (only if source is a local file or is a TURL with GSIFTP as transport protocol) In this case no catalogue operations are performed.
  • At the moment there is no way to address GFAL to use a specific file access protocol. If needed, this can be implemented.

    fts status (Paolo T.)

    FTS 2.0 offers support for SRM 2.2. It is under test since December 2006 on pilot FTS testbed. Deployment at CERN foreseen for April 2007 and at T1s one month after. For features introduced and plans, please check slides. It was observed that standard gridftp/SRM point to point tests should be made part of SAM in order to monitor overall performance of the network.

    SRM v1 to v2 migration strategy (Stephen)

    Please, check slides for a full introduction and details on the problem. Recommendations:
  • Sites should use alias as much as possible and keep the names of the SEs stable. The GSSD will try to push forward this recommendations.
  • On sites where the SRM v1 and v2 endpoints are disjoint (each with its own name service), it may be necessary for the VOs to decide which v1 entries need to be made available under v2. This implies making a physical copy of the data concerned if the v1 entries cannot simply be added to the v2 name service. Examples: CASTOR v1-->v2 at CERN, dCache-->CASTOR at RAL.
  • On sites where the SRM v1 and v2 endpoints are front-ends for the same name service, data written by v1 remain accessible by v2 and vice versa. Data written by v1 can only have storage class hints expressed through the name space: the SE admin can map a particular directory to a particular storage class. To ease the transition from v1 to v2 VOs could design name spaces such that the different storage class instances only appear high up the hierarchy, so that there will be not so many. Here is an example of how the namespace should be organized:
      Let's assume the following:
    
        RAW         --> T1D0
        DST         --> T1D0, but different tape set
        ESD_master  --> T1D1
        ESD_replica --> T0D1
        AOD         --> T0D1
    
        The namespace SHOULD NOT be organized as follows:
    
        /my_exp/data/year/run/reco_pass/stream/RAW
        /my_exp/data/year/run/reco_pass/stream/DST
        /my_exp/data/year/run/reco_pass/stream/ESD_master
        /my_exp/data/year/run/reco_pass/stream/ESD_replica
        /my_exp/data/year/run/reco_pass/stream/AOD
    
        Instead it SHOULD be organized as follows:
    
        /my_exp/data/RAW/year/run/reco_pass/stream
        /my_exp/data/DST/year/run/reco_pass/stream
        /my_exp/data/ESD_master/year/run/reco_pass/stream
        /my_exp/data/ESD_replica/year/run/reco_pass/stream
        /my_exp/data/AOD/year/run/reco_pass/stream 
    
        This will easy the transition from v2 back to v1 if necessary.
    
    
  • It is not advisable to go back to SRM v1 once the full functionality of SRM v2 has been used by an experiment. The goal is to test v2 thoroughly and then migrate to v2 and stay with it.
  • All implementations offer new features together with the SRM v2 interface. Therefore, upgrading to SRM v2 can imply more bugs to be sorted out. The experiments have agreed to collaborate to make v2 working fully.
  • We should be pay attention to situations where both SRM v1 and v2 endpoints are published. If the v2 endpoint is not visible for some reason, there is the possibility that client tools default to SRM v1, creating confusion and potential problems.

    Storage Issues from ATLAS (Miguel)

    Please, check slides.
  • Many scalability issues identified during transfer when reaching CASTOR LSF max num of pending jobs threshold. Existing CASTOR s/w could not cope with current ATLAS load. ATLAS pushed for earlier deployment of new CASTOR-LSF plugin to satisfy Tier-0 requirements. CASTOR problems served as a motivation for detailed low-level error analysis and to try to understand interactions between all m/w layers: FTS->SRM->GridFTP->Storage Elements. Network tests (GridFTP only) are now ongoing to BNL. ATLAS suggested WLCG should take over this responsibility: to constantly monitor state of the network.
  • Network tests (GridFTP only) are now ongoing to BNL. ATLAS suggested WLCG should take over this responsibility: to constantly monitor state of the network.
  • Problems with FTS doing the SRM handling while transferring. FTS 2.0 has a DB schema change that allows for decoupling SRM handling from file transfers, which is a high-priority item on the developer to-do list.
  • Noticed interference between servers and monitoring tools (CASTOR and GridView)
  • dCache problem noticed in Lyon when disk full (dCache was not counting big blocks allocated by the system for small control files). The problem has been fixed.
  • dCache disk full led to many different error messages. This needs to be addressed.
  • Similar dCache problem occurred at BNL with disk full because the log rotation algorithm was not working properly. This reduces the Tier0 throughput as CASTOR transfer slots are blocked to retry the failed operations.

    Storage Issues from LHCb (Nick)

    Please, check slides.
  • srm_get (SRM v1) not working properly for dCache since it returns TURL of a non-staged file. Problem acknowledged and fix being worked on.
  • Few authorization problems noticed at SARA but this could be a general issue (please, check slides for details).
  • Question: Should gsidcap be allowed to be used remotely ? Answer: The experiments can decide what to do. It is not recommendable to use gsidcap for remote transfers.
  • LHCb intends to use gsidcap for implementing pre-staging as a temporary workaround for the srm_get problem. SRM v2 offers functionality in this respect. Generic higher level tools could be used. (Check presentation by Jean-Philippe).
    Roberto Santinelli sent the following report after the meeting:
    "1 the gsidcap issue we had at NIKHEF (but this was rather a site configuration 
    problem that prevented LHCb to run reconstruction/reprocessing for many months 
    at SARA/NIKHEF, see https://twiki.cern.ch/twiki/bin/view/LHCb/DC06Activity 
    for more information
    
    2. the CASTOR2 issue at CNAF (Angelo Carbone could give 
    you a fully exhaustive description of this issue although 
    https://twiki.cern.ch/twiki/bin/view/LHCb/PreviousDC06Problems also gives 
    an overview of the problem and its solution). 
    
    3. At a certain point we also suffered about a slowness of transfers 
    to/from  CERN versus T2 centers and this was also due to a misconfiguration on the 
    HTAR routing. (see Olof's findings)
    https://twiki.cern.ch/twiki/pub/LCG/SCActionList/tier-2_transfers.doc)
    
    4. Very recently, but this has to be confirmed and further investigated, 
    we are experiencing a problem with dcache centers. It looks like, if under 
    heavy load, dcache storage system registers the file in the database 
    (read: in the dcache name-space /pnfs/...) but the replica is not 
    physically copied on the disk pools. 
    This problems seems to happen at transfer time (using lcg-utils) and this 
    is really worrying (if confirmed) because a large amount of data might be 
    lost (having sometimes only one single replica).
    
    Of course there is a plethora of problems related to the high instability 
    of the SRM endpoints that brought  a very low efficiency on our transfers 
    (as reported too many times at several bodies) and also a very bad 
    connection between application area (read:Root) and the middleware it self 
    (read: SRM) on the format of the turl returned by srm_get that is not 
    understandable by the application (see root:// rfio:// protocol saga at 
    CERN Castor2, the gsidcap protocol that wants prepended dcap:  in order to 
    be understood by ROOT, format of the rfio protocol returned by Castor1 
    at PIC)........but on that last respect Philippe is the most adequate 
    person for a full exhaustive  summary."
    
    

    Rollout plan (All)

  • SRM v1 will be supported till required (this is true for CASTOR, dCache and DPM).
  • The following test endpoints will be added to the S2 test suite:
       dCache:
       ------
       FNAL and DESY already there
       BNL
       FZK
       IN2P3
       This will happen within the end of April. All T1 endpoints will be available
       on dedicated test hardware. No existing data will be made available.
       Newly created data can be scratched. Experiments have agreed on this.
    
  • UK sites are interested in testing new versions of DPM and StoRM. Greig will send new endpoints to Flavia as soon as possible to be included in the S2 test suite.
  • New StoRM endpoint available in production at CNAF as of now. This will be added to the test endpoints.
  • All available endpoints will be S2 [stress-]tested at least till 15th of May.
  • Endpoints will be added to FTS pilot testbed for FTS tests at the same time.
  • All endpoints will be published by the test UI (lxb0728.cern.ch) that is available to the experiments to test lcg-utils/GFAL.
  • FTS 2.0 client will be also installed on test UI.
  • CASTOR2 with SRM 2.2 will be installed at CNAF in the next 2 coming weeks. This endpoint will also be S2 tested and made available for FTS/lcg-utils/GFAL testing on both the test UI and the FTS pilot testbed.
  • (This point has been checked and agreed with the pre-production testbed coordinators.) All available endpoints can be included in the pre-production testbed together with all available clients, even if uncertified. The experiments can test their software against the new versions of the middleware on the pre-production testbed. This inclusion in the pre-production testbed can start during the first 2 weeks of May if steps 1-8 above are concluded.
  • Tier-2s can use production resources for these tests. Experiments have to be careful and aware that newly produced data can be lost.
  • On the pre-production testbed a dedicated LFC is available for tests. Also, a dedicated BDII will be used.
  • Sites that install StoRM and join the pre-production testbed should also make available some CEs with GPFS directly mounted. Such CEs and SEs should be published properly in the pre-production BDII (CESEBind in particular).
  • This plan will be carried out and overviewed by the SRM effort at CERN.

    Storage Monitoring Tools (Mirco C.)

    Mirco reported on the analysis of the tools studied so far for retrieving storage information that can be useful for experiments. This info can be made available through the tools that are currently used for monitoring (dashboards, GridView, SAM, etc.).
  • It was suggested to make a summary of the information commonly available
  • The report should be circulated to the VOs for comments
  • VOs should give input on what they need
  • A first wrapper should be made available and possibly integrated with the existing monitoring tools.
  • Investigation on what is available but not at the moment retrievable from the various implementations should also be carried out.

    Next GSSD meetings

    A tentative schedule is published here.
  • There are minutes attached to this event. Show them.
      • 09:00 09:30
        Status of SRM 2.2 implementations 30m
        Report on the status of the implementations and the releases available
        Speaker: Flavia Donno
        Slides
      • 10:00 10:15
        lcg-utils status 15m
        A status report on the available version of the lcg-utils middleware
        Speaker: Jean-Philippe Baud (CERN)
        Slides
      • 10:30 10:50
        FTS new features and status 20m
        A report on the new FTS features made available for SRM 2.2. Status of the client.
        Speaker: Paolo Tedesco (Unknown)
        Slides
      • 11:30 12:00
        SRM v1-v2 migration and GLUE schema 30m
        How to migrate from v1 to v2. Strategic plan. Few notes on the new GLUE schema
        Speaker: Dr Stephen Burke (RUTHERFORD APPLETON LABORATORY)
        Slides
      • 14:00 14:30
        Storage Issues from ATLAS 30m
        Speaker: Mr Miguel Branco (CERN)
        Slides
      • 14:35 14:55
        Recent LHCb experience with dCache 20m
        Problems observed by LHCb using dCache 1.7
        Speaker: Nicholas Brook (University of Bristol)
        Slides
      • 15:00 15:20
        Rollout plannning for SRM Storage Services 20m
        Discussions with the sites that have declared to be ready to deploy the new versions of the Storage Services available. Planning.
        Speaker: All
      • 15:30 16:00
        Storage Monitoring Tools 30m
        What is needed in terms of storage ? What are the ongoing efforts ?
        Speaker: Mirco Ciriello (Unknown)
        Slides