Sep 2 – 9, 2007
Victoria, Canada
Europe/Zurich timezone
Please book accomodation as soon as possible.

Grid data storage on widely distributed worker nodes using Scalla and SRM

Sep 6, 2007, 5:30 PM
Lecture (Victoria, Canada)


Victoria, Canada

oral presentation Distributed data analysis and information management Distributed data analysis and information management


Mr Pavel Jakl (Nuclear Physics Institute, Academy of Sciences of the Czech Republic)


Facing the reality of storage economics, NP experiments such as RHIC/STAR have been engaged in a shift in the analysis model, and now heavily rely on using cheap disks attached to processing nodes, as such a model is extremely beneficial over expensive centralized storage. Additionally, exploiting storage aggregates with enhanced distributed computing capabilities such as dynamic space allocation (lifetime of spaces), file management on shared storages (lifetime of files, pinning file), storage policies or a uniform access to heterogeneous storage solutions is not an easy task. The Xrootd/Scalla system allows for storage aggregation. We will present an overview of the largest deployment of Scalla (Structured Cluster Architecture for Low Latency Access) in the world spanning over 1000 CPUs co-sharing the 350 TB Storage Elements and the experience on how to make such a model work in the RHIC/STAR standard analysis framework. We will explain the key features and approach on how to make access to mass storage (HPSS) possible in such a large deployment context. Furthermore, we will give an overview of a fully "gridified" solution using the plug-and-play features of Scalla architecture, replacing standard storage access with grid middleware SRM (Storage resource manager) components designed for space management and will compare the solution with the standard Scalla approach in use in STAR for the past 2 years. Integration details, future plans and status of development will be explained in the area of best transfer strategy between multiple-choice data pools and best placement with respect of load balancing and interoperability with other SRM aware tools or implementations.
Submitted on behalf of Collaboration (ex, BaBar, ATLAS) STAR

Primary authors

Dr Jerome LAURET (BROOKHAVEN NATIONAL LABORATORY) Mr Pavel Jakl (Nuclear Physics Institute, Academy of Sciences of the Czech Republic)


Mr Alex SIM (Lawrence Berkeley Laboratory) Mr Alexandru ROMOSAN (Lawrence Berkeley Laboratory) Mr Andrew HANUSHEVSKY (Stanford Linear Accelerator Center) Mr Arie SHOSHANI (Lawrence Berkeley Laboratory)

Presentation materials