Conveners
Computer Facilities, Production Grids and Networking
- Daniele Bonacorsi (University of Bologna)
Computer Facilities, Production Grids and Networking
- Andreas Heiss (Forschungszentrum Karlsruhe GmbH)
Computer Facilities, Production Grids and Networking
- Maria Girone (CERN)
Computer Facilities, Production Grids and Networking
- Daniele Bonacorsi (University of Bologna)
-
Dr Domenico Vicinanza (DANTE)21/05/2012, 13:30Computer Facilities, Production Grids and Networking (track 4)ParallelThe Large Hadron Collider (LHC) is currently running at CERN in Geneva, Switzerland. Physicists are using LHC to recreate the conditions just after the Big Bang, by colliding two beams of particles and heavy ions head-on at very high energy. The project is expected to generate 27 TB of raw data per day, plus 10 TB of "event summary data". This data is sent out from CERN to eleven Tier 1...Go to contribution page
-
Edoardo Martelli (CERN)21/05/2012, 13:55Computer Facilities, Production Grids and Networking (track 4)ParallelThe much-heralded exhaustion of the IPv4 networking address space has finally started. While many of the research and education networks have been ready and poised for years to carry IPv6 traffic, there is a well-known lack of academic institutes using the new protocols. One reason for this is an obvious absence of pressure due to the extensive use of NAT or that most currently still have...Go to contribution page
-
Mr Andrey Bobyshev (FERMILAB)21/05/2012, 14:20Computer Facilities, Production Grids and Networking (track 4)ParallelThe LHC is entering its fourth year of production operation. Many Tier1 facilities can count up to a decade of existence when development and ramp-up efforts are included. LHC computing has always been heavily dependent on high capacity, high performance network facilities for both the LAN and WAN data movement, particularly within the Tier1 centers. As a result, the Tier1 centers tend to...Go to contribution page
-
Jan Iven (CERN), Massimo Lamanna (CERN)21/05/2012, 14:45Computer Facilities, Production Grids and Networking (track 4)ParallelLarge-volume physics data storage at CERN is based on two services, CASTOR and EOS: * CASTOR - in production for many years - now handles the Tier0 activities (including WAN data distribution), as well as all tape-backed data; * EOS - in production since 2011 - supports the fast-growing need for high-performance low-latency (i.e. diskonly) data access for user analysis. In 2011, a large...Go to contribution page
-
Rapolas Kaselis (Vilnius University (LT))21/05/2012, 15:10Computer Facilities, Production Grids and Networking (track 4)ParallelCMS experiment possesses distributed computing infrastructure and its performance heavily depends on the fast and smooth distribution of data between different CMS sites. Data must be transferred from the Tier-0 (CERN) to the Tier-1 for storing and archiving, and time and good quality are vital to avoid overflowing CERN storage buffers. At the same time, processed data has to be distributed...Go to contribution page
-
Shawn Mc Kee (University of Michigan (US))21/05/2012, 16:35Computer Facilities, Production Grids and Networking (track 4)ParallelGlobal scientific collaborations, such as ATLAS, continue to push the network requirements envelope. Data movement in this collaboration is projected to include the regular exchange of petabytes of datasets between the collection and analysis facilities in the coming years. These requirements place a high emphasis on networks functioning at peak efficiency and availability; the lack thereof...Go to contribution page
-
Niko Neufeld (CERN)21/05/2012, 17:00Computer Facilities, Production Grids and Networking (track 4)ParallelThe upgraded LHCb experiment, which is supposed to go into operation in 2018/19 will require a massive increase in its compute facilities. A new 2 MW data-centre is planned at the LHCb site. Apart from the obvious requirement of minimizing the cost, the data-centre has to tie in well with the needs of online processing, while at the same time staying open for future and offline use. We present...Go to contribution page
-
Dr Horst Gรถringer (GSI)21/05/2012, 17:25Computer Facilities, Production Grids and Networking (track 4)ParallelGSI in Darmstadt (Germany) is a center for heavy ion research. It hosts an Alice Tier2 center and is the home of the future FAIR facility. The planned data rates of the largest FAIR experiments, CBM and Panda, will be similar to those of the current LHC experiments at Cern. gStore is a hierarchical storage system with unique name space and successfully in operation since more than...Go to contribution page
-
Dmitry Ozerov (Deutsches Elektronen-Synchrotron (DE)), Martin Gasthuber (Deutsches Elektronen-Synchrotron (DE)), Patrick Fuhrmann (DESY), Yves Kemp (Deutsches Elektronen-Synchrotron (DE))21/05/2012, 17:50Computer Facilities, Production Grids and Networking (track 4)ParallelWe present results on different approaches on mounted filesystems in use or under investigation at DESY. dCache, established since long as a storage system for physics data has implemented the NFS v4.1/pNFS protocol. New performance results will be shown with the most current version of the dCache server. In addition to the native usage of the mounted filesystem in a LAN environment, the...Go to contribution page
-
Wahid Bhimji (University of Edinburgh (GB))22/05/2012, 13:30Computer Facilities, Production Grids and Networking (track 4)ParallelWe describe recent I/O testing frameworks that we have developed and applied within the UK GridPP Collaboration, the ATLAS experiment and the DPM team, for a variety of distinct purposes. These include benchmarking vendor supplied storage products, discovering scaling limits of SRM solutions, tuning of storage systems for experiment data analysis, evaluating file access protocols, and...Go to contribution page
-
Erik Mattias Wadenstein (Unknown)22/05/2012, 13:55Computer Facilities, Production Grids and Networking (track 4)ParallelDistributed storage systems are critical to the operation of the WLCG. These systems are not limited to fulfilling the long term storage requirements. They also serve data for computational analysis and other computational jobs. Distributed storage systems provide the ability to aggregate the storage and IO capacity of disks and tapes, but at the end of the day IO rate is still bound by the...Go to contribution page
-
Brian Paul Bockelman (University of Nebraska (US))22/05/2012, 14:20Computer Facilities, Production Grids and Networking (track 4)ParallelWhile the LHC data movement systems have demonstrated the ability to move data at the necessary throughput, we have identified two weaknesses: the latency for physicists to access data and the complexity of the tools involved. To address these, both ATLAS and CMS have begun to federate regional storage systems using Xrootd. Xrootd, referring to a protocol and implementation, allows us to...Go to contribution page
-
Dr Xavier Espinal Curull (Universitat Autรฒnoma de Barcelona (ES))22/05/2012, 14:45Computer Facilities, Production Grids and Networking (track 4)ParallelScientific experiments are producing huge amounts of data, and they continue increasing the size of their datasets and the total volume of data. These data are then processed by researchers belonging to large scientific collaborations, with the Large Hadron Collider being a good example. The focal point of Scientific Data Centres has shifted from coping efficiently with PetaByte scale storage...Go to contribution page
-
Jason Alexander Smith (Brookhaven National Laboratory (US))22/05/2012, 15:10Computer Facilities, Production Grids and Networking (track 4)ParallelManaging the infrastructure of a large and complex data center can be extremely difficult without taking advantage of automated services. Puppet is a seasoned, open-source tool designed for enterprise-class centralized configuration management. At the RHIC/ATLAS Computing Facility at Brookhaven National Laboratory, we have adopted Puppet as part of a suite of tools, including Git, GLPI, and...Go to contribution page
-
Jason Zurawski (Internet2)24/05/2012, 13:30Computer Facilities, Production Grids and Networking (track 4)ParallelScientific innovation continues to increase requirements for the computing and networking infrastructures of the world. Collaborative partners, instrumentation, storage, and processing facilities are often geographically and topologically separated, as is the case with LHC virtual organizations. These separations challenge the technology used to interconnect available resources,...Go to contribution page
-
Steve Barnet (University of Wisconsin Madison)24/05/2012, 13:55Computer Facilities, Production Grids and Networking (track 4)ParallelBesides the big LHC experiments a number of mid-size experiments is coming online which need to define new computing models to meet the demands on processing and storage requirements of those experiments. We present the hybrid computing model of IceCube which leverages GRID models with a more flexible direct user model as an example of a possible solution. In IceCube a central datacenter at...Go to contribution page
-
Mr Pier Paolo Ricci (INFN CNAF)24/05/2012, 14:20Computer Facilities, Production Grids and Networking (track 4)ParallelThe storage solution currently used in production at the INFN Tier-1 at CNAF, is the result of several years of case studies, software development and tests. This solution, called the Grid Enabled Mass Storage System (GEMSS), is based on a custom integration between a fast and reliable parallel filesystem (IBM GPFS), with a complete integrated tape backend based on TIVOLI TSM Hierarchical...Go to contribution page
-
Artur Jerzy Barczyk (California Institute of Technology (US)), Azher Mughal (California Institute of Technology), sandor Rozsa (California Institute of Technology (CALTECH))24/05/2012, 14:45Computer Facilities, Production Grids and Networking (track 4)Parallel40Gb/s network technology is increasingly available today in the data centers as well as in the network backbones. We have built and evaluated storage systems equipped with the last generation of 40GbE Network Interface Cards. The recently available motherboards with the PCIe v3 bus provide the possibility to reach the full 40Gb/s rate per network interface. A fast caching system was built...Go to contribution page
-
Tim Bell (CERN)24/05/2012, 15:10Computer Facilities, Production Grids and Networking (track 4)ParallelThe CERN Computer Centre is reviewing strategies for optimizing the use of the existing infrastructure in the future, and in the likely scenario that any extension will be remote from CERN, and in the light of the way other large facilities are today being operated. Over the past six months, CERN has been investigating modern and widely-used tools and procedures used for virtualisation,...Go to contribution page