Conveners
Computing and Batch Systems
- Ofer Rind (BROOKHAVEN NATIONAL LABORATORY)
- Wolfgang Friebel (Deutsches Elektronen-Synchrotron (DE))
- Michele Michelotto (Universita e INFN (IT))
Computing and Batch Systems: Computing and Batch Systems
- Wolfgang Friebel (Deutsches Elektronen-Synchrotron (DE))
- Michele Michelotto (Universita e INFN (IT))
- Ofer Rind (BROOKHAVEN NATIONAL LABORATORY)
Computing and Batch Systems: Computing and Batch Systems
- Michele Michelotto (Universita e INFN (IT))
- Wolfgang Friebel (Deutsches Elektronen-Synchrotron (DE))
- Ofer Rind (BROOKHAVEN NATIONAL LABORATORY)
Computing and Batch Systems: Computing and Batch Systems
- Wolfgang Friebel (Deutsches Elektronen-Synchrotron (DE))
- Ofer Rind (BROOKHAVEN NATIONAL LABORATORY)
- Michele Michelotto (Universita e INFN (IT))
Computing and Batch Systems: Computing and Batch Systems
- Wolfgang Friebel (Deutsches Elektronen-Synchrotron (DE))
- Michele Michelotto (Universita e INFN (IT))
- Ofer Rind (BROOKHAVEN NATIONAL LABORATORY)
Peter Love
(Lancaster University (GB))
25/03/2015, 14:00
Computing & Batch Services
This contribution describes the usage and benchmarking of a commercial data centre running Openstack. Different cloud provisional tools are described highlighting the pros and cons of each system. A comparison is made between this facility and a standard grid T2 site in terms of job throughput and availability. Usage of the centre’s local object store is also described.
Dr
Tony Wong
(Brookhaven National Laboratory)
25/03/2015, 14:25
Computing & Batch Services
The RHIC-ATLAS Computing Facilty (RACF) at BNL has traditionally evaluated hardware on-site, with physical access to the systems. The effort to request evaluation hardware, shipping, set-up and testing has consumed an increasing amount of time and the process has become less productive over the years. To regain past productivity and shorten the evaluation process, BNL has started a pilot...
Gang Qin
(University of Glasgow (GB))
25/03/2015, 14:50
Computing & Batch Services
Modern Linux Kernels include a feature set that enables the
control and monitoring of system resources, called Cgroups. Cgroups
have been enabled on a production HTCondor pool sited at the Glasgow
site of the UKI-SCOTGRID distributed Tier-2. A system has been put in
place to collect and aggregate metrics extracted from Cgroups on all
worker nodes within the Condor pool. From this...
Manfred Alef
(Karlsruhe Institute of Technology (KIT))
25/03/2015, 15:15
Computing & Batch Services
In this talk we will provide information about the current status of the preliminary work to relaunch the HEPiX Benchmarking Working Group which will develop the next release of the HEP CPU benchmark.
Dr
Michele Michelotto
(INFN Padua & CMS)
25/03/2015, 16:05
Computing & Batch Services
The WLCG community has requested a fast benchmark to quickly assess the perfomances of a worker node. A good candidate is a python script used in LHCb
Dr
Lucia Morganti
(INFN)
25/03/2015, 16:30
Computing & Batch Services
Systems on Chip (SoCs), originally targeted for mobile and embedded technology, are becoming attractive for HEP and HPC scientific communities, given their low cost, huge worldwide shipments, low power consumption and increasing processing power - mostly associated with their GPUs.
A variety of development boards are currently available, making it foreseeable to use these power-efficient...
Liviu Valsan
(CERN)
25/03/2015, 16:55
Computing & Batch Services
x86 is the uncontested leader for server platforms in terms of market share and is currently the architecture of choice for High Energy Physics applications. But as more and more importance is given to power efficiency, physical density and total cost of ownership we are seeing new processor architectures emerging and some existing ones becoming more open. With the introduction of AArch64,...
Mr
David Power
(Boston Ltd.)
25/03/2015, 17:20
Computing & Batch Services
The talk's coverage will include Xeon Haswell, ARM and Open Compute Platforms
Daniel Gruber
(U)
26/03/2015, 12:20
Computing & Batch Services
- Introduction - DRMAA2 in a Nutshell
- The C Interface - Data Types, Monitoring Sessions, Job Sessions,
Working with Jobs, Job Templates, Error Handling and Dealing with
Enhancements
- Getting started with DRMAA2
- Example Applications - Job Monitoring Applications and Simple
Multi-Clustering
Jerome Belleman
(CERN)
27/03/2015, 09:00
Computing & Batch Services
The CERN Batch System comprises 4000 worker nodes, 60 queues and offers
a service for various types of large user communities. In light of the
developments driven by the Agile Infrastructure and the more demanding
processing requirements, it is faced with increasingly challenging scalability
and flexibility needs.
This production cluster currently runs IBM/Platform LSF. Over the last...
Manfred Alef
(Karlsruhe Institute of Technology (KIT))
27/03/2015, 09:25
Computing & Batch Services
The Grid Computing Centre Karlsruhe (GridKa) is using the Grid Engine batch system since 2011. In this presentation I will talk about the experiences with this batch system, including multi-core job support, and first experiences with cgroups.
Erik Mattias Wadenstein
(University of Umeå (SE))
27/03/2015, 09:50
Computing & Batch Services
An update on the current status of SLURM usage in the Nordics, as well as recent developments in improving support for LHC type jobs including tuning for efficient scheduling of multicore grid jobs. Also an overview of some remaining challenges will be given together with discussion on how to address them.
Mr
Michel Jouvin
(Laboratoire de l'Accelerateur Lineaire (FR))
27/03/2015, 10:15
Computing & Batch Services
I propose to give a summary of the Condor workshop, held at CERN mid-December.
Andrew David Lahiff
(STFC - Rutherford Appleton Lab. (GB))
27/03/2015, 11:05
Computing & Batch Services
After running Torque/Maui for many years, the RAL Tier-1 migrated to HTCondor during 2013 in order to benefit from improved reliability, scalability and additional functionality unavailable in Torque. This talk will discuss the deployment of HTCondor at RAL, our experiences and the evolution of our pool over the past two years, as well as our future plans.
Jerome Belleman
(CERN)
27/03/2015, 11:30
Computing & Batch Services
While we are taking measures to face the limitations discussed earlier on
in our IBM/Platform LSF cluster, we have been working on setting up a new
batch system based on HTCondor. There has been some progress with the pilot
service which we described last HEPiX. We also went on investigating some
of the more advanced functions which will lead up to the production state
of the new CERN...
Andrew David Lahiff
(STFC - Rutherford Appleton Lab. (GB))
27/03/2015, 11:55
Computing & Batch Services
With the increasing interest in HTCondor in Europe, an important question for sites considering migrating to HTCondor is how well it integrates with the standard grid middleware, in particular integration with the information system and APEL accounting. Also, with the increasing interest and usage of private clouds, how easily a batch system can be integrated with a private cloud is another...
Stephen Jones
(Liverpool University)
27/03/2015, 12:20
Computing & Batch Services
This talk describes DrainBoss, which is a proportional integral (PI) controller with conditional logic that strives to maintain the correct ratio between single-core and multi-core jobs in an ARC/HTCondor cluster. DrainBoss can be used instead of the HTCondor DEFRAG Daemon.