Dr
Simon Patton
(LAWRENCE BERKELEY NATIONAL LABORATORY)
03/09/2007, 08:00
The Unified Software Development Process (USDP) defines a process for developing
software from the initial inception to the final delivery. The process creates a
number of difference models of the final deliverable; the use case, analysis, design,
deployment, implementation and test models. These models are developed using an
iterative approach that breaks down into four main phases;...
Dr
Sven Hermann
(Forschungszentrum Karlsruhe)
03/09/2007, 08:00
Forschungszentrum Karlsruhe is one of the largest science and engineering research
institutions in Europe. The resource centre GridKa as part of this science centre is
building up a Tier 1 centre for the LHC project. Embedded in the European grid
initiative EGEE, GridKa also manages the ROC (regional operation centre) for the
German Swiss region. A ROC is responsible for regional...
Mr
Shahryar Khan
(Stanford Linear Acclerator Center)
03/09/2007, 08:00
The future of Computing in High Energy Physics (HEP) applications depends on
both the Network and Grid infrastructure. Some South Asian countries such as
India and Pakistan are making progress in this direction by not only building
Grid clusters, but also by improving their network infrastructure. However to
facilitate the use of these resources, they need to overcome the issues of...
Mr
Andrey Tsyganov
(Moscow Physical Engineering Inst. (MePhI))
03/09/2007, 08:00
CERN, the European Laboratory for Particle Physics, located in Geneva - Switzerland,
is currently building the LHC, a 27 km particle accelerator. The equipment life-cycle
management of this project is provided by the Engineering and Equipment Data
Management System (EDMS) Service. Using Oracle, it supports the management and
follow-up of different kinds of documentation through the whole...
Mr
Ulrich Fuchs
(CERN & Ludwig-Maximilians-Universitat Munchen)
03/09/2007, 08:00
ALICE is a dedicated heavy-ion detector to exploit the physics potential of
nucleus-nucleus (lead-lead) interactions at LHC energies. The aim is to study
the physics of strongly interacting matter at extreme energy densities, where
the formation of a new phase of matter, the quark-gluon plasma, is expected.
Running in heavy-ion mode the data rate from event building to permanent...
Mr
Belmiro Antonio Venda Pinto
(Faculdade de Ciencias - Universidade de Lisboa)
03/09/2007, 08:00
The ATLAS experiment uses a complex trigger strategy to be able to achieve the
necessary Event Filter rate output, making possible to optimize the storage and
processing needs of these data. These needs are described in the ATLAS Computing
Model which embraces Grid concepts. The output coming from the Event Filter will
consist of four main streams: the physical stream, express stream,...
Dr
David Malon
(Argonne National Laboratory)
03/09/2007, 08:00
In the ATLAS event store, files are sometimes "an inconvenient truth."
From the point of view of the ATLAS distributed data management system,
files are too small--datasets are the units of interest. From the point
of view of the ATLAS event store architecture, files are simply a physical
clustering optimization: the units of interest are event collections--
sets of events that...
Jos Van Wezel
(Forschungszentrum Karlsruhe (FZK/GridKa))
03/09/2007, 08:00
The disk pool managers in use in the HEP community focus on managing disk storage but
at the same time rely on a mass storage i.e. tape based system either to offload data
that has not been touched for a long time or for archival purposes. Traditionally tape
handling systems like HPSS by IBM or Enstore developed at FNAL are used because they
offer specialized features to overcome the...
Kaushik De
(UT-Arlington)
03/09/2007, 08:00
During 2006-07, the ATLAS experiment at the Large Hadron Collider launched a massive
Monte Carlo simulation production exercise to commission software and computing
systems in preparation for data in 2007. In this talk, we will describe the goals and
objectives of this exercise, the software systems used, and the tiered computing
infrastructure deployed worldwide. More than half a petabyte...
Dr
Monica Verducci
(European Organization for Nuclear Research (CERN))
03/09/2007, 08:00
One of the most challenging task faced by the LHC experiments will be the storage of
"non-event data" produced by calibration and alignment stream processes into the
Conditions Database. For the handling of this complex experiment conditions data the
LCG Conditions Database Project has implemented COOL, a new software product designed
to minimise the duplication of effort by developing a...
Nils Gollub
(CERN),
Nils Gollub
(University of Uppsala)
03/09/2007, 08:00
ATLAS Tile Calorimeter detector (TileCal) is presently involved in an intense phase
of commissioning with cosmic rays and subsystems integration. Various monitoring
programs have been developed at different level of the data flow to tune the set-up
of the detector running conditions and to provide a fast and reliable assessment of
the data quality.
The presentation will focus on the...
Dr
Amir Farbin
(European Organization for Nuclear Research (CERN))
03/09/2007, 08:00
The EventView Analysis Framework is currently the basis for much of the analysis software employed by various
ATLAS physics groups (for example the Top, SUSY, Higgs, and Exotics working groups). In ATLAS's central data
preparation, this framework provides an assessment of data quality and the first analysis of physics data for the
whole collaboration. An EventView is a self-consistent...
Mr
Bruno Hoeft
(Forschungszentrum Karlsruhe)
03/09/2007, 08:00
While many fields relevant to Grid security are already covered by existing
working groups, their remit rarely goes beyond the scope of the Grid
infrastructure itself. However, security issues pertaining to the internal set-up
of compute centres have at least as much impact on Grid security. Thus, this
talk will present briefly the EU ISSeG project (Integrated Site Security for Grids)....
Dr
Andreas Gellrich
(DESY)
03/09/2007, 08:00
As a partner of the international EGEE project in the German/Switzerland
federation (DECH) and as a member of the national D-GRID initiative, DESY
operates a large-scale production-grade Grid infrastructure with hundreds of CPU
cores and hundreds of Terabytes of disk storage.
As Tier-2/3 center for ATLAS and CMS DESY plays a leading role in Grid computing
in Germany. DESY strongly support...
Dr
Bockjoo Kim Kim
(University of Florida)
03/09/2007, 08:00
The CMS experiment will begin data collection at the end of 2007 and
released its software with new framework since the end of 2005.
The CMS experiment employs a tiered distributed computing
based on the Grids, the LHC Computing Grid (LCG) and the Open
Science Grid (OSG). There are approximately 37 tiered CMS centers around the world.
The number of the CMS software releases was three...
Dirk Hufnagel
(for the CMS Offline/Computing group)
03/09/2007, 08:00
With the upcoming LHC engineering run in November, the CMS Tier0 computing
effort will be the one of the most important activities of the experiment.
The CMS Tier0 is responsible for all data handling and processing of real
data events in the first period of their life, from when the data is
written by the DAQ system to a disk buffer at the CMS experiment site to
when it is transferred...
Dr
Ivan D. Reid
(School of Design and Engineering - Brunel University, UK)
03/09/2007, 08:00
Goodness-of-fit statistics measure the compatibility of random samples against some
theoretical probability distribution function. The classical one-dimensional
Kolmogorov-Smirnov test is a non-parametric statistic for comparing two empirical
distributions, which defines the largest absolute difference between the two
cumulative probability distribution functions as a measure of...
Mr
Georges Kohnen
(Université de Mons-hainaut)
03/09/2007, 08:00
The IceCube neutrino telescope is a cubic kilometer Cherenkov detector currently under construction in the deep ice at the geographic South Pole. As of 2007, it has reached more than 25 % of its final instrumented volume and is actively taking data. We will briefly describe the design and current status, as well as the physics goals of the detector. The main focus will, however, be on the...
Mr
Martin Gasthuber
(Deutsches Elektronen Synchrotron (DESY))
03/09/2007, 08:00
Based on todays understanding of LHC scale analysis requirements and
the clear dominance of fast and high capacity random access storage,
this talk will present a generic architecture for a national facility
based on existing components from various computing domains. The
following key areas will be discussed in detail and solutions will be
proposed, building the overall...
Craig Dowell
(Univ. of Washington)
03/09/2007, 08:00
The ATLAS Muon Spectrometer is constructed out of 1200 drift tube chambers
with a total area of nearly 7000 square meters. It must determine muon track
positions
to a very high precision despite its large size necessitating complex real-time
alignment measurements.
Each chamber, as well as approximately 50 alignment reference bars in the
endcap region,
are equipped with CCD cameras,...
Marco Clemencic
(European Organization for Nuclear Research (CERN))
03/09/2007, 08:00
The COOL software has been chosen by both Atlas and LHCb as the base of
their conditions database infrastructure. The main focus of the COOL project in
2007 will be the deployment, test and validation of Oracle-based COOL
database services at Tier0 and Tier1. In this context, COOL software
development will concentrate on service-related issues, and in particular on the
optimization...
Dr
Dantong Yu
(Brookhaven National Laboratory), Dr
Dimitrios Katramatos
(Brookhaven National Laboratory), Dr
Shawn McKee
(University of Michigan)
03/09/2007, 08:00
Computer facilities, production grids and networking
Supporting reliable, predictable, and efficient global movement of data in
high-energy physics distributed computing environments requires the capability to
provide guaranteed bandwidth to selected data flows and schedule network usage
appropriately. The DOE-funded TeraPaths project at Brookhaven National Laboratory
(BNL), currently in its second year, is developing methods and tools that...
Dr
Giuseppe Della Ricca
(Univ. of Trieste and INFN)
03/09/2007, 08:00
The electromagnetic calorimeter of the Compact Muon Solenoid experiment will play a
central role in the achievement of the full physics performance of the detector at
the LHC. The detector performance will be monitored using applications based on the
CMS Data Quality Monitoring (DQM) framework and running on the High-Level Trigger
Farm as well as on local DAQ systems. The monitorable...
Dr
Doris Ressmann
(Forschungszentrum Karlsruhe)
03/09/2007, 08:00
The grid era brings upon new and steeply rising demands in data storage.
The GridKa project at Forschungszentrum Karlsruhe delivers its share of
the computation and storage requirements of all LHC and 4 other HEP
experiments. Access throughput from the worker nodes to the storage can
be as high a 2 GB/s. At the same time a continuous throughput in the
order of 300-400 MB/s into and...
Dr
Manuela Cirilli
(University of Michigan)
03/09/2007, 08:00
The calibration of the 375000 ATLAS Monitored Drift Tubes will be a highly challenging task: a dedicated set of data
will be extracted from the second level trigger of the experiment and streamlined to three remote Tier-2 Calibration
Centres.
This presentation reviews the complex chain of databases envisaged to support
the MDT Calibration and describes the actual status of the...
Dr
Wolfgang Waltenberger
(Hephy Vienna)
03/09/2007, 08:00
A tool is presented that is capable of reading from and writing to several
different file formats. Currently supported file formats are ROOT, HBook, HDF,
XML, Sqlite3 and a few text file formats. A plugin mechanism decouples
the file-format specific "backends" from the main library.
All data are internally represented as "heterogenous hierarchic tuples"; no
other data structure exists in...
Ian Fisk
(Fermi National Accelerator Laboratory (FNAL))
03/09/2007, 08:00
CMS is preparing seven remote Tier-1 computing facilities to archive and serve
experiment data. These centers represent the bulk of CMS's data serving capacity, a
significant resource for reprocessing data, all of the simulation archiving capacity,
and operational support for Tier-2 centers and analysis facilities. In this paper we
present the progress on deploying the largest remote...
Irina Sourikova
(BROOKHAVEN NATIONAL LABORATORY)
03/09/2007, 08:00
After seven years of running and collecting 2 Petabytes
of physics data, PHENIX experiment at the Relativistic Heavy Ion Collider
(RHIC) has gained a lot of experience with database management systems ( DBMS ).
Serving all of the experiment's operations - data taking, production and analysis -
databases provide 24/7 access to calibrations and book-keeping information
for hundreds of...
Dr
Iosif Legrand
(CALTECH),
Ramiro Voicu
(CALTECH)
03/09/2007, 08:00
The efficient use of high-speed networks to transfer large data sets is an essential
component for many scientific applications including CERN’s LCG experiments.
We present an efficient data transfer application, Fast Data Transfer (FDT), and a
distributed agent system (LISA) able to monitor, configure, control and globally
coordinate complex, large scale data transfers.
FDT is an...
Prof.
Toby Burnett
(University of Washington)
03/09/2007, 08:00
Applications often need to have many parameters defined for execution. A few can be
done with the command line, but this does not scale very well. I present a simple use of
embedded Python that makes it easy to specify configuration data for applications,
avoiding wiring in constants, or writing elaborate parsing difficult to justify for small, or
one-off applications. But the...
Elisabetta Ronchieri
(INFN CNAF)
03/09/2007, 08:00
People involved in modular projects need to improve the build software process,
planning the correct execution order and detecting circular dependencies. The lack of
suitable tools may cause delays in the development, deployment and maintenance of the
software.
Experience in such projects has shown that the arranged use of version
control and build systems is not able to support the...
Mr
Alexander Withers
(Brookhaven National Laboratory)
03/09/2007, 08:00
The PostgreSQL database is a vital component of critical services at the
RHIC/USATLAS Computing Facility such as the Quill subsystem of the Condor
Project and both PNFS and SRM within dCache. Current deployments are
relatively unsophisticated, utilizing default configurations on small-scale
commodity hardware. However, a substantial increase in projected growth has
exposed deficiencies...
Dr
Maria Grazia Pia
(INFN Genova)
03/09/2007, 08:00
The Statistical Toolkit provides an extensive collection of algorithms for the
comparison of two data samples: in addition to the chisquared test, it includes all
the tests based on the empirical distribution function documented in literature for
binned and unbinned distributions.
Some of these tests, like the Kolmogorov-Smirnov one, are widely used; others, like
the Anderson-Darling...
Dr
Elliott Wolin
(Jefferson Lab)
03/09/2007, 08:00
EVIO is a lightweight event I/O package consisting of an
object-oriented layer on top of a pre-existing, highly efficient,
C-based event I/O package. The latter, part of the JLab CODA package,
has been in use in JLab high-speed DAQ systems for many years, but
other underlying disk I/O packages could be substituted. The event
format on disk, a packed tree-like hierarchy of banks, maps...
Dr
Jose Hernandez
(CIEMAT)
03/09/2007, 08:00
CMS undertakes periodic computing challenges of increasing scale and complexity
to test its computing model and Grid computing systems. The computing challenges are
aimed at establishing a working distributed computing system that implements the CMS
computing model based on an underlying multi-flavour grid infrastructure. CMS
dataflows and data processing workflows are exercised during a...
Mr
LUIS MARCH
(Instituto de Fisica Corpuscular)
03/09/2007, 08:00
The Spanish ATLAS Tier-2 is geographically distributed between three HEP institutes.
They are IFAE (Barcelona) and IFIC (Valencia) and UAM (Madrid). Currently it has a
computing power of about 400 kSI2k CPU, a disk storage capacity of 40 TB and a
network bandwidth connecting the three sites and the nearest Tier-1 of 1 Gb/s. These
resources will increase with time in parallel to those of...
Tomas Kouba
(Institute of Physics - Acad. of Sciences of the Czech Rep. (ASCR)
03/09/2007, 08:00
Each tier 2 site is monitored by various services from outside. The
Prague T2 is monitored by SAM tests, GSTAT monitoring, RTM from RAL, regional
nagios monitoring and experiment specific tools. Besides that we monitor our
own site for hardware and software failures and middleware status.
All these tools produce an output that must be regularly checked by
site administrators. We...
Mr
Alessandro Italiano
(INFN-CNAF)
03/09/2007, 08:00
Every day operations on a big computer center farm like that of a Tier1 can be numerous. Opening or closing a
host, changing batch system configuration, replacing a disk, reinstalling a host and so on, is just a short list of
what can and will really happen. In these conditions remembering all that has been done could be really difficult.
Typically a big farm is managed by a team so it...
Dr
Chadwick Keith
(Fermilab)
03/09/2007, 08:00
Fermilab supports a scientific program that includes experiments and scientists
located across the globe. In order to better serve this community, Fermilab has
placed its production computer resources in a Campus Grid infrastructure called
'FermiGrid'. The FermiGrid infrastructure allows the large experiments at Fermilab to
have priority access to their own resources, enables sharing of...
Prof.
Wolfgang Kuehn
(Univ. Giessen, II. Physikalisches Institut)
03/09/2007, 08:00
PANDA is a new universal detector for antiproton physics at the HESR facility at
FAIR/GSI. The PANDA data acquisition system has to handle interaction rates of
the order of 10**7 /s and data rates of several 100 Gb /s. FPGA based
compute nodes with multi-Gb/s bandwidth capability using the ATCA
architecture are designed to handle tasks such as event building, feature
extraction and...
Dr
Sven Gabriel
(Forschungszentrum Karlsruhe)
03/09/2007, 08:00
GridKa is the German Tier1 centre in the Worldwide LHC Computing Grid
(WLCG). It is part of the Institut für Wissenschaftliches Rechnen (IWR) at the
Forschungszentrum Karlsruhe (FZK). It started in 2002 as the successor of the
”Regional Data and Computing Centre in Germany” (RDCCG)
GridKa supports all four LHC experiments, ALICE, ATLAS, CMS and LHCb,
four non-LHC high energy physics...
Dr
Christopher Jones
(Cornell University)
03/09/2007, 08:00
When doing an HEP analysis, physicists typically repeat the same operations over and over while applying minor
variations. Doing the operations as well as remembering the changes done during each iteration can be a very
tedious process. HEPTrails in an analysis application written in Python and built on top of the University of Utah's
VisTrails system which provides workflow and full...
Dr
Enrico Mazzoni
(INFN Pisa)
03/09/2007, 08:00
We report about the tests performed in the INFN Pisa Computing Centre with some of
the latest generation storage devices. Fibre Channel and NAS solutions have been
tested in a realistic enviroment, both participating in Worldwide CMS's Service
Challenges, and simulating analysis patterns with more than 500 jobs accessing
concurrently]data files. Both usage pattern have evidentiated the...
Dr
David Bailey
(University of Manchester), Dr
Robert Appleby
(University of Manchester)
03/09/2007, 08:00
Understanding modern particle accelerators requires simulating charged
particle transport through the machine elements. These simulations can be very
time consuming due to the large number of particles and the need to consider
many turns of a circular machine. Stream computing offers an attractive way to
dramatically improve the performance of such simulations by calculating the...
Mr
Enrico Fattibene
(INFN-CNAF, Bologna, Italy), Mr
Federico Pescarmona
(INFN-Torino, Italy), Mr
Giuseppe Misurelli
(INFN-CNAF, Bologna, Italy), Mr
Stefano Dal Pra
(INFN-Padova, Italy)
03/09/2007, 08:00
In production quality Grid infrastructure accounting data play a key role on the
possibility to spot out how the allocated resources have been used. The different
types of Grid user have to be taken into account in order to provide different
subsets of accounting data based on the specific role covered by a Grid user.
Grid end users, VO (Virtual Organization) managers, site administrators...
Dr
Patricia Conde Muíño
(LIP-Lisbon)
03/09/2007, 08:00
ATLAS is one of the four major LHC experiments, designed to cover a wide range
of physics topics. In order to cope with a rate of 40MHz and 25 interactions per
bunch crossing, the ATLAS trigger system is divided in three different levels. The
first one (LVL1, hardware based) identifies signatures in 2 microseconds that are
confirmed by
the the following trigger levels (software based)....
Antonio Amorim
(Universidade de Lisboa (SIM and FCUL, Lisbon))
03/09/2007, 08:00
The ATLAS conditions databases will be used to manage information of quite diverse
nature and
level of complexity. The infrastructure in being built using the LCG COOL
infrastructure and
provides a powerful information sharing gateway upon many different systems. The
nature of
the stored information ranges from temporal series of simple values to very complex
objects
describing...
Luca dell'Agnello
(INFN-CNAF)
03/09/2007, 08:00
INFN CNAF is a multi experiment computing center acting as Tier-1 for LCG but also
supporting other HEP and non HEP experiments and Virtual Organizations.
The CNAF Tier-1 is one of the main Resource Centers of the Grid Infrastructure
(WLCG/EGEE); the preferred access method to the center is through WLCG/EGEE and
INFNGRID middleware and services.
Critical issues to be addressed to meet...
Prof.
Manuel Delfino Reznicek
(Port d'Informació Científica (PIC))
03/09/2007, 08:00
A new data center has been deployed for the MAGIC Gamma Ray Telescope, located in the
Roque de los Muchachos observatory in the Canary Islands, Spain, at the Port
d'Informació Científica in Barcelona. The MAGIC Datacenter at PIC recieves all the
raw data produced by MAGIC, either via the network or tape cartridges, and provides
archiving, rapid processing for quality control and...
Tomasz Wlodek
(Brookhaven National Laboratory)
03/09/2007, 08:00
Managing large number of heterogeneous grid servers with different service
requirements posts great challenges. We describe a cost-effective integrated
operation framework which manages hardware inventory, monitors services, raises
alarms with different severity levels and tracks the facility response to them.
The system is based on open source components: RT (Request Tracking) tracks...
Jonathan Butterworth
(University College London)
03/09/2007, 08:00
Accurate modelling of high energy hadron interactions is essential for the precision
analysis of data from the LHC. It is therefore imperative that the predictions of
Monte Carlos used to model this physics are tested against existing and future
measurements. These measurements cover a wide variety of reactions, experimental
observables and kinematic regions. To make this process more...
Antonio Amorim
(Universidade de Lisboa (SIM and FCUL, Lisbon))
03/09/2007, 08:00
The ATLAS Trigger and Data Acquisition systems (TDAQ) to the Conditions databases has
strong requirements on reliability and performance. Several applications were
developed to support the integration of Condition database access with the online
services in TDAQ like the interface to the Information Services and to the TDAQ
configuration..
The DBStressor was developed to test and stress...
Vincenzo Chiochia
(Universitat Zurich)
03/09/2007, 08:00
The CMS Pixel Detector is hosted inside the large solenoid generating a
magnetic field of 4 T.
The electron-hole pairs produced by particles traversing the pixel sensors will thus
experience the Lorentz force due to the combined presence of magnetic and electric field.
This results in a systematic shift of the charge distribution. In order to achieve a
high position resolution a...
Dr
Robert Bainbridge
(Imperial College London)
03/09/2007, 08:00
The CMS silicon strip tracker is unprecedented in terms of its size and
complexity, providing a sensitive area of >200 m^2 and comprising 10M
readout channels. Its data acquisition system is based around a custom
analogue front-end ASIC, an analogue optical link system and an
off-detector VME board that performs digitization, zero-suppression and
data formatting. These data are forwarded...
Mr
Philip DeMar
(FERMILAB)
03/09/2007, 08:00
Advances in wide area network service offerings, coupled with comparable developments
in local area network technology have enabled many HEP sites to keep their offsite
network bandwidth ahead of demand. For most sites, the more difficult and costly
aspect of increasing wide area network capacity is the local loop, which connects the
facility LAN to the wide area service provider(s). ...
Dr
Andreas Heiss
(Forschungszentrum Karlsruhe)
03/09/2007, 08:00
Within the Worldwide LHC Computing Grid (WLCG), a Tier-1 centre like the German
GridKa computing facility has to provide significant CPU and storage resources as
well as several Grid services with a high level of quality. GridKa currently supports
all four LHC Experiments, Alice, Atlas, CMS and LHCb as well as four non-LHC high
energy physics experiments, and is about to significantly...
Dr
David Lawrence
(Jefferson Lab)
03/09/2007, 08:00
The C++ reconstruction framework JANA has been written to support the
next generation of Nuclear Physics experiments at Jefferson Lab in
anticipation of the 12GeV upgrade. The JANA framework was designed to allow
multi-threaded event processing with a minimal impact on developers of
reconstruction software. As we enter the multi-core (and soon
many-core) era, thread-enabled code will...
Prof.
Manuel Delfino Reznicek
(Port d'Informació Científica (PIC))
03/09/2007, 08:00
Small files pose performance issues for Mass Storage Systems, particularly those
using magnetic tape. The ViVo project reported at CHEP06 solved some of these
problems by using Virtual Volumes based on ISO images containing the small files, and
only storing and retrieving these images from the MSS. Retrieval was handled using
Unix automounters, requiring deployment of ISO servers with a...
Ms
Geraldine Conti
(EPFL)
03/09/2007, 08:00
The LHCb warm magnet has been designed to provide an integrated field of 4
Tm for tracks coming from the primary vertex.To insure good momentum
resolution of a few per mil, an accurate description of the magnetic field map is
needed. This is achieved by combining the information from a TOSCA-based
simulation and data from measurements. The paper presents the fit method
applied to...
Dr
Sebastien Binet
(LBNL)
03/09/2007, 08:00
LHC experiments are entering in a phase where optimization in view of
data taking as well as robustness' improvements are of major importance. Any
reduction in event data size can bring very significant savings in the
amount of hardware (disk and tape in particular) needed to process
data. Another area of concern and potential major gains is reducing
the memory size and I/O bandwidth...
Miguel Coelho Dos Santos
(CERN)
03/09/2007, 08:00
We present our design, development and deployment of a portable monitoring system for
the CERN Archival and Storage System (Castor) based on its existing internal database
infrastructure and deployment architecture.
This new monitoring architecture is seen as an important requirement for future
development and support. Castor is now deployed at several sites which use...
Dr
Stefano Mersi
(INFN & Università di Firenze)
03/09/2007, 08:00
The CMS silicon strip tracker comprises a sensitive area of >200 m2 and 10M readout
channels. Its data acquisition system is based around a custom analogue front-end
ASIC, an analogue optical link system and an off-detector VME board that performs
digitization, zero-suppression and data formatting. The data acquisition system uses
the CMS online software framework, known as XDAQ, to...
Dr
Charles Leggett
(LAWRENCE BERKELEY NATIONAL LABORATORY)
03/09/2007, 08:00
Runtime memory usage in experiments has grown enormously in recent years,
especially in large experiments like Atlas. However, it is difficult
to break down total memory usage as indicated by OS-level tools, to
identify the precise users and abusers. Without a detailed knowledge
of memory footprints, monitoring memory growth as an experiment
evolves in order to control ballooning...
Mr
Sebastian Lopienski
(CERN)
03/09/2007, 08:00
Nowadays, IT departments provide, and people use, computing services of an
increasingly heterogeneous nature. There is thus a growing need for a status display
that groups these different services and reports status and availability in a uniform
way. The Service Level Status (SLS) system addresses these needs by providing a
web-based display that dynamically shows availability, basic...
Dr
Ilya Narsky
(California Institute of Technology)
03/09/2007, 08:00
SPR implements various tools for supervised learning such as boosting (3 flavors),
bagging, random forest, neural networks, decision trees, bump hunter (PRIM),
multi-class learner, logistic regression, linear and quadratic discriminant analysis,
and others. Presented at CHEP 2006, SPR has been extended with several important
features since then. The package has been stripped of CLHEP...
Mr
Andreas Unterkircher
(CERN)
03/09/2007, 08:00
We describe the methodology for testing gLite releases. Starting from the needs given
by the EGEE software management process we illustrate our design choices for testing
gLite.
For certifying patches different test scenarios have to be considered: regular
regression tests, stress tests and manual verification of bug fixes. Conflicts arise
if these tests are all carried out at the same...
Mr
Ian Gable
(University of Victoria)
03/09/2007, 08:00
The ATLAS Canada computing model consists of a Tier-1 computing centre located at
the TRIUMF Laboratory in Vancouver, Canada, and two distributed Tier-2 computing
centres: one in Eastern Canada and one in Western Canada. Each distributed Tier-2
computing centre is made up of a
group of universities. To meet the network requirements of each institution, HEPnet
Canada and CANARIE...
Alessandro De Salvo
(Istituto Nazionale di Fisica Nucleare Sezione di Roma 1)
03/09/2007, 08:00
The huge amount of resources available in the Grids, and the necessity
to have the most updated experiment software deployed in all the sites
within a few hours, have spotted the need for automatic installation
systems for the LHC experiments.
In this paper we describe the ATLAS system for the experiment software
installation in LCG/EGEE, based on the Lightweight Job Submission
Framework...
Ms
Elizabeth Sexton-Kennedy
(FNAL)
03/09/2007, 08:00
With the turn-on of the LHC, the CMS DAQ system is expecting to log petabytes of
experiment data in the coming years. The CMS Storage Manager system is a part of the
high bandwidth event data handling pipeline of the CMS high level DAQ. It has two
primary functions. Each Storage Manager instance collects data from the sub-farm, or
DAQ slice of the Event Filter farm it has been assigned...
Prof.
Gang Chen
(IHEP, China)
03/09/2007, 08:00
Beijing Electron Spectrometer (BESIII) experiment will produce 5 PB of data in next
five years. Grid is used to solve this challenge. This paper introduces BES grid
computing model and specific technologies, including automatic data replication,
fine-grained job scheduling and so on.
Obreshkov Emil
(INRNE/CERN)
03/09/2007, 08:00
The ATLAS offline software comprises over 1000 software packages organized into 10 projects that are built on a
variety of compiler and operating system combinations every night. File-level parallelism, package-level parallelism
and multi-core build servers are used to perform simultaneous builds of 6 platforms that are merged into a single
installation on AFS. This in turn is used to...
Mr
Sigve Haug
(LHEP University of Bern)
03/09/2007, 08:00
Since 2005 the Swiss ATLAS Grid is in production. It comprises
four clusters at one Tier 2 and two Tier 3 sites. About 800
heterogenous cores and 60 TB disk space are connected by
a dark fibre network operated at 10 Giga bit per second. Three different
operating systems are deployed. The Tier 2 cluster runs both LCG and
NorduGrid middleware (ARC) while the Tier 3 clusters run only the...
Dan Nae
(California Institute of Technology (CALTECH))
03/09/2007, 08:00
In this paper we present the design, implementation and evolution of the
mission-orientedUSLHCNet for HEP research. The design philosophy behind our network
is to help meet the dataintensive computing challenges of the next generation of
particle physics experiments with a comprehensive, network-focused approach. Instead
of treating the network as a static, unchanging and unmanaged set of...
Dr
Patricia Conde Muíño
(LIP-Lisbon)
03/09/2007, 08:00
With the project PHEASANT a DSVQL was proposed for the purpose of providing a tools
that could increase user's productivity while producing query code for data analysis.
The previous project aimed at the proof concept and methodology feasability by
introducing the concept of DSLs. We are now concetrated on implementation issues in
order to deploy a final tool.
The concept of domain...
Konstantinos Bachas
(Aristotle University of Thessaloniki)
03/09/2007, 08:00
The measurement of the muon energy deposition in the calorimeters is an integral part of muon
identification, track isolation and correction for catastrophic muon energy losses, which are the
prerequisites to the ultimate goal of refitting the muon track using calorimeter information as well. To this
end, an accurate energy loss measurement method in the calorimeters is developed which...
Kenneth Bloom
(University of Nebraska-Lincoln)
03/09/2007, 08:00
The CMS computing model relies heavily on the use of "Tier-2"
computing centers. At LHC startup, the typical Tier-2 center will have
1 MSpecInt2K of CPU resources, 200 TB of disk for data storage,
and a WAN connection of at least 1 Gbit/s. These centers will be the
primary sites for the production of large-scale simulation samples
and for the hosting of experiment data for user...
Dr
Ulrich Schwickerath
(CERN)
03/09/2007, 08:00
LSF 7, the latest version of Platform's batch workload management
system, addresses many issues which limited the ability of LSF 6.1 to
support large scale batch farms, such as the lxbatch service at CERN. In
this paper we will present the status of the evaluation and deployment
of LSF 7 at CERN, including issues concerning the integration of LSF 7
witht the gLite grid...
Mr
Colin Morey
(Manchester University)
03/09/2007, 08:00
Cfengine is a middle to high level policy language and autonomous agent for
building expert systems to administrate and configure large computer clusters. It is
ideal for large-scale cluster management and is highly portable across varying
computer platforms, allowing the management of multiple architectures and node types
within the same farm.
As well as being a highly capable...
Mr
Andrey Bobyshev
(FERMILAB)
03/09/2007, 08:00
At Fermilab, there is a long history of utilizing network flow data collected from
site routers for various analyses, including network performance characterization,
anomalous traffic detection, investigation of computer security incidents, network
traffic statistics and others. Fermilab’s flow analysis model is currently built as a
distributed system that collects flow data from the site...
Dr
David Alexander
(Tech-X Corporation)
03/09/2007, 08:00
Nuclear and high-energy physicists routinely execute data processing and data
analysis jobs on a Grid and need to be able to monitor their jobs execution at an
arbitrary site at any time. Existing Grid monitoring tools provide abundant
information about the whole system, but are geared towards production jobs and well
suited for Grid administrators, while the information tailored towards...
Prof.
Gordon Watts
(University of Washington)
03/09/2007, 08:00
ROOT is firmly based on C++ and makes use of many of its features –
templates and multiple inheritance, in particular. Many modern languages like
Java and C# and python are missing these features or have radically different
implementations. These programming languages, however, have many
advantages to offer scientists including improved programming paradigms,
development...
Hegoi Garitaonandia Elejabarrieta
(Instituto de Fisica de Altas Energias (IFAE))
03/09/2007, 08:00
The ATLAS Trigger & Data Acquisition System has been designed to use more than
2000 CPUs. During the current development stage it is crucial to test the system on a
number of CPUs of similar scale. A dedicated farm of this size is difficult to find, and
can only be made available for short periods. On the other hand many large farms
have become available recently as part of computing...
Prof.
Harvey Newman
(CALTECH)
03/09/2007, 08:00
The main objective of the VINCI project is to enable data intensive applications to
efficiently use and coordinate shared, hybrid network resources, to improve the
performance and throughput of global-scale grid systems, such as those used in high
energy physics. VINCI uses a set of agent-based services implemented in the MonALISA
framework to enable the efficient use of network resources,...
Dr
Tony Chan
(BROOKHAVEN NATIONAL LAB)
03/09/2007, 08:00
The Brookhaven Computing Facility provides for the computing needs of the
RHIC experiments, supports the U.S. Tier 1 center for the ATLAS experiment
at the LHC and provides computing support for the LSST experiment. The
multi-purpose mission of the facility requires a complex computing infrastructure
to meet different requirements and can result in duplication of services with a
large...
Tejinder Virdee
(CERN/Imperial College)
03/09/2007, 09:15
The current status of the LHC machine and the experiments, especially the general-purpose experiments, will be given. Also discussed will be the preparations for the physics run in 2008. The prospects for physics, with an emphasis on what can be expected with an integrated luminosity of 1 fb-1, will be outlined.
Les Robertson
(CERN)
03/09/2007, 10:00
The talk will review the progress so far in setting up the distributed computing services for LHC
data handling and analysis and look at some of the challenges we face when the real data
begins to flow.
Dr
Amir Farbin
(European Organization for Nuclear Research (CERN))
03/09/2007, 14:00
As we near the collection of the first data from the Large Hadron Collider, the
ATLAS collaboration is preparing the software and computing infrastructure to
allow quick analysis of the first data and support of the long-term steady-state
ATLAS physics program. As part of this effort considerable attention has been
payed to the "Analysis Model", a vision of the interplay of the...
Dr
Steven Goldfarb
(University of Michigan)
03/09/2007, 14:00
I report on major current activities in the domain of Collaborative Tools, focusing
on development for the LHC
collaborations and HEP, in general, including audio and video conferencing, web
archiving, and more. This
presentation addresses the follow-up to the LCG RTAG 12 Final Report (presented at
CHEP 2006), including the
formation of the RCTF (Remote Collaboration Task Force) to...
Dr
Andrew Maier
(CERN)
03/09/2007, 14:00
Distributed data analysis and information management
oral presentation
Ganga, the job-management system (http://cern.ch/ganga), developed as an ATLAS- LHCb common project,
offers a simple, efficient and consistent user experience in a variety of heterogeneous environments: from local
clusters to global Grid systems. Ganga helps end-users to organise their analysis activities on the Grid by providing
automatic persistency of the job's metadata. A user has...
Dr
Jamie Shiers
(CERN)
03/09/2007, 14:00
Computer facilities, production grids and networking
oral presentation
This talk summarises the main lessons learnt from deploying WLCG production services,
with a focus on Reliability, Scalability, Accountability, which lead to both
manageability and usability.
Each topic is analysed in turn. Techniques for zero-user-visible downtime for the
main service interventions are described, together with pathological cases that need
special treatment. The...
Prof.
Adele Rimoldi
(Pavia University & INFN)
03/09/2007, 14:00
The ATLAS detector is entering the final phases of construction and
commissioning in order to be ready to take data during the first LHC
commissioning run, foreseen by the end of 2007. A good understanding of
the experiment performance from the beginning is essential to
efficiently debug the detector and assess its physics potential in view
of the physics runs which are going to take...
Dr
Akram Khan
(Brunel University)
03/09/2007, 14:20
Distributed data analysis and information management
oral presentation
ASAP is a system for enabling distributed analysis for CMS physicists. It was
created with the aim of simplifying the transition from a locally running application
to one that is distributed across the Grid. The experience gained in operating the
system for the past 2 years has been used to redevelop a more robust, performant and
scalable version. ASAP consists of a client for job...
Prof.
Nobuhiko Katayama
(High Energy Accelerator Research Organization)
03/09/2007, 14:20
We developed the original CABS language more than 10 years ago. The main
objective of the language was to describe a decay of a particle as simply as
possible in the context of usual HEP data analysis. A decay mode, for example,
can be defined as follows:
define Cand Dzerobar kpi 2 { K+ identified pi- identified }
hist 1d inv_mass 0 80 1.5 2.3 ``all momentum''
cut inv_mass...
Mr
Philippe Galvez
(California Institute of Technology)
03/09/2007, 14:20
The EVO (Enabling Virtual Organizations) system is based on a new distributed and
unique architecture, leveraging the 10+ years of unique experience of developing and
operating the large distributed production based VRVS collaboration system. The
primary objective being to provide to the High Energy and Nuclear Physics experiments
a system/service that meet their unique requirements of...
Sunanda Banerjee
(Fermilab/TIFR)
03/09/2007, 14:20
The CMS simulation based on the Geant4 toolkit and the CMS object-oriented framework
has been in production for more than three years and has delivered a total of more
than 200 M physics events for the CMS Data Challenges and Physics Technical Design
Report studies. The simulation software has been successfully ported to the new CMS
Event-Data-Model based software framework and is used in...
Dr
Markus Schulz
(CERN)
03/09/2007, 14:20
Computer facilities, production grids and networking
oral presentation
Today's production Grids connect large numbers of distributed hosts using high
throughput networks and hence are valuable targets for attackers. In the same way
users transparently access any Grid service independently of its location, an
attacker may attempt to propagate an attack to different sites that are part of a
Grid. In order to contain and resolve the incident, and since such an...
Dr
Oliver Gutsche
(FERMILAB)
03/09/2007, 14:20
The CMS computing model to process and analyze LHC collision data
follows a data-location driven approach and is using the WLCG
infrastructure to provide access to GRID resources. As a preparation
for data taking beginning end of 2007, CMS tests its computing model
during dedicated data challenges.
Within the CMS computing model, user analysis plays an important role
in the CMS...
Mr
Emmanuel Ormancey
(CERN)
03/09/2007, 14:40
The need for Single Sign On has always been restricted by the lack of cross
platform solutions: a single sign on working only on one platform or technology
is nearly useless. The recent improvements in Web Services Federation (WS-
Federation) standard enabling federation of identity, attribute, authentication
and authorization information can now provide real extended Single Sign On...
Dr
Amber Boehnlein
(FERMI NATIONAL ACCELERATOR LABORATORY)
03/09/2007, 14:40
High energy physics experiments periodically reprocess data, in order to take
advantage of improved understanding of the detector and the data processing code.
Between February and May 2007, the DZero experiment will reprocess a substantial
fraction of its dataset. This consists of half a billion events, corresponding to
more than 100 TB of data, organized in 300,000 files.
The...
Mr
Jan Fiete Grosse Oetringhaus
(CERN)
03/09/2007, 14:40
Distributed data analysis and information management
oral presentation
ALICE (A Large Ion Collider Experiment) at the LHC plans to use a PROOF cluster at CERN (CAF - Cern Analysis
Facility) for fast analysis. The system is especially aimed at the prototyping phase of analyses that need a high
number of development iterations and thus desire a short response time. Typical examples are the tuning of cuts
during the development of an analysis as well as...
Mrs
Ruth Pordes
(FERMILAB)
03/09/2007, 14:40
Computer facilities, production grids and networking
oral presentation
The Open Science Grid (OSG) is receiving five years of funding across six program offices of the Department of
Energy Office of Science and the National Science Foundation. OSG is responsible for operating a secure
production-quality distributed infrastructure, a reference software stack including the Virtual Data Toolkit (VDT),
extending the capabilities of the high throughput virtual...
Thomas Paul
(Northeastern University)
03/09/2007, 14:40
The Pierre Auger Observatory aims to discover the nature and
origins of the highest energy cosmic rays. The large number of
physicists involved in the project and the diversity of simulation
and reconstruction tasks pose a challenge for the offline analysis
software, not unlike the challenges confronting software for very large
high energy physics experiments. Previously we have...
Dr
Stuart Paterson
(CERN)
03/09/2007, 15:00
Distributed data analysis and information management
oral presentation
The LHCb distributed data analysis system consists of the Ganga job submission
front-end and the DIRAC Workload and Data Management System. Ganga is jointly
developed with ATLAS and allows LHCb users to submit jobs on several backends
including: several batch systems, LCG and DIRAC. The DIRAC API provides a
transparent and secure way for users to run jobs to the Grid and is the default...
James William Monk
(Department of Physics and Astronomy - University College London)
03/09/2007, 15:00
The Durham HepData database has for many years provided an up-to-date archive of
published numerical data from HEP experiments worldwide. In anticipation of the
abundance of new data expected from the LHC, the database is undergoing a complete
metamorphosis to add new features and improve the scope for use of the database by
external applications. The core of the HepData restructuring is...
Tadashi Maeno
(Brookhaven National Laboratory)
03/09/2007, 15:00
A new distributed software system was developed in the fall of 2005 for the ATLAS
experiment at the LHC. This system, called PanDA, provides an integrated service
architecture with late binding of jobs, maximal automation through layered services,
tight binding with ATLAS distributed data management (DDM) system, advanced error
discovery and recovery procedures, and other features. In this...
Prof.
Richard McClatchey
(University of the West of England)
03/09/2007, 15:00
The Health-e-Child (HeC) project is an EC Framework Programme 6 Integrated Project
that aims at developing an integrated healthcare platform for paediatrics. Through
this platform biomedical informaticians will integrate heterogeneous data and perform
epidemiological studies across Europe.
The main objective of the project is to gain a comprehensive view of a child's health
by...
Dr
Jeremy Coles
(RAL)
03/09/2007, 15:00
Computer facilities, production grids and networking
oral presentation
Over the last few years, UK research centres have provided significant computing
resources for many high-energy physics collaborations under the guidance of the
GridPP project. This paper reviews recent progress in the Grid deployment and
operations area including findings from recent experiment and infrastructure service
challenges. These results are discussed in the context of how GridPP...
Dr
Maria Grazia Pia
(INFN Genova)
03/09/2007, 15:20
Computational tools originating from high energy physics developments provide
solutions to common problems in other disciplines: this study presents quantitative
results concerning the application of HEP simulation and analysis tools, and of the
grid technology, to dosimetry for oncological radiotherapy.
The study concerned all the three major radiotherapy techniques: therapy...
Dr
Pavel Murat
(Fermilab)
03/09/2007, 15:20
Computer facilities, production grids and networking
oral presentation
CDFII detector at Fermilab is taking physics data since 2002.
The architechture of the CDF computing system has substantially
evolved during the years of the data taking and currently it reached stable
configuration which will allow experiment to process and analyse the data
until the end of Run II.
We describe major architechtural components of the CDF offline
computing - dedicated...
Dr
Johannes Elmsheuser
(Ludwig-Maximilians-Universität München)
03/09/2007, 15:20
Distributed data analysis and information management
oral presentation
The distributed data analysis using Grid resources is one of the
fundamental applications in high energy physics to be addressed
and realized before the start of LHC data taking. The needs to
manage the resources are very high. In every experiment up to a
thousand physicist will be submitting analysis jobs into the Grid.
Appropriate user interfaces and helper applications have to be
made...
Dr
Alberto Di Meglio
(CERN)
03/09/2007, 15:20
The ETICS system is a distributed software configuration, build and test system
designed to fulfill the needs to improve the quality, reliability and
interoperability of distributed software in general and grid software in
particular. The ETICS project is a consortium of five partners (CERN, INFN,
Engineering Ingegneria Informatica, 4D Soft and the University of Wisconsin-
Madison)....
Dr
Frank Gaede
(DESY IT)
03/09/2007, 15:20
The International Linear Collider is the next large accelerator project in
High Energy Physics. The Large Detector Concept (LDC) study is one of four
international working groups that are developing a detector concept for the
ILC. The LDC uses a modular C++ application framework (Marlin) that is
based on the international data format LCIO. It allows the distributed
development of...
Mr
Lars Fischer
(Nordic Data Grid Facility)
03/09/2007, 15:40
Computer facilities, production grids and networking
oral presentation
The Tier-1 facility operated by the Nordic DataGrid Facility (NDGF) differs
significantly from other Tier-1s in several aspects: It is not located one or a few
locations but instead distributed throughout the Nordic, it is not under the
governance of a single organization but instead a "virtual" Tier-1 build out of
resources under the control of a number of different national...
Giulio Eulisse
(Northeastern University)
03/09/2007, 15:40
We describe a relatively new effort within CMS to converge on a set of web based
tools, using state of the art industry techniques, to engage with the CMS offline
computing system. CMS collaborators require tools to monitor various components of
the computing system and interact with the system itself. The current state of the
various CMS web tools is described along side current planned...
Mr
Adam Kocoloski
(MIT)
03/09/2007, 15:40
Distributed data analysis and information management
oral presentation
Modern Macintosh computers feature Xgrid, a distributed computing architecture built
directly into Apple's OS X operating system. While the approach is radically
different from those generally expected by the Unix based Grid infrastructures (Open
Science Grid, TeraGrid, EGEE), opportunistic computing on Xgrid is nonetheless a
tempting and novel way to assemble a computing cluster with a...
Dr
Lee Lueking
(FERMILAB)
03/09/2007, 15:40
Distributed data analysis and information management
oral presentation
The CMS Dataset Bookkeeping Service (DBS) has been developed to catalog all CMS event
data from Monte Carlo and Detector sources. It includes the ability to identify MC or
trigger source, track data provenance, construct datasets for analysis, and discover
interesting data. CMS requires processing and analysis activities at various service
levels and the system provides support for...
Dr
Patrick Fuhrmann
(DESY)
03/09/2007, 16:30
With the start of the Large Hardron Collider at CERN, end of 2007, the associated
experiments will feed the major share of their data into the dCache Storage
Element technology at most of the Tier I centers and many of the Tier IIs
including the larger sites.
For a project, not having its center of gravity at CERN, and receiving contributions
from various loosely coupled sites in...
Leandro Franco
(CERN)
03/09/2007, 16:30
Distributed data analysis and information management
oral presentation
Particle accelerators produce huge amounts of information in every
experiment and such quantity cannot be stored easily in a personal
computer. For that reason, most of the analysis is done using remote
storage servers (this will be particularly true when the Large Hadron
Collider starts its operation in 2007). Seeing how the bandwidth has
increased in the last few years, the biggest...
Dr
Richard Mount
(SLAC)
03/09/2007, 16:30
Computer facilities, production grids and networking
oral presentation
The PetaCache project started at SLAC in 2004 with support from DOE
Computer Science and the SLAC HEP program. PetaCache focuses on using
cost-effective solid state storage for the hottest data under analysis. We chart
the evolution of metrics such as accesses per second per dollar for different
storage technologies and deduce the near inevitability of a massive use of solid-
state...
Prof.
Gordon Watts
(University of Washington)
03/09/2007, 16:30
The DZERO experiment records proton-antiproton collisions at the Fermilab
Tevatron collider. The DZERO Level 3 data acquisition (DAQ) system is required
to transfer event fragments of approximately 1-20 kilobytes from
63 VME crate sources to any of approximately 240 processing nodes at a rate
of 1 kHz. It is built upon a Cisco 6509 Ethernet switch, standard PCs, and
commodity VME...
Dr
Ivana Hrivnacova
(IPN)
03/09/2007, 16:30
The Virtual Monte Carlo (VMC) provides the abstract interface
into the Monte Carlo transport codes: Geant3, Geant4 and Fluka.
The user VMC based application, independent from the specific
Monte Carlo codes, can be then run with all three simulation programs.
The VMC has been developed by the ALICE Offline Project and since
then it draw attention in more experimental...
Dr
Gerd Behrmann
(Nordic Data Grid Facility)
03/09/2007, 16:50
The LCG collaboration is encompased by a number of Tier 1 centers. The nordic LCG
Tier 1 is in contrast to other Tier 1 centers distributed over most of Scandinavia. A
distributed setup was chosen for both political and technical reasons, but also
provides a number of unique challenges. dCache is well known and respected as a
powerfull distributed storage resource manager, and was chosen...
Dr
Tsukasa Aso
(Toyama National College of Maritime Technology, JST CREST)
03/09/2007, 16:50
The GEANT4 Monte Carlo code provides many powerful functions for conducting
particle transport simulations with great reliability and flexibility. GEANT4 has
been extending the application fields for not only the high energy physics but
also medical physics. Using the reliable simulation for the radiation therapy, it
will become possible to validate treatment planning and select the...
Dr
Giuseppe Lo Presti
(CERN/INFN)
03/09/2007, 16:50
Computer facilities, production grids and networking
oral presentation
In this paper we present the architecture design of the CERN Advanced Storage system
(CASTOR) and its new disk cache management layer (CASTOR2).
Mass storage systems at CERN have evolved over time to meet growing requirements,
both in terms of scalability and fault resiliency. CASTOR2 has been designed as a
Grid-capable storage resource sharing facility, with a database-centric...
Marco Clemencic
(European Organization for Nuclear Research (CERN))
03/09/2007, 16:50
The COOL project provides software components and tools for the handling of
the LHC experiment conditions data. COOL software development is the result
of a collaboration between the CERN IT Department and Atlas and LHCb, the
two experiments that have chosen it as the base of their conditions database
infrastructure. COOL supports persistency for several relational technologies...
Lassi Tuura
(Northeastern University)
03/09/2007, 16:50
Distributed data analysis and information management
oral presentation
The CMS experiment will need to sustain uninterrupted high reliability, high throughput and very diverse data
transfer activities as the LHC operations start. PhEDEx, the CMS data transfer system, will be responsible for the
full range of the transfer needs of the experiment. Covering the entire spectrum is a demanding task: from the
critical high-throughput transfers between CERN and...
Dr
Simon George
(Royal Holloway)
03/09/2007, 16:50
The High Level Trigger (HLT) of the ATLAS experiment at the Large Hadron Collider
receives events which pass the LVL1 trigger at ~75 kHz and has to reduce the rate to
~200 Hz while retaining the most interesting physics. It is a software trigger and
performs the reduction in two stages: the LVL2 trigger should take ~10 ms and the
Event Filter (EF) ~1 s.
At the heart of the HLT is the...
Dr
Douglas Smith
(Stanford Linear Accelerator Center)
03/09/2007, 17:10
Distributed data analysis and information management
oral presentation
The BaBar high energy experiment has been running for many years now,
and has resulted in a data set of over a petabyte in size, containing
over two million files. The management of this set of data has to
support the requirements of further data production along with a
physics community that has vastly different needs. To support these
needs the BaBar bookkeeping system was developed,...
Dr
Horst Goeringer
(GSI)
03/09/2007, 17:10
Computer facilities, production grids and networking
oral presentation
GSI in Darmstadt (Germany) is a center for heavy ion research
and hosts an Alice Tier2 center.
For the future FAIR experiments at GSI,
CBM and Panda, the planned data rates
will reach those of the current LHC experiments at Cern.
Since more than ten years gStore, the GSI Mass Storage System,
is successfully in operation.
It is a hierarchical storage system with a unique name...
Barbara Martelli
(Italian INFN National Center for Telematics and Informatics (CNAF))
03/09/2007, 17:10
Database replication is a key topic in the LHC Computing GRID environment to allow
processing of data in a distributed environment. In particular LHCb computing model
relies on the LHC File Catalog (LFC). LFC is the database catalog which stores
informations about files spread across the GRID, their logical names and physical
locations of all their replicas. The LHCb computing model...
Prof.
Vladimir Ivantchenko
(CERN, ESA)
03/09/2007, 17:10
Current status of the Standard EM package of the Geant4 toolkit is described.
The precision of simulation results is discussed with the focus on LHC
experiments. The comparisons of the simulation with the experimental data are
shown.
Paul Avery
(University of Florida)
03/09/2007, 17:30
Computer facilities, production grids and networking
oral presentation
UltraLight is a collaboration of experimental physicists and network engineers whose
purpose is to provide the network advances required to enable and facilitate
petabyte-scale analysis of globally distributed data. Existing Grid-based
infrastructures provide massive computing and storage resources, but are currently
limited by their treatment of the network as an external, passive, and...
Andrew Cameron Smith
(CERN)
03/09/2007, 17:30
Distributed data analysis and information management
oral presentation
The LHCb Computing Model describes the dataflow model for all stages in the
processing of real and simulated events and defines the role of LHCb associated Tier1
and Tier2 computing centres. The WLCG ‘dressed rehearsal’ exercise aims to allow LHC
experiments to deploy the full chain of their Computing Models, making use of all
underlying WLCG services and resources, in preparation for real...
Dr
Maria Grazia Pia
(INFN Genova)
03/09/2007, 17:30
A project is in progress for a systematic, quantitative validation of Geant4 physics
models against experimental data.
Due to the complexity of Geant4 physics, the validation of Geant4 hadronic models
proceeds according to a bottom-up approach (i.e. from the lower energy range up to
higher energies): this approach, which is different from the one adopted in the LCG
Simulation Validation...
Mr
Mario Lassnig
(CERN & University of Innsbruck, Austria)
03/09/2007, 17:30
The ATLAS detector at CERN's Large Hadron Collider presents data handling requirements on an unprecedented
scale. From 2008 on the ATLAS distributed data management system (DQ2) must manage tens of petabytes of
event data per year, distributed globally via the LCG, OSG and NDGF computing grids, now known as the WLCG.
Since its inception in 2005 DQ2 has continuously managed all datasets...
Mr
Michael DePhillips
(BROOKHAVEN NATIONAL LABORATORY)
03/09/2007, 17:30
Database demands resulting from offline analysis and production of data at
The STAR experiment at Brookhaven National Laboratory's Relativistic Heavy-Ion
Collider has
steadily increased over the last 6 years of data taking activities. With each year
STAR more than doubles events taken with an anticipation of reaching a billion event
capabilities as early as next year. The challenges...
Leonard Apanasevich
(University of Chicago at Illinois)
03/09/2007, 17:40
The High Level Trigger (HLT) that runs in the 1000 dual-CPU box Filter
Farm of the CMS experiment is a set of sophisticated software tools
for selecting a very small fraction of interesting events in real
time. The coherent tuning of these algorithms to accommodate multiple
physics channels is a key issue for CMS, one that literally defines
the reach of the experiment's physics program....
Ms
Alessandra Forti
(University of Manchester)
03/09/2007, 17:50
Computer facilities, production grids and networking
oral presentation
The HEP department of the University of Manchester has purchased a 1000
nodes cluster. The cluster is dedicated to run EGEE and LCG software and is currently
supporting 12 active VOs. Each node is equipped with
2x250 GB disks for a total amount of 500 GB and there is no tape storage behind nor
raid arrays are used. Three different storage solutions are
currently being deployed to...
Wolfgang Ehrenfeld
(Univ. of Hamburg/DESY)
03/09/2007, 17:50
The simulation of the ATLAS detector is largely dominated by the
showering of electromagnetic particles in the heavy parts of the
detector, especially the electromagnetic barrel and endcap
calorimeters. Two procedures have been developed to accelerate the
processing time of EM particles in these regions: (1) a fast shower
parameterization and (2) a frozen shower library. Both work...
Dr
Caitriana Nicholson
(University of Glasgow)
03/09/2007, 17:50
The ATLAS Tag Database is an event-level metadata system, designed to
allow efficient identification and selection of interesting events for
user analysis. By making first-level cuts using queries on a relational
database, the size of an analysis input sample could be greatly reduced
and thus the time taken for the analysis reduced. Deployment of such a
Tag database is underway, but to be...
Dr
Syed Naqvi
(CoreGRID Network of Excellence)
03/09/2007, 17:50
Security requirements of service oriented architectures (SOA) are reasonably higher
than the classical information technology (IT) architectures. Loose coupling – the
inherent benefit of SOA – stipulates security as a service so as to circumvent tight
binding of the services. The services integration interfaces are developed with
minimal assumptions between the sending and receiving...
Teresa Maria Fonseca Martin
(CERN)
03/09/2007, 17:55
The ATLAS experiment under construction at CERN is due to begin operation at the end
of 2007. The detector will record the results of proton-proton collisions at a
centre-of-mass energy of 14 TeV. The trigger is a three-tier system designed to
identify in real-time potentially interesting events that are then saved for detailed
offline analysis. The trigger system will select...
Dr
Ian Fisk
(FNAL)
04/09/2007, 11:00
Computer facilities, production grids and networking
oral presentation
In preparation for the start of the experiment, CMS has conducted computing, software, and analysis challenges to
demonstrate the functionality, scalability, and useability of the computing and software components. These
challenges are designed to validate the CMS distributed computing model by demonstrating the functionality of
many components simultaneously. In the challenges CMS...
Dr
Marianne Bargiotti
(European Organization for Nuclear Research (CERN))
04/09/2007, 11:00
The DIRAC Data Management System (DMS) relies on both WLCG Data Management services
(LCG File Catalogues, Storage Resource Managers and FTS) and LHCb specific components
(Bookkeeping Metadata File Catalogue).
The complexity of both the DMS and its interactions with numerous WLCG components as
well as the instability of facilities concerned, has turned frequently into
unexpected problems...
Dr
Roger Jones
(LANCAS)
04/09/2007, 11:00
Distributed data analysis and information management
oral presentation
The ATLAS Computing Model was constructed after early tests and was captured in the ATLAS Computing TDR in
June 2005. Since then, the grid tools and services have evolved and their performance is starting to be understood
through large-scale exercises. As real data taking becomes immanent, the computing model continues to evolve,
with robustness and reliability being the watchwords for...
Dr
Simone Pagan Griso
(University and INFN Padova)
04/09/2007, 11:20
Distributed data analysis and information management
oral presentation
The upgrades of the Tevatron collider and of the CDF detector have considerably
increased the demand on computing resources in particular for Monte Carlo production
for
the CDF experiment. This has forced the collaboration to move beyond the usage of
dedicated resources and start exploiting Grid resources.
The CDF Analysis Farm (CAF) model has been reimplemented into
LcgCAF ...
Mr
Michel Jouvin
(LAL / IN2P3)
04/09/2007, 11:20
Computer facilities, production grids and networking
oral presentation
Quattor is a tool aimed at efficient management of fabrics with hundred or
thousand of Linux machines, still being easy enough to manage smaller
clusters. It has been originally developed inside the European Data Grid (EDG)
project. It is now in use at more than 30 grid sites running gLite middleware,
ranging from small LCG T3 to very large one like CERN.
Main goals and specific...
Prof.
Shahram Rahatlou
(Univ di Roma La Sapienza), Dr
Tommaso Boccali
(INFN Sezione di Pisa)
04/09/2007, 11:20
At the end of 2007 the first colliding beams from LHC are expected. The CMS Computing
model enforces the use of the same software (with different performance settings) for
offline and online(HLT) operations; this is particularly true for the reconstruction
software: the different settings must allow a processing time per event
(typically, numbers for 2x10e33 luminosity are given) of 50 ms...
Swagato Banerjee
(University of Victoria)
04/09/2007, 11:40
BaBar Abstract #8 - Track 2 (Event processing)
Experience with validating GEANT4 v7 and v8 against v6 in BaBar
S. Banerjee, P. Kim, W. Lockman, and D. Wright for the BaBar Computing Group
The BaBar experiment at SLAC has been using the GEANT 4 package
version 6 for simulation of the detector response to passage of
particles through its material.
Since 2005 and 2006, respectively,...
Torsten Antoni
(Forschungszentrum Karlsruhe)
04/09/2007, 11:40
Computer facilities, production grids and networking
oral presentation
The organization and management of the user support in a global e-science computing
infrastructure such as EGEE is one of the challenges of the grid. Given the widely
distributed nature of the organisation, and the spread of expertise for installing,
configuring, managing and troubleshooting the grid middleware services, a standard
centralized model could not be deployed in EGEE. This...
Dr
Hartmut Stadie
(Universitaet Hamburg)
04/09/2007, 11:40
Distributed data analysis and information management
oral presentation
The detector and collider upgrades for the HERA-II running at DESY have considerably
increased the demand on computing resources for the ZEUS experiment.
To meet the demand, ZEUS commissioned an automated Monte Carlo(MC) production capable
of using Grid resources in November 2004. Since then, more than one billion events
have been simulated and reconstructed on the Grid which corresponds...
Mr
Philippe Canal
(FERMILAB)
04/09/2007, 11:50
For the last several months the main focus of development
in the ROOT I/O package has been code consolidation and
performance improvements.
Access to remote files is affected both by bandwidth and
latency. We introduced a pre-fetch mechanism to minimize
the number of transactions between client and server and
hence reducing the effect of latency. We will review the...
Dr
Ashok Agarwal
(University of Victoria)
04/09/2007, 12:00
Distributed data analysis and information management
oral presentation
The present paper highlights the approach used to design and implement a web services
based BaBar Monte Carlo (MC) production grid using Globus Toolkit version 4. The grid
integrates the resources of two clusters at the University of Victoria, using the
ClassAd mechanism provided by the Condor-G metascheduler. Each cluster uses the
Portable Batch System (PBS) as its local resource...
Mr
Luigi Zangrando
(INFN Padova)
04/09/2007, 12:00
Modern GRID middlewares are built around components providing basic
functionality, such as data storage, authentication and security, job
management, resource monitoring and reservation. In this paper we
describe the Computing Resource Execution and Management (CREAM)
service. CREAM provides a Web service-based job execution and
management capability for Grid systems; in particular, it is...
Mr
Antonio Retico
(CERN)
04/09/2007, 12:00
Computer facilities, production grids and networking
oral presentation
Grids have the potential to revolutionise computing by providing ubiquitous, on
demand access to computational services and resources. They promise to allow for on
demand access and composition of computational services provided by multiple
independent sources. Grids can also provide unprecedented levels of parallelism for
high-performance applications. On the other hand, grid...
Mr
Andrei Gheata
(CERN/ISS)
04/09/2007, 12:05
The ROOT geometry modeller (TGeo) offers powerful tools for detector geometry
description. The package provides several functionalities like: navigation, geometry
checking, enhanced visualization, geometry editing GUI and many others, using ROOT
I/O. A new interface module g4root was recently developed to take advantage of ROOT
geometry navigation optimizations in the context of GEANT4...
04/09/2007, 14:05
oral presentation
Effective security needs resources and support from senior management. This session will look at some ways of gaining that support by establishing a common understanding of risk.
04/09/2007, 15:00
This session will look to establish a common understanding of risk and introduce the ISSeG risk assessment questionnaire.
04/09/2007, 16:30
This session will look at some of the emerging recommendations that can be used at sites to improve security.
Marco Mambelli
(University of Chicago)
05/09/2007, 08:00
A Data Skimming Service (DSS) is a site-level service for rapid event filtering and
selection from locally resident datasets based on metadata queries to associated
"tag" databases. In US ATLAS, we expect most if not all of the AOD-based datasets to
be be replicated to each of the five Tier 2 regional facilities in the US Tier 1
"cloud" coordinated by Brookhaven National Laboratory. ...
Dr
Torsten Harenberg
(University of Wuppertal)
05/09/2007, 08:00
Today, one of the major challenges in science is the processing of large datasets.
The LHC experiments will produce an enormous amount of results that are
stored in databases or files. These data are processed by a large
number of small jobs that read only chunks.
Existing job monitoring tools inside the LHC Computing Grid (LCG) provide
just limited functionality to the user.
These...
Dr
Silvio Pardi
(University of Naples ``Federico II'' - C.S.I. and INFN)
05/09/2007, 08:00
The user interface is a crucial service to guarantee the Grid accessibility. The goal
to achieve, is the implementation of an environment able to hide the grid complexity
and offer a familiar interface to the final user.
Currently many graphical interfaces have been proposed to simplify the grid access,
but the GUI approach appears not very congenital to UNIX developers and...
Valentin Kuznetsov
(Cornell University)
05/09/2007, 08:00
The CMS Dataset Bookkeeping System (DBS) search page is a
web-based application used by physicists and production managers
to find data from the CMS experiment. The main challenge in the
design of the system was to map the complex, distributed data
model embodied in the DBS and the Data Location Service (DLS) to
a simple, intuitive interface consistent with the mental model...
Mr
Giacinto Piacquadio
(Physikalisches Institut - Albert-Ludwigs-Universität Freiburg)
05/09/2007, 08:00
A new inclusive secondary vertexing algorithm which exploits the topological
structure of weak b- and c-hadron decays inside jets is presented. The primary goal
is the application to b-jet tagging. The fragmentation of a b-quark results in a
decay chain composed of a secondary vertex from the weakly decaying b-hadron and
typically one or more tertiary vertices from c-hadron decays. The...
Dr
Sebastien Incerti
(CENBG-IN2P3)
05/09/2007, 08:00
Detailed knowledge of the microscopic pattern of energy deposition related to the
particle track structure is required to study radiation effects in various domains,
like electronics, gaseous detectors or biological systems.
The extension of Geant4 physics down to the electronvolt scale requires not only new
physics models, but also adequate design technology. For this purpose a...
Mr
Pablo Martinez
(Insitituto de Física de Cantabria)
05/09/2007, 08:00
A precise alignment of Muon System is one of the requirements to fulfill the CMS
expected performance to cover its physics program. A first prototype of the
software and computing tools to achieve this goal has been successfully tested
during the CSA06, Computing, Software and Analysis Challenge in 2006. Data was
exported from Tier-0 to Tier-1 and Tier-2, where the alignment software...
Dr
Josva Kleist
(Nordic Data Grid Facility)
05/09/2007, 08:00
AliEn or Alice Environment is the Gridware developed and used within the ALICE
collaboration for storing and processing data in a distributed manner. ARC (Advanced
Resource Connector) is the Grid middleware deployed across the Nordic countries and
gluing together the resources within the Nordic Data Grid Facility (NDGF). In this
paper we will present our approach to integrate AliEn and...
Mr
Luca Magnoni
(INFN-CNAF)
05/09/2007, 08:00
In a Grid environment the naming capability allows users to refer to specific data
resources in a physical storage system using a high level logical identifier. This
logical identifier is typically organized in a file system like structure, a
hierarchical tree of names. Storage Resource Manager (SRM) services map the logical
identifier to the physical location of data evaluating a set of...
Stephane Chauvie
(INFN Genova)
05/09/2007, 08:00
An original model is presented for the simulation of the energy loss of negatively
charged hadrons: it calculates the stopping power by regarding the target atoms as an
ensemble of quantum harmonic oscillators.
This approach allows to account for charge dependent effects in the stopping power,
which are relevant at low energy: the differences between the stopping powers of
positive and...
Dr
Jerome Lauret
(BROOKHAVEN NATIONAL LABORATORY)
05/09/2007, 08:00
Secure access to computing facilities has been increasingly on demand of practical
tools as the world of cyber-security infrastructure has changed the landscape to
access control via gatekeepers or gateways. However, the venue of two factor
authentication (SSH keys for example) preferred over simpler Unix based login has
introduced the challenging task of managing private keys and its...
Dr
Josva Kleist
(Nordic Data Grid Facility)
05/09/2007, 08:00
The Nordic Data Grid Facility (NDGF) consists of Grid resources running ARC
middleware in Scandinavia and other countries. These resources serve many virtual
organisations and contribute a large fraction of total worldwide resources for the
ATLAS experiment, whose data is distributed and managed by the DQ2 software. Managing
ATLAS data within NDGF and between NDGF and other Grids used by...
Rolf Seuster
(University of Victoria)
05/09/2007, 08:00
The ATLAS Liquid Argon Calorimter consists of precision electromagnetic
accordion calorimeters in the barrel and endcaps, hadronic calorimeters
in the endcaps, and calorimeters in the forward region.
The initial high energy collision data at the LHC experiments is
expected in the spring of 2008. While tools for the reconstruction of
the calorimeter data are quite developed through years...
Dr
Daniela Rebuzzi
(INFN, Sezione di Pavia), Dr
Nectarios Benekos
(Max-Planck-Institut fur Physik)
05/09/2007, 08:00
The ATLAS detector, currently being installed at CERN, is designed
to make precise measurements of 14 TeV proton-proton collisions at
the LHC, starting in 2007. Arguably the clearest signatures for
new physics, including the Higgs Boson and supersymmetry, will involve
the production of isolated final-stated muons. The identification and
precise reconstruction of muons are performed using...
Dr
Ricardo Vilalta
(University of Houston)
05/09/2007, 08:00
Advances in statistical learning have placed at our disposal a rich set of
classification algorithms (e.g., neural networks, decision trees, Bayesian
classifiers, support vector machines, etc.) with little or no guidelines on how to
select the analysis technique most appropriate for the task at hand. In this paper we
present a new approach for the automatic selection of predictive models...
Dr
Andrew McNab
(University of Manchester)
05/09/2007, 08:00
GridSite has extended the industry-standard Apache webserver for use within Grid
projects, by adding support for Grid security credentials such as GSI and VOMS. With
the addition of
the GridHTTP protocol for bulk file transfer via HTTP and the development of a mapping
between POSIX filesystem operations and HTTP requests we have extended this scope of
GridSite into bulk data transfer and...
Dr
Douglas Benjamin
(Duke University)
05/09/2007, 08:00
The CDF experiment at Fermilab produces Monte Carlo data files using computing
resources on both the Open
Science Grid (OSG) and LHC Computing Grid (LCG) grids. This data produced must be
brought back to Fermilab
for archival storage. In the past CDF produced Monte Carlo data on dedicated
computer farms through out
the world. The data files were copied directly from the worker nodes to...
Dr
Daniele Bonacorsi
(INFN-CNAF, Bologna, Italy)
05/09/2007, 08:00
The CMS experiment operated a Computing, Software and Analysis Challenge in 2006 (CSA06). This activity is part
of the constant work of CMS in computing challenges of increasing complexity to demonstrate the capability to
deploy and operate a distributing computing system at the desired scale in 2008. The CSA06 challenge was a 25%
exercise, and included several workflow elements: event...
Dr
Andreas Nowack
(III. Physikalisches Institut (B), RWTH Aachen)
05/09/2007, 08:00
In Germany, several university institutes and research centres take
part in the CMS experiment. Concerning the data analysis, a couple of
computing centres at different Tier levels, ranging from Tier 1 to
Tier 3, exists at these places. The German Tier 1 centre GridKa at the
research centre at Karlsruhe serves all four LHC experiments as
well as for four non-LHC experiments. With respect...
Prof.
Alexander Read
(University of Oslo, Department of Physics)
05/09/2007, 08:00
Computing and storage resources connected by the Nordugrid ARC middleware in the
Nordic countries, Switzerland and Slovenia are a part of the ATLAS computing grid.
This infrastructure is being commissioned with the ongoing ATLAS Monte Carlo
simulation production in preparation for the commencement of data taking in late
2007. The unique non-intrusive architecture of ARC, it's...
Prof.
Richard McClatchey
(UWE)
05/09/2007, 08:00
We introduce the concept, design and deployment of the DIANA meta-scheduling approach
to solving the challenge of the data analysis being faced by the CERN experiments.
The DIANA meta-scheduler supports data intensive bulk scheduling, is network aware
and follows a policy centric meta-scheduling that will be explained in some detail.
In this paper, we describe a Physics analysis case...
Dr
Domenico Giordano
(Dipartimento Interateneo di Fisica)
05/09/2007, 08:00
The CMS Silicon Strip Tracker (SST), consisting of more than 10 millions of channels,
is organized in about 16,000 detector modules and it is the largest silicon strip
tracker ever built for high energy physics experiments.
In the first half of 2007 the CMS SST project is facing the important milestone of
commissioning and testing a quarter of the entire SST with cosmic muons.
The full...
Mr
Tigran Mkrtchyan Mkrtchyan
(Deutsches Elektronen-Synchrotron DESY)
05/09/2007, 08:00
Starting June 2007, all WLCG data management services have to be ready and prepared
to move terabytes of data from CERN to the Tier 1 centers world wide, and from the
Tier 1s to their corresponding Tier 2s. Reliable file transfer services, like FTS, on
top of the SRM v2.2 protocol are playing a major role in this game. Nevertheless,
moving large junks of data is only part of the...
Mr
Enrico Fattibene
(INFN-CNAF, Bologna, Italy), Mr
Giuseppe Misurelli
(INFN-CNAF, Bologna, Italy)
05/09/2007, 08:00
A monitoring tool for complex Grid systems can gather a huge amount of information
that have to be presented to the users in the most comprehensive way. Moreover
different types of consumers could be interested in inspecting and analyzing
different subsets of data. The main goal in designing a Web interface for the
presentation of monitoring information is to organize the huge amount of...
Dr
Ricardo Graciani Diaz
(Universidad de Barcelona)
05/09/2007, 08:00
DIRAC Services and Agents are defined in the context of the DIRAC system (the LHCb's
Grid Workload and Data Management system), and how they cooperate to build functional
sub-systems is presented. How the Services and Agents are built from the low level
DIRAC framework tools is described.
Practical experiente in the LHCb production system has directed the creation of the
current DIRAC...
Mr
Adrian Casajus Ramo
(Universitat de Barcelona)
05/09/2007, 08:00
The DIRAC system is made of a number of cooperating Services and Agents that interact
between them with a Client-Server architecture. All DIRAC components rely on a low
level framework that provides the necessary basic functionality.
In the current version of DIRAC these components have been identified as: DISET, the
secure communication protocol for remote procedure call and file...
Gianluca Castellani
(European Organization for Nuclear Research (CERN))
05/09/2007, 08:00
LHCb accesses Grid through DIRAC, its WorkLoad and Data Management
system.
In DIRAC all the jobs are stored in central task queues and then pulled onto
worker nodes via generic Grid jobs called Pilot Agents. These task queues are
characterized by different requirements about CPUtime and destination.
Because the whole LHCb community is divided in sets of physicists, developers,...
Dr
Andrei Tsaregorodtsev
(CNRS-IN2P3-CPPM, Marseille)
05/09/2007, 08:00
The DIRAC system was developed in order to provide a complete solution
for using distributed computing resources of the LHCb experiment at CERN
for data production and analysis. It allows a concurrent use of over 10K CPUs and
10M file replicas distributed over many tens of sites. The sites can be part of a
computing grid such as WLCG or standalone computing clusters all integrated in a...
Andrew Cameron Smith
(CERN)
05/09/2007, 08:00
DIRAC, LHCb’s Grid Workload and Data Management System, utilises WLCG resources and
middleware components to perform distributed computing tasks satisfying LHCb’s
Computing Model. The Data Management System (DMS) handles data transfer and data
access within LHCb. Its scope ranges from the output of the LHCb Online system to
Grid-enabled storage for all data types. It supports metadata for...
Mr
Claude Charlot
(Ecole Polytechnique)
05/09/2007, 08:00
We describe the strategy developed for electron reconstruction in CMS. Emphasis is
put on isolated electrons and on recovering the bremsstrahlung losses due to the
presence of the material before the ECAL. Following the strategy used for the high
level triggers, a first
filtering is obtained building seeds from the clusters reconstructed in the ECAL. A
dedicated trajectory building is...
Dr
Vincenzo Ciaschini
(INFN CNAF)
05/09/2007, 08:00
While starting to use the grid in production, applications have begun to demand the
implementation of complex policies regarding the use of resources. Some want to
divide their users in different priority brackets and classify the resources in
different classes, others again content themselves with considering all users and
resources equal. Resource managers have to work into enabling...
Mr
Sergey Gorbunov
(GSI), Dr
alexander glazov
(DESY)
05/09/2007, 08:00
Stand-alone event reconstruction was developed for the Forward and the
Backward Silicon Trackers of the H1 experiment at HERA. The
reconstruction module includes the pattern recognition algorithm, a
track fitter and primary vertex finder. The reconstruction
algorithm shows high efficiency and speed. The detector alignment
was performed to within an accuracy of 10 um which...
Mr
Trunov Artem
(CC-IN2P3 (Lyon) and EKP (Karlsruhe))
05/09/2007, 08:00
We present our experience in setting up an xrootd storage cluster at CC-IN2P3 - a LCG
Tier-1 computing Center. The solution consists of xrootd storage cluster made of NAS
boxes and includes an interface to dCache/SRM, and Mass Storage System. The feature
of this system is integration of PROOF for facilitation of analysis. The setup allows
to take advantage of ease of administrative burden,...
Mr
Kyu Park
(Department of Electrical and Computer Engineering, University of Florida)
05/09/2007, 08:00
A primary goal of the NSF-funded UltraLight Project is to expand existing
data-intensive grid computing infrastructures to the next level by enabling a managed
network that provides dynamically constructed end-to-end paths (optically or
virtually, in whole or in part). Network bandwidth used to be the primary limiting
factor, but with the recent advent of 10Gb/s network paths end-to-end,...
296.
Extension of the DIRAC workload-management system to allow use of distributed Windows resources
Ms
Ying Ying Li
(University of Cambridge)
05/09/2007, 08:00
The DIRAC workload-management system of the LHCb experiment allows
coordinated use of globally distributed computing power and data storage. The
system was initially deployed only on Linux platforms, where it has been used
very successfully both for collaboration-wide production activities and for single-
user physics studies. To increase the resources available to LHCb, DIRAC has...
Dr
Klaus Goetzen
(GSI Darmstadt)
05/09/2007, 08:00
As one of the primary experiments to be located at the new Facility for Antiproton
and Ion Research in Darmstadt the PANDA experiment
aims for high quality hadron spectroscopy from antiproton proton collisions.
The versatile and comprehensive projected physics program requires an elaborate
detector design. The detector for the PANDA experiment will be a very complex machine
consisting of...
Dr
Manuel Venancio Gallas Torreira
(CERN)
05/09/2007, 08:00
Based on the ATLAS TileCal 2002 test-beam setup example, we present
here the technical, software aspects of a possible solution to the
problem of using two different simulation engines, like Geant4 and
Fluka, with the common geometry and digitization code. The specific
use case we discuss here, which is probably the most common one, is
when the Geant4 application is already implemented....
Mr
Edmund Widl
(Institut für Hochenergiephysik (HEPHY Vienna))
05/09/2007, 08:00
The Kalman alignment algorithm (KAA) has been specifically developed to cope with the
demands that arise from the specifications of the CMS Tracker. The algorithmic
concept is based on the Kalman filter formalism and is designed to avoid the
inversion of large matrices.
Most notably, the KAA strikes a balance between conventional global and local
track-based alignment algorithms, by...
Ted Hesselroth
(Fermi National Accelerator Laboratory)
05/09/2007, 08:00
gPlazma is the authorization mechanism for the distributed storage system dCache.
Clients are authorized based on a grid proxy and may be allowed various privileges
based on a role contained in the proxy. Multiple authorization mechanisms may be
deployed through gPlazma, such as legacy dcache-kpwd, grid-mapfile, grid-vorolemap,
or GUMS. Site-authorization through SAZ is also supported....
Alexandre Vaniachine
(Argonne National Laboratory)
05/09/2007, 08:00
To process the vast amount of data from high energy physics experiments, physicists rely
on Computational and Data Grids; yet, the distribution, installation, and updating of a
myriad of different versions of different programs over the Grid environment is
complicated, time-consuming, and error-prone.
We report on the development of a Grid Software Installation Management Framework...
Ms
Alessandra Forti
(University of Manchester)
05/09/2007, 08:00
System Management Working Group (SMWG) of sys admins from Hepix and
grid sites has been setup to address the fabric management problems
that HEP sites might have. The group is open and its goal is
not to implement new tools but to share what is already in use at
sites according to existing best practices. Some sites are already
publicly sharing their tools and sensors and some other...
Prof.
Nobuhiko Katayama
(High Energy Accelerator Research Organization)
05/09/2007, 08:00
The Belle experiment operates at the KEKB accelerator, a high luminosity
asymmetric energy e+ e- collider. The Belle collaboration studies CP violation in
decays of B meson to answer one of the fundamental questions of Nature, the
matter-anti-matter asymmetry. Currently, Belle accumulates more than one
million B Bbar meson pairs that correspond to about 1.2 TB of raw data in one...
Alfonso Mantero
(INFN Genova)
05/09/2007, 08:00
A component of the Geant4 toolkit is responsible for the simulation of atomic
relaxation: it is part of a modelling approach of electromagnetic interactions that
takes into account the detailed atomic structure of matter, by describing particle
interactions at the level of the atomic shells of the target material.
The accuracy of Geant4 Atomic Relaxation has been evaluated against the...
Dr
Daniela Rebuzzi
(INFN Pavia and Pavia University)
05/09/2007, 08:00
The Atlas Muon Spectrometer is designed to reach a very high transverse momentum
resolution for muons in a pT range extending from 6 GeV/c up to 1 Tev/c. The most
demanding design goal is an overall uncertainty of 50 microns on the sagitta of a
muon with pT = 1 TeV/c. Such precision requires an accurate control of the positions
of the muon detectors and of their movements during the...
Aatos Heikkinen
(Helsinki Institute of Physics, HIP)
05/09/2007, 08:00
We introduce a new implementation of Liege cascade INCL4 with ABLA evaporation in
Geant4.
INCL4 treats hadron, Deuterium, Tritium, and Helium beams up to 3 GeV energy,
while ABLA provides treatment for light evaporation residues.
The physics models in INCL4 and ABLA and are reviewd with focus on recent additions.
Implementation details, such as first version of object oriented...
Timur Perelmutov
(FERMI NATIONAL ACCELERATOR LABORATORY)
05/09/2007, 08:00
The Storage Resource Manager (SRM) and WLCG collaborations recently
defined version 2.2 of the SRM protocol, with the goal of satisfying
the requirement of the LCH experiments. The dCache team has now
finished the implementation of all SRM v2.2 elements required by the
WLCG. The new functions include space reservation, more advanced data
transfer, and new namespace and permission...
Mr
Thomas Doherty
(University of Glasgow)
05/09/2007, 08:00
AMI is an application which stores and allows access to dataset metadata for the ATLAS
experiment. It provides a set of generic tools for managing database applications. It
has a
three-tier architecture with a core that supports a connection to any RDBMS using
JDBC and
SQL. The middle layer assumes that the databases have an AMI compliant self-describing
structure. It provides a...
Dr
Robert Harakaly
(CERN)
05/09/2007, 08:00
Configuration is an essential part of the deployment process of any software product.
In the case of Grid middleware the variety and complexity of grid services coupled
with multiple deployment scenarios make the provision of a coherent configuration
both more important and more difficult. The configuration system must provide a
simple interface which strikes a balance between the...
Dr
Iosif Legrand
(CALTECH)
05/09/2007, 08:00
MonaLISA (Monitoring Agents in A Large Integrated Services Architecture) provides a
distributed service for monitoring, control and global optimization of complex
systems including the grids and networks used by the LHC experiments. MonALISA is
based on an ensemble of autonomous multi-threaded, agent-based subsystems which able
to collaborate and cooperate to perform a wide range of...
Gianluca Castellani
(CERN)
05/09/2007, 08:00
Facilities offered by WLCG are extensively used by LHCb in all aspects of their
computing activity. A real time knowledge of the status of all Grid components
involved is needed to optimize their exploitation. This is achieved by employing
different monitoring services each one supplying a specific overview of the
system. SAME tests are used in LHCb for monitoring the status of CE...
Dr
Sergio Andreozzi
(INFN-CNAF)
05/09/2007, 08:00
GridICE is an open source distributed monitoring tool for Grid systems that is
integrated in the gLite middleware and provides continuous monitoring of the EGEE
infrastructure. The main goals of GridICE are: to provide both summary and detailed
view of the status and availability of Grid resource, to highlight a number of
pre-defined fault situations and to present usage information. In...
Mr
Sylvain Reynaud
(IN2P3/CNRS)
05/09/2007, 08:00
Advanced capabilities available in nowadays batch systems are fundamental for
operators of high-performance computing centers in order to provide a high-
quality service to their local users. Existing middleware allow sites to expose
grid-enabled interfaces of the basic functionalities offered by the site’s
computing service. However, they do not provide enough mechanisms for...
Dr
Graeme Stewart
(University of Glasgow)
05/09/2007, 08:00
When operational, the Large Hadron Collider experiments at CERN will
collect tens of petabytes of physics data per year. The worldwide LHC
computing grid (WLCG) will distribute this data to over two hundred
Tier-1 and Tier-2 computing centres, enabling particle physicists
around the globe to access the data for analysis. Different middleware
solutions exist for effective management of...
Mr
Martin Radicke
(DESY Hamburg)
05/09/2007, 08:00
The dCache software has become a major storage element in the WLCG, providing
high-speed file transfers by caching datasets on potentially thousands of disk
servers in front of tertiary storage. Currently dCache's model of separately
connecting all disk servers to the tape backend leads to locally controlled flush and
restore behavior has shown some inefficiencies in respect of tape drive...
Dr
Marco La Rosa
(The University of Melbourne)
05/09/2007, 08:00
With the proliferation of multi-core x86 processors, it is reasonable to
ask whether the supporting infrastructure of the system (memory
bandwidth, IO bandwidth etc) can handle as many jobs as there are cores.
Furthermore, are traditional benchmarks like SpecINT and SpecFloat
adequate for assessing multi-core systems in real computing situations.
In this paper we present the results of...
Mr
Alexander Kulyavtsev
(FNAL)
05/09/2007, 08:00
dCache is a distributed storage system which today stores and serves
petabytes of data in several large HEP experiments. Resilient dCache
is a top level service within dCache, created to address reliability
and file availability issues when storing data for extended periods of
time on disk-only storage systems. The Resilience Manager
automatically keeps the number of copies within...
Dr
Gregory Dubois-Felsmann
(SLAC)
05/09/2007, 08:00
The BaBar experiment currently uses approximately 4000 KSI2k on
dedicated Tier 1 and Tier 2 compute farms to produce Monte Carlo
events and to create analysis datasets from detector and Monte Carlo
events. This need will double in the next two years requiring
additional resources.
We describe enhancements to the BaBar experiment's distributed system
for the creation of skimmed...
Dr
Maria Grazia Pia
(INFN GENOVA)
05/09/2007, 08:00
Journal publication plays a fundamental role in scientific research, and has
practical effects on researchers’ academic career and towards funding agencies.
An analysis is presented, also based on the author’s experience as a member of the
Editorial Board of a major journal in Nuclear Technology, of publications about high
energy physics computing in refereed journals.
The statistical...
Prof.
Sridhara Dasu
(University of Wisconsin)
05/09/2007, 08:00
We describe the ideas and present performance results from a rapid-response adaptive computing environment
(RACE) that we setup at the UW-Madison CMS Tier-2 computing center. RACE uses Condor technologies to allow
rapid-response to certain class of jobs, while suspending the longer running jobs temporarily. RACE allows us to use
our entire farm for long running production jobs, but also...
Nancy Marinelli
(University of Notre Dame)
05/09/2007, 08:00
A seed/track finding algorithm has been developed for
reconstruction of e+e- from converted photons. It combines
the information of the electromagnetic calorimeter with
the accurate information provided by the tracker.
An Ecal seeded track finding is used to locate the approximate
vertex of the conversion. Tracks found with this method are then
used as input to further inside-out...
Dr
Andy Buckley
(Durham University)
05/09/2007, 08:00
The Rivet system is a framework for validation of Monte Carlo event generators
against archived experimental data, and together with JetWeb and HepData forms a core
element of the CEDAR event generator tuning programme. It is also an essential tool
in the development of next generation event generators by members of the MCnet
network. Written primarily in C++, Rivet provides a uniform...
Emmanuel Ormancey
(CERN)
05/09/2007, 08:00
Nearly every large organization use a tool to broadcast messages and
information across the internal campus (messages like alerts announcing
interruption in services or just information about upcoming events). The tool
typically allows administrators (operators) to send "targeted" messages which
is sent only to specific group of users or computers (for instance only those
ones...
Dr
Gregory Dubois-Felsmann
(SLAC)
05/09/2007, 08:00
The BaBar experiment needs fast and efficient procedure for distributing
jobs to produce a large amount of simulated events for analysis purpose.
We discuss the benefits/drawbacks gained mapping the traditional
production schema on the grid paradigm, and describe the structure
implemented on the standard "public" resources of INFN-Grid project.
Data access/distribution on sites...
Dr
Steven Goldfarb
(University of Michigan)
05/09/2007, 08:00
"Shaping Collaboration 2006" was a workshop held in Geneva, on December 11-13, 2006, to examine the status
and future of collaborative tool technology and its usage for large global scientific collaborations, such as those of the
CERN LHC (Large Hadron Collider). The workshop brought together some of the leading experts in the field of
collaborative tools (WACE 2006) with physicists and...
Dr
Yaodong Cheng
(Institute of High Energy Physics,Chinese Academy of Sciences)
05/09/2007, 08:00
Currently more and more heterogeneous resources are integrated into LCG. Sharing LCG
files across different platforms, including different OS and grid middlewares, is a
basic issue. We implemented web service interface for LFC and simulated LCG file
access client by using globus Java CoG Kit.
Dr
Dorian Kcira
(University of Louvain)
05/09/2007, 08:00
With a total area of more than 200 square meters and about 16000 silicon detectors
the Tracker of the CMS experiment will be the largest silicon detector ever built.
The CMS silicon Tracker will detect charged tracks and will play a determinant role
in lepton reconstruction and heavy flavour quark tagging.
A general overview of the Tracker data handling software, which allows the...
Dr
Paul Miyagawa
(University of Manchester)
05/09/2007, 08:00
The ATLAS solenoid produces a magnetic field which enables the Inner
Detector to measure track momentum by track curvature. This solenoidal
magnetic field was measured using a rotating-arm mapping machine and, after
removing mapping machine effects, has been understood to the 0.05% level.
As tracking algorithms require the field strength at many different points,
the representation of...
Dr
Pavel Nevski
(Brookhaven National Laboratory (BNL))
05/09/2007, 08:00
In order to be ready for the physics analysis ATLAS experiment is running
a world wide
Monte Carlo production for many different physics samples with different
detector conditions.
Job definition is the starting point of ATLAS production system. This is a
common interface for the ATLAS community to submit jobs for processing by
the Distrubuted production system used for all...
Robert Petkus
(Brookhaven National Laboratory)
05/09/2007, 08:00
The RHIC/USATLAS Computing Facility at BNL has evaluated high-performance, low-cost
storage solutions in order to complement a substantial distributed file system
deployment of dCache (>400 TB) and xrootd (>130 TB). Currently, these file systems
are spread across disk-heavy computational nodes providing over 1.3 PB of aggregate
local storage. While this model has proven sufficient to...
Prof.
Vladimir Ivantchenko
(CERN, ESA)
05/09/2007, 08:00
The testing suite for validation of Geant4 hadronic generators with the data of
thin target experiments is presented. The results of comparisons with the
neutron and pion production data of are shown for different Geant4 hadronic
generators for the beam momentum interval 0.5 – 12.9 GeV/c.
Tapio Lampen
(Helsinki Institute of Physics HIP)
05/09/2007, 08:00
We demonstrate the use of a ROOT Toolkit for Multivariate Data
Analysis (TMVA) in tagging b-jets associated with heavy
neutral MSSM Higgs bosons at the LHC.
The associated b-jets can be used to extract Higgs events from the
Drell-Yan background, for which the associated jets are mainly light
quark and gluon jets.
TMVA provides an evaluation for different multivariate
classification...
Suren Chilingaryan
(The Institute of Data Processing and Electronics, Forschungszentrum Karlsruhe)
05/09/2007, 08:00
For the reliable and timely forecasts of dangerous conditions of Space Weather
world-wide networks of particle detectors are located at different latitudes,
longitudes and altitudes. To provide better integration of these networks the DAS
(Data Acquisition System) is facing a challenge to establish reliable data exchange
between multiple network nodes which are often located in hardly...
Dr
Solveig Albrand
(LPSC/IN2P3/UJF Grenoble France)
05/09/2007, 08:00
AMI was chosen as the ATLAS dataset selection interface in July 2006. It should
become the main interface for searching for ATLAS data using physics metadata criteria.
AMI has been implemented as a generic database management framework which allows
parallel searching over many catalogues, which may have differing schema. The main
features of the web interface will be described; in...
Dr
Andy Buckley
(Durham University)
05/09/2007, 08:00
Monte Carlo event generators are an essential tool for modern particle physics; they
simulate aspects of collider events ranging from the parton-level "hard process" to
cascades of QCD radiation in both initial and final states, non-perturbative
hadronization processes, underlying event physics and specific particle decays. LHC
events in particular are so complex that event generator...
Dr
Daniele Bonacorsi
(INFN-CNAF, Bologna, Italy)
05/09/2007, 08:00
Early in 2007 the CMS experiment deployed a traffic load generator infrastructure, aimed at providing CMS
Computing Centers (Tiers of the WLCG) with a means for debugging, load-testing and commissioning data
transfer routes among them. The LoadTest is built upon, and relies on, the PhEDEx dataset transfer tool as a
reliable data replication system in use by CMS. On top of PhEDEx, the CMS...
Dr
Andrew McNab
(University of Manchester)
05/09/2007, 08:00
We describe the operation of www.gridpp.ac.uk, the website provided for GridPP and
its precursor, UK HEP Grid, since 2000, and explain the operational procedures of the
service and the various collaborative tools and components that were adapted or
developed for use on the site. We pay particular attention to the security issues
surrounding such a prominent site, and how the GridSite...
Dr
Raja Nandakumar
(Rutherford Appleton Laboratory)
05/09/2007, 08:00
The worldwide computing grid is essential to the LHC experiments in analysing the
data collected by the detectors. Within LHCb, the computing model aims to simulate
data at Tier-2 grid sites as well as non-grid resources. The reconstruction,
stripping and analysis of the produced LHCb data will primarily place at the Tier-1
centres. The computing data challenge DC06 started in May 2006...
Mr
Rudolf Frühwirth
(Inst. of High Energy Physics, Vienna)
05/09/2007, 08:00
We present the "LiC Detector Toy'' ("LiC'' for Linear Collider) program, a simple
but powerful software tool for detector design, modification and geometry studies. It
allows the user to determine the resolution of reconstructed track parameters for the
purpose of comparing and optimizing various detector set-ups. It consists of a
simplified simulation of the detector measurements, taking...
Mr
Antonio Retico
(CERN)
05/09/2007, 08:00
The WLCG/EGEE Pre-Production Service (PPS) is a grid infrastructure whose
goal is to give early access to new services to WLCG/EGEE users in order to
evaluate new features and changes in the middleware before new versions are
actually deployed in PPS.
The PPS grid counts about 30 sites providing resources and manpower.
The service contributes to the overall quality of the grid...
Dr
Winfried A. Mitaroff
(Institute of High Energy Physics (HEPHY) of the Austrian Academy of Sciences, Vienna)
05/09/2007, 08:00
A detector-independent toolkit (RAVE) is being developed for the reconstruction of the common
interaction vertices from a set of reconstructed tracks. It deals both with "finding" (pattern
recognition of track bundles) and with "fitting" (estimation of vertex position and track momenta).
The algorithms used so far include robust adaptive filters which are derived from the CMS...
Dr
Fabio Cossutti
(INFN)
05/09/2007, 08:00
The CMS Collaboration has developed a detailed simulation of the
electromagnetic calorimeter (ECAL), which has been fully integrated in
the collaboration software framework CMSSW. The simulation is based on
the Geant4 detector simulation toolkit for the modelling of the passage
of particles through matter and magnetic field. The geometrical
description of the detector is being...
Dr
Sergio Andreozzi
(INFN-CNAF)
05/09/2007, 08:00
A key advantage of Grid systems is the capability of sharing
heterogeneous resources and services across traditional
administrative and organizational domains. This capability enables
the creation of virtual pools of resources that can be assigned to
groups of users. One of the problems that the utilization of such
pools presents is the awareness of the resources, i.e., the fact
that...
Mr
Riccardo Zappi
(INFN-CNAF)
05/09/2007, 08:00
In Grid systems, a core resource being shared among geographically-dispersed
communities of users is the storage. For this resource, a standard interface
specification (Storage Resource Management or SRM) was defined and is being evolved
in the context of the Open Grid Forum. By implementing this interface, all storage
resources part of a Grid could be managed in an homogenous fashion. In...
Dr
Piergiulio Lenzi
(Dipartimento di Fisica)
05/09/2007, 08:00
The first application of one of the official CMS tracking algorithm,
known as Combinatorial Track Finder, on cosmic
muon real data is described.
The CMS tracking system consists of a silicon pixel vertex detector and a
surrounding silicon microstrip detector.
The silicon strip tracker consists of 10 barrel layers and 12 endcap disks on each side.
The system is currently going through...
Dr
Andrea Fontana
(INFN-Pavia)
05/09/2007, 08:00
The concept of Virtual Monte Carlo allows to use different Monte Carlo
programs to simulate particle physics detectors without changing the
geometry definition and the detector response simulation.
In this context, to study the reconstruction capabilities of a detector,
the availability of a tool to extrapolate the track parameters and their
associated errors due to magnetic field,...
Dr
Gabriele Compostella
(University Of Trento INFN Padova)
05/09/2007, 08:00
When the CDF experiment was developing its software infrastructure,
most computing was done on dedicated clusters. As a result,
libraries, configuration files, and large executable were
deployed over a shared file system.
As CDF started to move into the Grid world, the assumption of having a
shared file system showed its limits. In a widely distributed computing model,
such as the...
Mr
Andreas Weindl
(FZ Karlsruhe / IK), Dr
Harald Schieler
(FZ Karlsruhe / IK)
05/09/2007, 08:00
The KASCADE-Grande experiment is a multi-detector installation at the site of the
Forschungszentrum Karlsruhe, Germany, to measure and study extensive air showers
induced in the atmosphere by primary cosmic rays in the energy range from 10^14 to
10^18 eV. For three of the detector components, WEB based online event displays have
been implemented. They provide in a fast and simplified way...
Mr
Laurence Field
(CERN)
05/09/2007, 09:00
Over recent years a number of grid projects have emerged which have built grid infrastructures
that are now the computing backbones for various user communities. A significant number of these
user communities are artificially limited to only one grid due the different middleware used in each
grid project. Grid interoperation is trying to bridge these differences and enable virtual...
Prof.
Frank Wuerthwein
(UCSD)
05/09/2007, 09:30
Harvey Newman
(California Institute of Technology (CALTECH))
05/09/2007, 11:00
Networks of sufficient and rapidly increasing end-to-end capability, as well as a high degree of reliability are vital for the LHC and other major HEP programs. Our bandwidth usage on the major national backbones and intercontinental links used by our field has progressed by a factor of several hundred over the past decade, and the outlook is for a similar increase over the next decade. This...
Paul Nilsson
(UT-Arlington)
05/09/2007, 14:00
The PanDA software provides a highly performant distributed production and
distributed analysis system. It is the first system in the ATLAS experiment to use a
pilot based late job delivery technique. In this talk, we will describe the
architecture of the pilot system used in Panda. Unique features have been implemented
for high reliability automation in a distributed environment....
Marco Clemencic
(European Organization for Nuclear Research (CERN))
05/09/2007, 14:00
Distributed data analysis and information management
oral presentation
The LHCb Conditions Database project provides the necessary tools to handle non-event
time-varying data. The main users of conditions are reconstruction and analysis
processes, which are running on the Grid. To allow efficient access to the data, we
need to use a synchronized replica of the content of the database located at the same
site as the event data file, i.e. the LHCb Tier1. The...
Dirk Duellmann
(CERN)
05/09/2007, 14:00
Computer facilities, production grids and networking
oral presentation
Relational database services are a key component of the computing models for the Large Hadron Collider (LHC). A
large proportion of non-event data including detector conditions, calibration, geometry and production
bookkeeping metadata require reliable storage and query services in the LHC Computing Grid (LCG). Also core grid
services to catalogue and distribute data cannot operate...
Dr
Matthias Wittgen
(SLAC)
05/09/2007, 14:00
The BaBar slow control system uses EPICS (Experimental Physics and
Industrial Control System) running on 17 VME based single board computers (SBCs).
EPICS supports the real-time operating systems vxWorks and RTEMS.
During the 2004/05 shutdown BaBar started to install a new detector
component, the Limited Streamer Tubes (LST), adding over 20000 high
voltage channels and about 350...
Boris Mangano
(University of California, San Diego)
05/09/2007, 14:00
With nominal collision energies of 14 TeV at luminosities of 10^34
cm^-2 s^-1, the LHC will explore energies an order of magnitude higher
than colliders before. This poses big challenges for the tracking
system and the tracking software to reconstruct tracks in the primary
collision and the ~20 underlying events.
CMS has built a full silicon tracking system consisting of an inner
pixel...
Dr
Lee Lueking
(FERMILAB)
05/09/2007, 14:20
Distributed data analysis and information management
oral presentation
The CMS experiment at the LHC has established an infrastructure using the FroNTier
framework to deliver conditions (i.e. calibration, alignment, etc.) data to
processing clients worldwide. FroNTier is a simple web service approach providing
client HTTP access to a central database service. The system for CMS has been
developed to work with POOL which provides object relational mapping...
Dr
Stuart Paterson
(CERN)
05/09/2007, 14:20
The LHCb DIRAC Workload and Data Management System employs advanced optimization
techniques in order to dynamically allocate resources. The paradigms realized by
DIRAC, such as late binding through the Pilot Agent approach, have proven to be
highly successful. For example, this has allowed the principles of workload
management to be applied not only at the time of user job submission to...
Dr
Xavier Espinal
(PIC/IFAE)
05/09/2007, 14:20
Computer facilities, production grids and networking
oral presentation
In preparation for first data at the LHC, a series of Data Challenges, of
increasing scale and complexity, have been performed. Large quantities of
simulated data have been produced on three different Grids, integrated into
the ATLAS production system. During 2006, the emphasis moved towards providing
stable continuous production, as is required in the immediate run-up to first
data, and...
Mr
Federico Carminati
(CERN)
05/09/2007, 14:20
Since 1998 the ALICE Offline Project has developed an integrated offline framework (AliRoot) and a distributed
computing environment (AliEn) to process the data of the ALICE experiment. These systems are integrated with the
LCG computing infrastructure, and in particular with the ROOT system and with the WLCG Grid middleware, but
they also present a number of original solutions, which...
Mr
Filimon Roukoutakis
(CERN & University of Athens)
05/09/2007, 14:20
ALICE is one of the experiments under installation at CERN Large Hadron Collider,
dedicated to the study of Heavy-Ion Collisions. The final ALICE Data Acquisition
system has been installed and is being used for the testing and commissioning of
detectors. Data Quality Monitoring
(DQM) is an important aspect of the online procedures for a HEP experiment. In this
presentation we overview the...
Mr
Sergio Gonzalez-Sevilla
(Instituto de Fisica Corpuscular (IFIC) UV-CSIC)
05/09/2007, 14:20
It is foreseen that the Large Hadron Collider will start its
operations and collide proton beams during November 2007. ATLAS is one of the
four LHC experiments currently under preparation. The alignment of the ATLAS
tracking system is one of the challenges that the experiment must solve in
order to achieve its physics goals. The tracking system comprises two silicon
technologies: pixel...
Mr
Serguei Kolos
(University of California Irvine)
05/09/2007, 14:35
Data Quality Monitoring (DQM) is an important and integral part of the data taking
and data reconstruction of HEP experiments. In an online environment, DQM
provides the shift crew with live information beyond basic monitoring. This is used to
overcome problems promptly and help avoid taking faulty data. During the off-line
reconstruction DQM is used for more complex analysis of physics...
Mr
Jose Hernandez Calama
(CIEMAT)
05/09/2007, 14:40
Computer facilities, production grids and networking
oral presentation
Monte Carlo production in CMS has received a major boost in performance and
scale since last CHEP conference. The production system has been re-engineered
in order to incorporate the experience gained in running the previous system
and to integrate production with the new CMS event data model, data management
system and data processing framework. The system is interfaced to the two...
Alexandre Vaniachine
(Argonne National Laboratory)
05/09/2007, 14:40
Distributed data analysis and information management
oral presentation
In preparation for ATLAS data taking in ATLAS database activities a coordinated shift from
development towards operations has occurred. In addition to development and
commissioning activities in databases, ATLAS is active in the development and deployment
(in collaboration with the WLCG 3D project) of the tools that allow the worldwide
distribution and installation of databases and...
Dr
Yuri Fisyak
(BROOKHAVEN NATIONAL LABORATORY)
05/09/2007, 14:40
The STAR experiment was primarily designed to detect signals of a possible phase
transition in nuclear matter. Its layout, typical for a collider experiment, contains
a large Time Projection Chamber (TPC) in a Solenoid Magnet, a set of four layers of
combined silicon strip and silicon drift detectors for secondary vertex
reconstruction plus other detectors. In this presentation, we will...
Mr
Marco Cecchi
(INFN cnaf)
05/09/2007, 14:40
The gLite Workload Management System (WMS) is a collection of components providing a
service responsible for the distribution and management of tasks across resources
available on a Grid. The main purpose is to accept a request of execution of a job
from a client, find appropriate resources to satisfy it and follow it until
completion. Different aspects of job management are accomplished...
Dr
William Badgett
(Fermilab)
05/09/2007, 14:50
We present the Online Web Based Monitoring (WBM) system of the CMS experiment,
consisting of a web services framework based on Jakarta/Tomcat and the
Root data display package. Due to security concerns, many monitoring
applications of the CMS experiment cannot be run outside of the experimental
site. As such, in order to allow remote users access to CMS experimental
status information,...
Dr
Markus Stoye
(Inst. f. Experimentalphysik, Universitaet Hamburg)
05/09/2007, 15:00
The CMS silicon tracker comprises about 17000 silicon modules. Its radius
and length of 120 cm and 560 cm, respectively, make it the largest silicon tracker
ever built. To fully exploit the precise hit measurements, it is necessary to
determine the positions and orientations of the silicon modules to the level of mum and
murad, respectively.
Among other track based alignment algorithms,...
Maria Girone
(CERN)
05/09/2007, 15:00
Computer facilities, production grids and networking
oral presentation
Physics meta-data stored in relational databases play a crucial role in the Large Hadron Collider (LHC) experiments
and also in the operation of the Worldwide LHC Computing Grid (WLCG) services. A large proportion of non-event
data such as detector conditions, calibration, geometry and production bookkeeping relies heavily on databases.
Also, the core Grid services that catalogue and...
Dr
Douglas Smith
(Stanford Linear Accelerator Center)
05/09/2007, 15:00
Distributed data analysis and information management
oral presentation
There is a need for a large dataset of simulated events for use in
analysis of the data from the BaBar high energy physics experiment.
The largest cycle of this production in the history of the experiment
was just completed in the past year, simulating events against all
detector conditions in the history of the experiment, resulting in over
eleven billion events in eighteen months. ...
Mr
Igor Soloviev
(CERN/PNPI)
05/09/2007, 15:05
This paper describes challenging requirements on the configuration service. It
presents the status of the implementation and testing one year before the
start of the ATLAS experiment at CERN providing details of:
- capabilities of underlying OKS* object manager to store and to archive
configuration descriptions, it's user and programming interfaces;
- the organization of configuration...
Dr
Martin Weber
(RWTH Aachen, Germany)
05/09/2007, 15:20
The full-silicon tracker of the CMS experiment with its 15148 strip and 1440
pixel modules is of an unprecedented size. For optimal track-parameter
resolution, the position and orientation of its modules need to be determined
with a precision of a few micrometer.
Starting from the inclusion of survey measurements, the use of a hardware
alignment system, and track based alignment, this...
Dr
Michael Wilson
(European Organisation for Nuclear Research (CERN))
05/09/2007, 15:20
Assessing the quality of data recorded with the Atlas detector
is crucial for commissioning and operating the detector to achieve
sound physics measurements.
In particular, the fast assessment of complex quantities obtained during
event reconstruction and the ability to easily track them over time are
especially important given the large data throughput and the
distributed nature of the...
Smirnov Yuri
(Brookhaven National Laboratory)
05/09/2007, 15:20
Computer facilities, production grids and networking
oral presentation
The Open Science Grid infrastructure provides one of the largest distributed
computing systems deployed in the ATLAS experiment at the LHC. During the CSC
exercise in 2006-2007, OSG resources provided about one third of the worldwide
distributed computing resources available in ATLAS. About half a petabyte of ATLAS MC
data is stored on OSG sites. About 2000k SpecInt2000 CPU's is available....
Ms
Helen McGlone
(University of Glasgow/CERN)
05/09/2007, 15:20
Distributed data analysis and information management
oral presentation
The ATLAS TAG database is a multi-terabyte event-level metadata selection system,
intended to allow discovery, selection of and navigation to events of interest to an
analysis. The TAG database encompasses file- and relational-database-resident
event-level metadata, distributed across all ATLAS Tiers.
...
Dr
Sanjay Padhi
(University of Wisconsin-Madison)
05/09/2007, 15:20
With the evolution of various Grid Technologies along with foreseen
first LHC collision this year, a homogeneous and interoperable Production
system for ATLAS is a necessity. We present the CRONUS, which a Condor
Glide-in based ATLAS Production Executor. The Condor glide-in daemons
traverse to the Worker nodes, submitted via Condor-G or gLite RB. Once
activated, they preserve the...
Dr
Yao Zhang
(Institute of High Energy Physics, Chinese Academy of Sciences)
05/09/2007, 15:20
The BESIII detector will be commissioned at the upgraded Beijing Electron
Positron Collider (BEPCII) at the end of 2007. The drift chamber(MDC), which is
one of the most important sub-detectors of the BESIII detector, is expected to
provide good momentum resolution (0.5%@1GeV/c) and tracking efficiency in a
range of 0.1~2.0 GeV/c. This makes stringent demands on the performance of...
Vardan Gyurjyan
(Jefferson Lab)
05/09/2007, 15:35
AFECS is a pure Java based software framework for designing and implementing
distributed control systems. AFECS creates a control system environment as a
collection of software agents behaving as finite state machines. These agents can
represent real entities, such as hardware devices, software tasks, or control
subsystems. A special control oriented ontology language (COOL), based on RDFS...
Mr
Dave Evans
(Fermi National Laboratory)
05/09/2007, 15:40
Computer facilities, production grids and networking
oral presentation
The CMS production system has undergone a major architectural upgrade from its
predecessor, with the goals of reducing the operations manpower requirement and
preparing for the large scale production required by the CMS physics plan.
This paper discusses the CMS Monte Carlo Workload Management architecture. The
system consist of 3 major components: ProdRequest, ProdAgent, and ProdMgr...
Mr
Juan Manuel Guijarro
(CERN)
05/09/2007, 15:40
The Database and Engineering Services Group of CERN's Information
Technology Department provides the Oracle based Central Data Base services used
in many activities at CERN.
In order to provide High Availability and ease management for those
services, a NAS (Network Attached Storage) based infrastructure has been set
up. It runs several instances of the Oracle RAC (Real...
Dr
Conrad Steenberg
(Caltech)
05/09/2007, 15:40
Distributed data analysis and information management
oral presentation
We describe how we have used the Clarens Grid Portal Toolkit to develop powerful
application and browser-level interfaces to ROOT and Pythia. The Clarens Toolkit is a
codebase that was initially developed under the auspices of the Grid Analysis
Environment project at Caltech, with the goal of enabling LHC physicists engaged in
analysis to bring the full power of the Grid to their desktops,...
Dr
Stefano Spataro
(II Physikalisches Institut, Universität Giessen (Germany))
05/09/2007, 15:40
The PANDA detector will be located at the future GSI accelerator FAIR. Its
primary objective is the investigation of strong interaction with anti-proton
beams, in the range up to 15 GeV/c as momentum of the incoming anti-proton.
The PANDA offline simulation framework is called “PandaRoot”, as it is based
upon the ROOT 5.12 package. It is characterized by a high versatility; it allows...
Mr
Philippe Canal
(FERMILAB)
05/09/2007, 16:30
We will review the architecture and implementation of the accounting
service for the Open Science Grid. Gratia's main goal is to provide the OSG
stakeholders with a reliable and accurate set of views of the usage of resources
across the OSG.
We will review the status of deployment of Gratia across the OSG and its
upcoming development. We will also discuss some aspects of current OSG...
Mr
Nicholas Robinson
(CERN)
05/09/2007, 16:30
CERN has long been committed to the free dissemination of scientific
research results and theories. Towards this end, CERN's own
institutional repository, the CERN Document Server (CDS) offers access
to CERN works and to all related scholarly literature in the HEP
domain. Hosting over 500 document collections containing more than
900,000 records, CDS provides access to anything from...
Mr
Philip DeMar
(FERMILAB)
05/09/2007, 16:30
Computer facilities, production grids and networking
oral presentation
Fermilab hosts the American Tier-1 Center for the LHC/CMS experiment. In preparation
for the startup of CMS, and building upon extensive experience supporting TeVatron
experiments and other science collaborations, the Laboratory has established high
bandwidth, end-to-end (E2E) circuits with a number of US-CMS Tier2 sites, as well as
other research facilities in the collaboration. These...
Mrs
Maria Del Carmen Barandela Pazos
(University of Vigo)
05/09/2007, 16:45
In a High Energy Physics experiment it is fundamental to handle information
related to the status of the detector and its environment at the time of the
acquired event. This type of time-varying non-event data are often grouped
under the term “conditions”. The LHCb’s Experiment Control System groups all
the infrastructure for the configuration, control and monitoring of all the...
Martin Flechl
(IKP, Uppsala Universitet)
05/09/2007, 16:50
A Grid is defined as being ``coordinated resource sharing and problem solving in
dynamic, multi-institutional virtual organizations''. Over recent years a number of
grid projects, many of which have a strong regional presence, have emerged to help
coordinate institutions and enable grids. Today, we face a situation where a number
of grid projects exist, most of which have slightly...
Dr
Douglas Smith
(Stanford Linear Accelerator Center)
05/09/2007, 16:50
International multi-institutional high energy physics experiments require easy
means for collaborators to communicate coherently in a global community. To
fill this need, the HyperNews system has been widely used in HEP. HyperNews is
a discussion management system which is a hybrid between a web-base forum
system and a mailing list system. Its goal is to provide a tool for
distributed...
Mr
Maxim Grigoriev
(FERMILAB)
05/09/2007, 16:50
Computer facilities, production grids and networking
oral presentation
The LHC experiments will start very soon, creating immense data volumes capable of
demanding allocation of an entire network circuit for task-driven applications.
Circuit-based alternate network paths are one solution to meeting the LHC high
bandwidth network requirements. The Lambda Station project is aimed at addressing
growing requirements for dynamic allocation of alternate network...
O Solovyanov
(IHEP, Protvino, Russia)
05/09/2007, 17:00
An online control system to calibrate and monitor ATLAS Barrel hadronic calorimeter
(TileCal) with a movable radioactive source, driven by liquid flow, is described.
To read out and control the system an online software has been developed, using ATLAS
TDAQ components like DVS (Diagnostic and Verification System) to verify the HW before
running, IS (Information Server) for data and...
Dr
Matt Crawford
(FERMILAB)
05/09/2007, 17:10
Computer facilities, production grids and networking
oral presentation
Due to shortages of IPv4 address space - real or artificial - many HEP
computing installations have turned to NAT and application gateways.
These workarounds carry a high cost in application complexity and
performance. Recently a few HEP facilities have begun to deploy IPv6
and it is expected that many more must follow within several years.
While IPv6 removes the problem of address...
Dr
Thijs Cornelissen
(CERN)
05/09/2007, 17:10
While most high energy experiments use track fitting software that is
based on the Kalman technique, the ATLAS offline reconstruction has
several global track fitters available. One of these is the global chi^2
fitter, which is based on the scattering angle formulation of the track
fit. One of the advantages of this method over the Kalman fit is that it
can provide the scattering angles...
Mr
Jeremy Herr
(University of Michigan)
05/09/2007, 17:10
Large scientific collaborations as well as universities have a growing need for
multimedia archiving of meetings and courses. Collaborations need to disseminate
training and news to their wide-ranging members, and universities seek to provide
their students with more useful studying tools. The University of Michigan ATLAS
Collaboratory Project has been involved in the recording and...
Mr
Sebastian Robert Bablok
(Department of Physics and Technology, University of Bergen)
05/09/2007, 17:15
The ALICE HLT is designed to perform event analysis including calibration of the
different ALICE detectors online. The detector analysis codes process data using the
latest calibration and condition settings of the experiment. This requires a high
reliability on the interfaces to the various other systems operating ALICE.
In order to have a comparable analysis with the results from...
Mr
Maxim Grigoriev
(FERMILAB)
05/09/2007, 17:30
Computer facilities, production grids and networking
oral presentation
End-to-end (E2E) circuits are used to carry high impact data movement into and out of
the US CMS Tier-1 Center at Fermilab. E2E circuits have been implemented to
facilitate the movement of raw experiment data from Tier-0, as well as processed data
to and from a number of the US Tier-2 sites. Troubleshooting and monitoring those
circuits presents a challenge, since the circuits typically...
Dr
Ronan McNulty
(University College Dublin, School of Physics)
05/09/2007, 17:30
As programming and their environments become increasingly complex, more
effort must be invested in presenting the user with a simple yet comprehensive
interface. Feicim is a tool that unifies the representation of data and
algorithms. It provides resource discovery of data-files, data-content and
algorithm implementation through an intuitive graphical user interface. It allows...
Dr
Antonio Pierro
(INFN-BARI)
05/09/2007, 17:30
The monitoring of the grid user activity and application performance is extremely
useful to plan resource usage strategies particularly in cases of complex applications.
Large VO's , like the LHC ones, do their monitoring by means of dashboards. Other
VO's or communities, like for example the BioinforGRID one, are characterized by a
greater diversification of the application types: so...
Tumanov Alexander
(T.W. Bonner Nuclear Laboratory)
05/09/2007, 17:45
Unprecedented data rates that are expected at the LHC put high demand on the
speed of the detector data acquisition system. The CSC subdetector located in
the Muon Endcaps of the CMS detector has a data readout system equivalent in
size to that of a whole Tevatron detector (60 VME crates in the CSC DAQ equal
to the whole D0 DAQ size).
As a part of the HLT, the CSC data unpacking...
Dr
Christopher Jones
(Cornell University)
05/09/2007, 17:50
The CMS offline software suite uses a layered approach to provide several different environments suitable for a
wide range of analysis styles.
At the heart of all the environments is the ROOT-based event data model file format. The simplest environment
uses "bare" ROOT to read files directly, without the use of any CMS-specific supporting libraries. This is useful for
performing...
Dr
Luc Goossens
(CERN)
05/09/2007, 17:50
Computer facilities, production grids and networking
oral presentation
ATLAS is a multi-purpose experiment at the LHC at CERN,
which will start taking data in November 2007.
To handle and process the unprecedented data rates expected
at the LHC (at nominal operation, ATLAS will record about
10 PB of raw data per year) poses a huge challenge on the
computing infrastructure.
The ATLAS Computing Model foresees a multi-tier hierarchical
model to perform this...
Dietrich Liko
(CERN)
06/09/2007, 08:30
Dietrich Liko:
Dietrich Liko is researcher at the Institute for High Energy Physics of the Austrian Academy of Sciences. He is currently on leave to participate in the devlopement of analysis tools for the grid with the EGEE project and as ATLAS Distributed Analysis Coordinator.
Dr
Amber Boehnlein
(FERMI NATIONAL ACCELERATOR LABORATORY)
06/09/2007, 09:00
Dr
Richard Mount
(SLAC)
06/09/2007, 11:00
Peter Clarke
(School of Physics - University of Edinburgh)
06/09/2007, 12:00
Dr
Lucas Taylor
(Northeastern University, Boston)
06/09/2007, 14:00
Distributed data analysis and information management
oral presentation
The CMS experiment is about to embark on its first physics run at the LHC. To
maximize the effectiveness of physicists and technical experts at CERN and
worldwide and to facilitate their communications, CMS has established several
dedicated and inter-connected operations and monitoring centers. These
include a traditional “Control Room” at the CMS site in France, a “CMS Centre”
for...
Dr
Stephen Burke
(Rutherford Appleton Laboratory, UK)
06/09/2007, 14:00
A common information schema for the description of Grid resources and services
is an essential requirement for interoperating Grid infrastructures, and its
implementation interacts with every Grid component. In this context, the GLUE
information schema was originally defined in 2002 as a joint project between
the European DataGrid and DataTAG projects and the US iVDGL (the...
Mr
Tomasz Maciej Frueboes
(Institute of Experimental Physics - University of Warsaw)
06/09/2007, 14:00
The CMS detector will start its operation in the end of 2007. Until
that time great care must be taken in order to assure that hardware
operation is fully understood. We present an example of how emulation
software helps achieving this goal in the CMS Level-1 RPC Trigger
system.
The design of the RPC trigger allows to insert sets of so-called test
pulses at any stage of the hardware...
Dr
Lukas Nellen
(I. de Ciencias Nucleares, UNAM)
06/09/2007, 14:00
Computer facilities, production grids and networking
oral presentation
The EELA project aims at building a grid infrastructure in Latin
America and at attracting users to this infrastructure. The EELA
infrastructure is based on the gLite middleware, developed by the EGEE
project. A test-bed, including several European and Latin American
countries, was set up in the first months of the project. Several
applications from different areas, especially...
Dr
Ivana Hrivnacova
(IPN, Orsay, France)
06/09/2007, 14:00
The Virtual Geometry Model (VGM) was introduced at CHEP in 2004,
where its concept, based on the abstract interfaces to geometry
objects, has been presented. Since then, it has undergone a
design evolution to pure abstract interfaces, it has been consolidated
and completed with more advanced features. Currently it is used
in Geant4 VMC for the support of TGeo geometry definition...
Dr
Alexei Klimentov
(BNL)
06/09/2007, 14:20
Computer facilities, production grids and networking
oral presentation
ATLAS Distributed Data Management Operations Team unites experts from
Tier-1s and Tier-2s computer centers. The group is responsible for all day
by day ATLAS data distribution between different sites and centers.
In our paper we describe ATLAS DDM operation model and address the
data management and operation issues. A serie of Functional Tests have
been conducted in the past and is in...
Dr
Helen Hayward
(University of Liverpool)
06/09/2007, 14:20
The inner detector of the ATLAS experiment is in the process of being
commissioned using cosmic ray events. First tests were performed in
the SR1 assembly hall at CERN with both barrel and endcaps for all
different detector technologies (pixels and microstrips silicon
detectors as well as straw tubes with additional transition radiation
detection). Integration with the rest of the ATLAS...
Dr
John Kennedy
(LMU Munich)
06/09/2007, 14:20
Distributed data analysis and information management
oral presentation
The ATLAS production system is responsible for the distribution of
O(100,000) jobs per day to over 100 sites worldwide.
The tracking and correlation of errors and resource usage within such a
large distributed system is of extreme importance.
The monitoring system presented here is designed to abstract the
monitoring information away form the central database of jobs....
Dr
Martin Purschke
(BROOKHAVEN NATIONAL LABORATORY)
06/09/2007, 14:20
The PHENIX experiment at the Relativistic Heavy Ion Collider (RHIC) has
commissioned several new detector systems which are part of the
general readout for the first time in the RHIC Run 7, which is
currently under way.
In each of the RHIC Run periods since 2003, PHENIX has collected about
0.5 PB of data. For Run 7 we expect record luminosities for
the Au-Au beams, which will lead to...
Lassi Tuura
(Northeastern University)
06/09/2007, 14:40
The CMS experiment at LHC has a very large body of software of its own and uses extensively software from
outside the experiment. Understanding the performance of such a complex system is a very challenging task,
not the least because there are extremely few developer tools capable of profiling software systems of this scale,
or producing useful reports.
CMS has mainly used IgProf,...
Dr
Daniele Bonacorsi
(INFN-CNAF, Bologna, Italy)
06/09/2007, 14:40
Computer facilities, production grids and networking
oral presentation
The CMS experiment is gaining experience towards the data taking in several computing preparation activities, and a
roadmap towards a mature computing operations model stands as a primary target. The responsibility of the
Computing Operations projects in the complex CMS computing environment spawns a wide area and aims at
integrating the management of the CMS Facilities Infrastructure,...
Dr
Tofigh Azemoon
(Stanford Linear Accelerator Center)
06/09/2007, 14:40
Distributed data analysis and information management
oral presentation
Petascale systems are in existence today and will become widespread in the
next few years. Such systems are inevitably very complex, highly distributed
and heterogeneous. Monitoring a petascale system in real time and
understanding its status at any given moment without impacting its
performance is a highly intricate task. Common approaches and off the shelf
tools are either...
Giuseppe Bagliesi
(INFN Sezione di Pisa)
06/09/2007, 14:40
Tau leptons play surely a key role in the physics studies at the
LHC. Interests in using tau leptons include (but are not limited to)
their ability to offer a relatively low background environment, a
competitive way of probing new physics as well as the possibility to
explore new physics regions not accessible otherwise.The Tau
identification and reconstruction algorithms developed for...
Dr
Andrew McNab
(University of Manchester)
06/09/2007, 14:40
Components of the GridSite system are used within WLCG and gLite to process security
credentials and access policies. We describe recent extensions to this system to
include the Shibboleth authentication framework of Internet2, and how the GridSite
architecture can now import a wide variety of credential types, including onetime
passcodes, X.509, GSI, VOMS, Shibboleth and OpenID and then...
Dr
Benedetto Gorini
(CERN)
06/09/2007, 14:50
During 2006 and early 2007, integration and commissioning of trigger and data acquisition (TDAQ) equipment in
the ATLAS experimental area have progressed. Much of the work has focussed on a final prototype setup consisting
of around 80 computers representing a subset of the full TDAQ system. There have been a series of technical runs
using this setup. Various tests have been run...
Mr
Riccardo Zappi
(INFN-CNAF)
06/09/2007, 15:00
In the near future, data on the order of hundred of Petabytes will be spread in
multiple storage systems worldwide dispersed in, potentially, billions of replicated
data items. Users, typically, are agnostic about the location of their data and they
want to get access by either specifying logical names or using some lookup mechanism.
A global namespace is a logical layer that allows...
Ricardo Rocha
(CERN)
06/09/2007, 15:00
Distributed data analysis and information management
oral presentation
The ATLAS Distributed Data Management (DDM) system is evolving to
provide a production-quality service for data distribution and data
management support for production and users' analysis.
Monitoring the different components in the system has emerged as one of
the key issues to achieve this goal. Its distributed nature over
different grid infrastructures (EGEE, OSG and NDGF)...
Dr
Sebastien Binet
(LBNL)
06/09/2007, 15:00
Python does not, as a rule, allow many optimizations, because there
are too many things that can change dynamically. However, a lot of HEP analysis
work consists of logically immutable blocks of code that are executed many
times: looping over events, fitting data samples, making plots. In fact, most
parallelization relies on this. There is therefore room for optimizations.
There are...
Mr
Pavel Reznicek
(IPNP, Charles University in Prague)
06/09/2007, 15:00
The LHC experiments will search for physics phenomena beyond the Standard Model
(BSM). Highly sensitive tests of beauty hadrons will represent an alternative
approach to this research. The analyzes of complex decay chains of beauty hadrons
will require involving several nodes, and detector tracks made by these reactions
must be extracted efficiently from other events to make...
Luca dell'Agnello
(INFN-CNAF)
06/09/2007, 15:00
Computer facilities, production grids and networking
oral presentation
Performance, reliability and scalability in data access are key issues when
considered in the context of HEP data processing and analysis applications.
The importance of these topics is even larger when considering the quantity of data
and the request load that a LHC data centers has to support.
In this paper we give the results and the technical details of a large scale
validation,...
Dr
Fons Rademakers
(CERN)
06/09/2007, 15:20
Distributed data analysis and information management
oral presentation
The goal of PROOF (Parallel ROOt Facility) is to enable interactive
analysis of large data sets in parallel on a distributed cluster or
multi-core machine. PROOF represents a high-performance alternative
to a traditional batch-oriented computing system.
The ALICE collaboration is planning to use PROOF at the CERN Analysis Facility
(CAF) and has been stress testing the system since mid...
Dr
Robert Bainbridge
(Imperial College London)
06/09/2007, 15:20
The CMS silicon strip tracker, providing a sensitive area of >200 m^2 and comprising
10M readout channels, is undergoing final assembly at the tracker integration
facility at CERN. The strip tracker community is currently working to develop and
integrate the online and offline software frameworks, known as XDAQ and CMSSW
respectively, for the purposes of data acquisition and detector...
Dr
Kirill Prokofiev
(University of Sheffield)
06/09/2007, 15:20
In the harsh environment of the Large Hadron Collider at CERN (design
luminosity of 10^34 cm-2s-1) efficient reconstruction of the signal primary
vertex is crucial for many physics analyses. Described in this paper are
primary vertex reconstruction strategies implemented in the ATLAS software
framework Athena. The implementation of the algorithms follows a very
modular design based on...
Jan van ELDIK
(CERN)
06/09/2007, 15:20
Computer facilities, production grids and networking
oral presentation
This paper presents work, both completed and planned, for streamlining the
deployment, operation and re-tasking of Castor2 instances. We present a summary of
what has recently been done to reduce the human intervention necessary for bringing
systems into operation; including the automation of Grid host certificate requests
and deployment in conjunction with the CERN Trusted CA and...
Dr
Jörg Stelzer
(CERN, Switzerland)
06/09/2007, 15:20
The ATLAS detector at CERN's LHC will be exposed to proton-proton
collisions at a rate of 40 MHz. To reduce the data rate, only
potentially interesting events are selected by a three-level trigger
system. The first level is implemented in custom-made electronics,
reducing the data output rate to less than 100 kHz. The second and
third levels are software triggers with a final output rate...
Mr
Levente Hajdu
(BROOKHAVEN NATIONAL LABORATORY)
06/09/2007, 15:35
Keeping a clear and accurate experiment log is important for any scientific
experiment. The concept is certainly not new but keeping accurate while useful
records for a Nuclear Physics experiment such as RHIC/STAR is not a priori a simple
matter – STAR operates 24 hours a day for six months out of the year with more then
24 shift crews operating 16 different subsystems (some located...
Mr
Fabrizio Furano
(INFN sez. di Padova)
06/09/2007, 15:40
Distributed data analysis and information management
oral presentation
HEP data processing and analysis applications typically deal
with the problem of accessing and processing data at high speed.
Recent study, development and test work has shown that the latencies
due to data access can often be hidden by parallelizing them
with the data processing, thus giving the ability
to have applications which process remote data with a high level of...
Mr
Timur Perelmutov
(FERMILAB)
06/09/2007, 15:40
Computer facilities, production grids and networking
oral presentation
The Storage Resource Manager (SRM) and WLCG collaborations recently
defined version 2.2 of the SRM protocol, with the goal of satisfying
the requirement of the LCH experiments. The dCache team has now
finished the implementation of all SRM v2.2 elements required by the
WLCG. The new functions include space reservation, more advanced data
transfer, and new namespace and permission...
Valentin Kuznetsov
(Cornell University)
06/09/2007, 15:40
We disscuss the rapid development of a large scale data discovery
service for the CMS experiment using modern AJAX techniques and
the Python language. To implement a flexible interface capable of
accommodating several different versions of the DBS databse, we used
a "stack" approach. Asynchronous JavaScript and XML (AJAX) together
with an SQL abstraction layer, template engine, code...
Dr
Haleh Hadavand
(Southern Methodist University)
06/09/2007, 16:30
The ATLAS experiment of the LHC is now taking its first data by
collecting cosmic ray events. The full reconstruction chain including
all sub-systems (inner detector, calorimeters and muon spectrometer)
is being commissioned with this kind of data for the first time.
Specific adaptations to deal with particles not coming from the
interaction point and not synchronized with the readout...
Dr
Maxim Potekhin
(BROOKHAVEN NATIONAL LABORATORY)
06/09/2007, 16:30
Computer facilities, production grids and networking
oral presentation
The simulation program for the STAR experiment at Relativistic Heavy Ion Collider at
Brookhaven National Laboratory is growing in scope and responsiveness to the needs of
the research conducted by the Physics
Working Groups. In addition, there is a significant ongoing R&D activity aimed at
future upgrades of the STAR detector, which also requires extensive simulations
support. The...
Dr
Mikhail Kirsanov
(Institute for Nuclear Research (INR))
06/09/2007, 16:30
The Generator Services project collaborates with the Monte Carlo
generators authors and with the LHC experiments in order to prepare
validated LCG compliant code for both the theoretical and the
experimental communities at the LHC. On the one side it provides the
technical support as far as the installation and the maintenance of
the generators packages on the supported platforms is...
Dan Flath
(SLAC)
06/09/2007, 16:30
Distributed data analysis and information management
oral presentation
The Data Handling Pipeline ("Pipeline") has been developed for the Gamma-Ray Large
Area Space Telescope (GLAST) launching at the end of 2007. Its goal is to generically
process graphs of dependent tasks, maintaining a full record of its state, history
and data products. In cataloging the relationship between data, analysis results,
software versions, as well as statistics (memory usage,...
Dr
Greig A Cowan
(University of Edinburgh)
06/09/2007, 16:30
The start of data taking this year at the Large Hadron Collider will
herald a new era in data volumes and distributed processing in
particle physics. Data volumes of 100s of Terabytes will be shipped
to Tier-2 centres for analysis by the LHC experiments using the
Worldwide LHC Computing Grid (WLCG).
In many countries Tier-2 centres are distributed between a number of
institutes, e.g.,...
Igor Sfiligoi
(Fermilab)
06/09/2007, 16:50
Computer facilities, production grids and networking
oral presentation
Pilot jobs are becoming increasingly popular in the Grid world. Experiments like
ATLAS and CDF are
using them in production, while others, like CMS, are actively evaluating them.
Pilot jobs enter Grid sites using a generic pilot credential, and once on a worker
node, call home to fetch the job of an actual user.
However, this operation mode poses several new security problems when...
Mrs
Ianna Osborne
(Northeastern University)
06/09/2007, 16:50
The event display and data quality monitoring visualisation
systems are especially crucial for commissioning CMS in the
imminent CMS physics run at the LHC. They have already proved
invaluable for the CMS magnet test and cosmic challenge.
We describe how these systems are used to navigate and filter
the immense amounts of complex event data from the CMS detector
and prepare clear and...
Dr
Vitaly Choutko
(Massachusetts Institute of Technology (MIT))
06/09/2007, 16:50
Distributed data analysis and information management
oral presentation
The AMS-02 detector will be installed on ISS ifor at least 3 years. The data will be
transmitted from ISS to NASA Marshall Space Flight Center (MSFC, Huntsvile, Alabama)
and transfered to CERN (Geneva Switzerland) for processing and analysis.
We are presenting the AMS-02 Ground Data Handling scenario and requirements to AMS
ground centers: the Payload Operation and Control Center (POCC)...
Mr
Ian Gable
(University of Victoria)
06/09/2007, 16:50
Deployment of HEP application in heterogeneous grid environments can be challenging
because many of the applications are dependent on specific OS versions and have a
large number of complex software dependencies. Virtual machine monitors such as Xen
could ease the deployment burden by allowing applications to be packaged complete
with their execution environments. Our previous work has...
Dr
Peter Elmer
(Princeton University)
06/09/2007, 16:50
Modern HEP experiments at colliders typically require offline software
systems consisting of many millions of lines of code. The software is
developed by hundreds of geographically distributed developers and is
often used actively for 10-15 years or longer. The tools and technologies
to support this HEP software development model have long been an interesting
topic at CHEP conferences....
Dr
Andrea Dotti
(Università and INFN Pisa)
06/09/2007, 17:10
The Tile Calorimeter (TileCal) is the central hadronic calorimeter of the
ATLAS experiment presently in an advanced state of installation and
commissioning at the LHC accelerator.
The complexity of the experiment, the number of electronics channels and the
high rate of acquired events requires a detailed commissioning of the detector,
during the installation phase of the experiment and...
Dr
Simone Campana
(CERN/IT/PSS)
06/09/2007, 17:10
Computer facilities, production grids and networking
oral presentation
The ATLAS experiment has been running continuous simulated events production since
more than two years. A considerable fraction of the jobs is daily submitted and
handled via the gLite Workload Management System, which overcomes several limitations
of the previous LCG Resource Broker. The gLite WMS has been tested very intensively
for the LHC experiments use cases for more than six months,...
Dr
Alfredo Pagano
(INFN/CNAF, Bologna, Italy)
06/09/2007, 17:10
Worldwide grid projects such as EGEE and WLCG need services with high availability,
not only for grid usage, but also for associated operations. In particular, tools
used for daily activities or operational procedures are considered critical.
In this context, the goal of the work done to solve the EGEE failover problem is to
propose, implement and document well-established mechanisms and...
Dr
Nicola De Filippis
(INFN Bari)
06/09/2007, 17:10
Distributed data analysis and information management
oral presentation
The Tracker detector has been taking real data with cosmics at the
Tracker Integration Facility (TIF) at CERN.
First DAQ checks and on-line monitoring tasks are executed at the
Tracker Analysis Centre (TAC) which is a dedicated Control Room at TIF with
limited computing resources. A set of software agents were developed
to perform the real-time data conversion in a standard Event...
Dr
Marcin Nowak
(Brookhaven National Laboratory)
06/09/2007, 17:30
In anticipation of data taking, ATLAS has undertaken a program of work
to develop an explicit state representation of the experiment's complex transient
event data model. This effort has provided both an opportunity to
consider explicitly the structure, organization, and content of the ATLAS persistent
event store before writing tens of petabytes of data (replacing simple...
Mr
Pavel Jakl
(Nuclear Physics Institute, Academy of Sciences of the Czech Republic)
06/09/2007, 17:30
Distributed data analysis and information management
oral presentation
Facing the reality of storage economics, NP experiments such as RHIC/STAR have been
engaged in a shift in the analysis model, and now heavily rely on using cheap disks
attached to processing nodes, as such a model is extremely beneficial over expensive
centralized storage. Additionally, exploiting storage aggregates with enhanced
distributed computing capabilities such as dynamic space...
Mr
Robert Stober
(Platform Computing)
06/09/2007, 17:30
Universus refers to an extension to Platform LSF that provides a secure,
transparent, one-way interface from an LSF cluster to any foreign cluster. A
foreign cluster is a local or remote cluster managed by a non-LSF workload
management system. Universus schedules work to foreign clusters as it would
to any other execution host.
Beyond its ability to interface with foreign workload...
Mr
Sergey Chechelnitskiy
(Simon Fraser University)
06/09/2007, 17:30
Computer facilities, production grids and networking
oral presentation
SFU is responcible for running two different clusters - one is designed for WestGrid internal
jobs with its specific software and the other should run Atlas jobs only. In addition to
different software configuration the Atlas cluster should have a diffener networking
confirugation. We would also like to have a flexibility of running jobs on different
hardware. That is why it has been...
Ms
Zhenping Liu
(BROOKHAVEN NATIONAL LABORATORY)
06/09/2007, 17:50
Computer facilities, production grids and networking
oral presentation
BNL ATLAS Computing Facility needs to provide a Grid-based storage system with these
requirements: a total of one gigabyte per second of incoming and outgoing data rate
between BNL and ATLAS T0, T1 and T2 sites, thousands of reconstruction/analysis jobs
accessing locally stored data objects, three petabytes of disk/tape storage in 2007
scaling up to 25 petabytes by 2011, and a...
Mr
Giulio Eulisse
(Northeastern University of Boston)
06/09/2007, 17:50
CMS software depends on over one hundred external packages, it's therefore obvious that being able to manage
the way they are built, deployed and configured and their dependencies (both among themselves and with
respect to core CMS software) is a critical part of the system.
We present a completely new system used to build and distribute CMS software which has
enabled us to go from...
Mr
Andreas Salzburger
(University of Innsbruck & CERN)
06/09/2007, 17:50
The track reconstruction of modern high energy physics experiments is a very complex
task that puts stringent requirements onto the software realisation. The ATLAS track
reconstruction software has been in the past dominated by a collection of individual
packages, each of which incorporating a different intrinsic event data model,
different data flow sequences and calibration data. The...
Daniele Spiga
(Universita degli Studi di Perugia)
06/09/2007, 17:50
Distributed data analysis and information management
oral presentation
Starting from 2007 the CMS experiment will produce several Pbytes of data each
year, to be distributed over many computing centers located in many different
countries. The CMS computing model defines how the data are to be distributed such
that CMS physicists can access them in an efficient manner in order to
perform their physics analyses. CRAB (CMS Remote Analysis Builder) is a...
Prof.
Roger Jones
(Lancaster University)
07/09/2007, 08:50
Mr
Federico Carminati
(CERN)
07/09/2007, 09:10
Patricia McBride
(Fermi National Accelerator Laboratory (FNAL))
07/09/2007, 09:30
Kors Bos
(NIKHEF)
07/09/2007, 10:30
Alberto Pace
(CERN)
Plenary
oral presentation
This talk will introduce identity management concepts and discuss various
issues associated with its implementation. The presentation will try to highlight
technical, legal, and social aspects that must been foreseen when defining the
numerous processes that an identity management infrastructure must support.
Mr
Jose Miguel Dana Perez
(CERN), Mr
Xavier Grehant
(CERN)
Grid middleware and tools
oral presentation
Today virtualization is used in computing centers to supply execution environments to
a variety of users and applications. Appropriate flavours and configurations can be
booted depending on the requirement, and in the same time the resources of a single
server can be shared while preserving isolation between the environments.
In order to optimize distributed resource sharing,...