T.M. Steinbeck
(KIRCHHOFF INSTITUTE OF PHYSICS, RUPRECHT-KARLS-UNIVERSITY HEIDELBERG, for the Alice Collaboration)
27/09/2004, 14:00
Track 1 - Online Computing
oral presentation
The Alice High Level Trigger (HLT) is foreseen to consist of a
cluster of 400 to 500 dual SMP PCs at the start-up of the
experiment. It's input data rate can be up to 25GB/s. This has to be
reduced to at most 1.2 GB/s before the data is sent to DAQ through
event selection, filtering, and data compression. For these
processing purposes, the data is passed through the cluster in...
M. Sutton
(UNIVERSITY COLLEGE LONDON)
27/09/2004, 14:20
Track 1 - Online Computing
oral presentation
The architecture and performance of the ZEUS Global Track Trigger
(GTT) are described. Data from the ZEUS silicon Micro Vertex
detector's HELIX readout chips, corresponding to 200k channels, are
digitized by 3 crates of ADCs and PowerPC VME board computers push
cluster data for second level trigger processing and strip data for
event building via Fast and GigaEthernet network...
A. Di Mattia
(INFN)
27/09/2004, 14:40
Track 1 - Online Computing
oral presentation
The Atlas Level-2 trigger provides a software-based event selection
after the initial Level-1 hardware trigger. For the muon events, the
selection is decomposed in a number of broad steps: first, the Muon
Spectrometer data are processed to give physics quantities
associated to the muon track (standalone features extraction) then,
other detector data are used to refine the extracted...
M. Ye
(INSTITUTE OF HIGH ENERGY PHYSICS, ACADEMIA SINICA)
27/09/2004, 15:00
Track 1 - Online Computing
oral presentation
This article introduces a Embedded Linux System based on vme series
PowerPC as well as the base method on how to establish the system.
The goal of the system is to build a test system of VMEbus device.
It also can be used to setup the data acquisition and control
system. Two types of compiler are provided by the developer system
according to the features of the system and the...
G. CHEN
(COMPUTING CENTER,INSTITUTE OF HIGH ENERGY PHYSICS,CHINESE ACADEMY OF SCIENCES)
27/09/2004, 15:20
Track 1 - Online Computing
oral presentation
BES is an experiment on Beijing Electron-Positron Collider (BEPC).
BES computing environment consists of PC/Linux cluster and mainly relies on the free
software. OpenPBS and Ganglia are used as job schedule and monitor system. With
helps from CERN IT Division, CASTOR was implemented as storage management system.
BEPC is being upgraded and luminosity will increase one hundred times...
H-J. Mathes
(FORSCHUNGSZENTRUM KARLSRUHE, INSTITUT FรผR KERNPHYSIK)
27/09/2004, 15:40
Track 1 - Online Computing
oral presentation
S.Argiro`(1), A. Kopmann (2), O.Martineau (2), H.-J. Mathes (2)
for the Pierre Auger Collaboration
(1) INFN, Sezione Torino
(2) Forschungszentrum Karlsruhe
The Pierre Auger Observatory currently under construction in Argentina will
investigate extensive air showers at energies above 10^18 eV. It
consists of a ground array of 1600 Cherenkov water detectors and 24
fluorescence...
I. Sourikova
(BROOKHAVEN NATIONAL LABORATORY)
27/09/2004, 16:30
Track 1 - Online Computing
oral presentation
To benefit from substantial advancements in Open Source database
technology and ease deployment and development concerns with
Objectivity/DB, the Phenix experiment at RHIC is migrating its principal
databases from Objectivity to a relational database management system
(RDBMS). The challenge of designing a relational DB schema to store a
wide variety of calibration classes was
...
D. Winter
(COLUMBIA UNIVERSITY)
27/09/2004, 16:50
Track 1 - Online Computing
oral presentation
The PHENIX detector consists of 14 detector subsystems. It is designed such
that individual subsystems can be read out independently in parallel as well
as a single unit. The DAQ used to read the detector is a highly-pipelined
parallel system. Because PHENIX is interested in rare physics events, the DAQ
is required to have a fast trigger, deep buffering, and very high bandwidth.
The...
D Chapin
(Brown University)
27/09/2004, 17:10
Track 1 - Online Computing
oral presentation
The DZERO Level 3 Trigger and Data Aquisition (L3DAQ) system has been
running continuously since Spring 2002. DZERO is loacated at one of the
two interaction points in the Fermilab Tevatron Collider. The L3DAQ
moves front-end readout data from VME crates to a trigger processor
farm. It is built upon a Cisco 6509 Ethernet switch, standard PCs, and
commodity VME single board computers. We...
M. ZUREK
(CERN, IFJ KRAKOW)
27/09/2004, 17:30
Track 1 - Online Computing
oral presentation
The talk presents the experience gathered during the testbed
administration (~100 PC and 15+ switches) for the ATLAS Experiment at
CERN.
It covers the techniques used to resolve the HW/SW conflicts, network
related problems, automatic installation and configuration of the
cluster nodes as well as system/service monitoring in the heterogeneous
dynamically changing...
M. Dobson
(CERN)
27/09/2004, 17:50
Track 1 - Online Computing
oral presentation
The ATLAS collaboration had a Combined Beam Test from May until
October 2004. Collection and analysis of data required integration
of several software systems that are developed as prototypes for
the ATLAS experiment, due to start in 2007. Eleven different detector
technologies were integrated with the Data Acquisition system and were
taking data synchronously. The DAQ was integrated...
G. unel
(UNIVERSITY OF CALIFORNIA AT IRVINE AND CERN)
27/09/2004, 18:10
Track 1 - Online Computing
oral presentation
The ATLAS Trigger and DAQ system is designed to use the Region of
Interest (RoI)mechanism to reduce the initial Level 1 trigger rate of
100 kHz down to about 3.3 kHz Event Building rate.
The DataFlow component of the ATLAS TDAQ system is responsible
for the reading of the detector specific electronics via 1600 point
to point readout links, the collection and provision of RoI to the...
R. Itoh
(KEK)
29/09/2004, 14:00
Track 1 - Online Computing
oral presentation
A sizeable increase in the machine luminosity of KEKB accelerator is expected in
coming years. This may result in a shortage in the data storage resource for the Belle
experiment in the near future and it is desired to reduce the data flow as much as
possible before writing the data to the storage device.
For this purpose, a realtime event reconstruction farm has been installed in...
M. Richter
(Department of Physics and Technology, University of Bergen, Norway)
29/09/2004, 14:20
Track 1 - Online Computing
oral presentation
The ALICE experiment at LHC will implement a High Level Trigger
System, where the information from all major detectors are combined,
including the TPC, TRD, DIMUON, ITS etc. The largest computing
challenge is imposed by the TPC, requiring realtime pattern
recognition. The main task is to reconstruct the tracks in the TPC,
and in a final stage combine the tracking information from all...
P. Sheldon
(VANDERBILT UNIVERSITY)
29/09/2004, 14:40
Track 1 - Online Computing
oral presentation
The BTeV experiment, a proton/antiproton collider experiment at the Fermi National
Accelerator Laboratory, will have a trigger that will perform complex computations
(to reconstruct vertices, for example) on every collision (as opposed to the more
traditional approach of employing a first level hardware based trigger). This
trigger requires large-scale fault adaptive embedded software: ...
R. Panse
(KIRCHHOFF INSTITUTE FOR PHYSICS - UNIVERSITY OF HEIDELBERG)
29/09/2004, 15:00
Track 1 - Online Computing
oral presentation
Super-computers will be replaced more and more by PC cluster
systems. Also future LHC experiments will use large PC clusters.
These clusters will consist of off-the-shelf PCs, which in general
are not built to run in a PC farm. Configuring, monitoring and
controlling such clusters requires a serious amount of time
consuming and administrative effort.
We propose a cheap and easy...
A. Campbell
(DESY)
29/09/2004, 15:20
Track 1 - Online Computing
oral presentation
We present the scheme in use for online high level
filtering, event reconstruction and classification
in the H1 experiment at HERA since 2001.
The Data Flow framework ( presented at CHEP2001 ) will
be reviewed. This is based on CORBA for all data transfer,
multi-threaded C++ code to handle the data flow and
synchronisation and fortran code for reconstruction and
event selection. A...
T. Shears
(University of Liverpool)
29/09/2004, 15:40
Track 1 - Online Computing
oral presentation
The Level 1 and High Level triggers for the LHCb experiment are
software triggers which will be implemented on a farm of about 1800
CPUs, connected to the detector read-out system by a large Gigabit
Ethernet LAN with a capacity of 8 Gigabyte/s and some 500 Gigabit
Ethernet links. The architecture of the readout network must be
designed to maximise data throughput, control data flow,...
V. Gyurjyan
(Jefferson Lab)
29/09/2004, 16:30
Track 1 - Online Computing
oral presentation
A general overview of the Jefferson Lab data acquisition run control system is presented.
This run control system is designed to operate the configuration, control, and
monitoring of all Jefferson Lab experiments. It controls data-taking activities by
coordinating the operation of DAQ sub-systems, online software components and
third-party software such as external slow control...
F. Carena
(CERN)
29/09/2004, 16:50
Track 1 - Online Computing
oral presentation
The Experiment Control System (ECS) is the top level of control of the ALICE
experiment.
Running an experiment implies performing a set of activities on the online systems
that control the operation of the detectors. In ALICE, online systems are the
Trigger, the Detector Control Systems (DCS), the Data-Acquisition System (DAQ) and
the High-Level Trigger (HLT).
The ECS provides a...
D. Liko
(CERN)
29/09/2004, 17:10
Track 1 - Online Computing
oral presentation
The unprecedented size and complexity of the ATLAS TDAQ system
requires a comprehensive and flexible control system. Its role
ranges from the so-called run-control, e.g. starting and stopping
the datataking, to error handling and fault tolerance. It also
includes intialisation and verification of the overall system.
Following the traditional approach a hierachical system of...
G. Watts
(UNIVERSITY OF WASHINGTON)
29/09/2004, 17:30
Track 1 - Online Computing
oral presentation
The DZERO Collider Expermient logs many of its Data Aquisition Monitoring
Information in long term storage. This information is most frequently used to
understand shift history and efficiency. Approximately two kilobytes of information
is stored every 15 second. We describe this system and the web interface provided.
The current system is distributed, running on Linux for the back end...
L. Abadie
(CERN)
29/09/2004, 17:50
Track 1 - Online Computing
oral presentation
The aim of the LHCb configuration database is to store all the
controllable devices of the detector. The experimentโs control system
(that uses PVSS) will configure, start up and monitor the detector
from the information in the configuration database. The database will
contain devices with their properties, connectivity and hierarchy. The
ability to rapidly store and retrieve huge amounts...
T.M. Steinbeck
(KIRCHHOFF INSTITUTE OF PHYSICS, RUPRECHT-KARLS-UNIVERSITY HEIDELBERG, for the Alice Collaboration)
29/09/2004, 18:10
Track 1 - Online Computing
oral presentation
The Alice High Level Trigger (HLT) cluster is foreseen to consist of
400 to 500 dual SMP PCs at the start-up of the experiment. The
software running on these PCs will consist of components
communicating via a defined interface, allowing flexible software
configurations. During Alice's operation the HLT has to be
continuously active to avoid detector dead time. To ensure that the...