-
Klaus SCHOSSMAIER (CERN)13/02/2006, 14:00Online Computingoral presentationThe data-acquisition software framework DATE for the ALICE experiment at the LHC has evolved over a period of several years. The latest version DATE V5 is geared for deployment during the test and commissioning phase. The DATE software is designed to runs on several hundred machines being installed with Scientific Linux CERN (SLC) to handle the data streams of approximatly 400 optical...Go to contribution page
-
Marco Pieri (University of California, San Diego, San Diego, California, USA)13/02/2006, 14:20Online Computingoral presentationThe CMS Data Acquisition system is designed to build and filter events originating from approximately 500 data sources from the detector at a maximum Level 1 trigger rate of 100 kHz and with an aggregate throughput of 100 GByte/sec. For this purpose different architectures and switch technologies have been evaluated. Events will be built in two stages: the first stage, the FED Builder,...Go to contribution page
-
Dr marc dobson (CERN)13/02/2006, 14:45Online Computingoral presentationThe needs of ATLAS experiment at the upcoming LHC accelerator, CERN, in terms of data transmission rates and processing power require a large cluster of computers (of the order of thousands) administrated and exploited in a coherent and optimal manner. Requirements like stability, robustness and fast recovery in case of failure impose a server-client system architecture with servers...Go to contribution page
-
Dr Beat Jost (CERN)13/02/2006, 15:05Online Computingoral presentationLHCb is one of the four experiments currently under construction at Cern's LHC accelerator. It is a single arm spectrometer designed to study CP violation the B-meson system with high precision. This paper will describe the LHCb online system, which consists of three sub-systems: - The Timing and Fast Control (TFC) system, responsible for distributing the clock and trigger decisions...Go to contribution page
-
Prof. Ryosuke ITOH (KEK)13/02/2006, 16:00Online Computingoral presentationThe Belle experiment, which is a B-factory experiment at KEK in Japan, is currently taking data with a DAQ system based on FASTBUS readout, switchless event building and higher level trigger(HLT) farm. To cope with a higher trigger rate from the expected sizeable increase in the accelerator luminosity in coming years, the upgrade of the DAQ system is in progress. FASTBUS modules are...Go to contribution page
-
Dr Hans G. Essel (GSI)13/02/2006, 16:20Online Computingoral presentationAt the upcoming new Facility for Antiproton and Ion Research FAIR at GSI the Compressed Baryonic Matter experiment CBM requires a new architecture of front-end electronics, data acquisition, and event processing. The detector systems of CBM are a Silicon Tracker System, RICH detectors, a TRD, RPCs, and an electromagnetic calorimeter. The envisioned interaction rate of 10~MHz produces a...Go to contribution page
-
Dr William Badgett (Fermilab)13/02/2006, 16:40Online Computingoral presentationThe CDF Experiment's control and configuration system consists of several database applications and supportive application interfaces in both Java and C++. The CDF Oracle database server runson a SunOS platform and provide both configuration data, real-time monitoring information and historical run conditions archiving. The Java applications running on the Scientific Linux operating system...Go to contribution page
-
Mr Sylvain Chapeland (CERN)13/02/2006, 17:00Online Computingoral presentationALICE (A Large Ion Collider Experiment) is the heavy-ion detector designed to study the physics of strongly interacting matter and the quark-gluon plasma at the CERN Large Hadron Collider (LHC). A large bandwidth and flexible Data Acquisition System (DAQ) is required to collect sufficient statistics in the short running time available per year for heavy ion and to accommodate very...Go to contribution page
-
Dr Benedetto Gorini (CERN)14/02/2006, 14:00Online Computingoral presentationThe Trigger and Data Acquisition system (TDAQ) of the ATLAS experiment at the CERN Large Hadron Collider is based on a multi-level selection process and a hierarchical acquisition tree. The system, consisting of a combination of custom electronics and commercial products from the computing and telecommunication industry, is required to provide an online selection power of 105 and a total...Go to contribution page
-
Mr Sebastian Neubert (Technical University Munich)14/02/2006, 14:25Online Computingoral presentationPANDA is a universal detector system, which is being designed in the scope of the FAIR-Project at Darmstadt, Germany and is dedicated to high precision measurements of hadronic systems in the charm quark mass region. At the HESR storage ring a beam of antiprotons will interact with internal targets to achieve the desired luminosity of 2x10^32cm^-2s^-1. The experiment is designed for event...Go to contribution page
-
Sebastian Robert Bablok (Department of Physics and Technology, University of Bergen, Norway)14/02/2006, 14:45Online Computingoral presentationThe HLT, integrating all major detectors of ALICE, is designed to analyse LHC events online. A cluster of 400 to 500 dual SMP PCs will constitute the heart of the HLT system. To synchronize the HLT with the other online systems of ALICE (Data Acquisition (DAQ), Detector Control System (DCS), Trigger (TRG)) the Experiment Control System (ECS) has to be interfaced. In order to do so, the...Go to contribution page
-
Gordon Watts (University of Washington)14/02/2006, 15:05Online Computingoral presentationDร, one of two collider experiments at Fermilab's Tevatron, upgraded its DAQ system for the start of Run II. The run started in March 2001, and the DAQ system was fully operational shortly afterwards. The DAQ system is a fully networked system based on Single Board Computers (SBCs) located in VME readout crates which forward their data to a 250 node farm of commodity processors for trigger...Go to contribution page
-
Dr Da-Peng JIN (IHEP (Institute of High Energy Physics, Beijing, China))14/02/2006, 16:00Online Computingoral presentationPhysical study is the base of the hardware designs of the BES3 trigger system. It includes detector simulations, generation and optimization of the sub-detectorsโ trigger conditions, main trigger simulations (Combining the trigger conditions from different detectors to find out the trigger efficiencies of the physical events and the rejection factors of the backgrounds events.) and...Go to contribution page
-
Hans von der Schmitt (MPI for Physics, Munich), Hans von der Schmitt (ATLAS)14/02/2006, 16:20Online Computingoral presentationThe ATLAS detector at CERN's LHC will be exposed to proton-proton collisions at a nominal rate of 1 GHz from beams crossing at 40 MHz. A three-level trigger system will select potentially interesting events in order to reduce this rate to about 200 Hz. The first trigger level is implemented in custom-built electronics and firmware, whereas the higher trigger levels are based on software. A...Go to contribution page
-
Mr Gianluca Comune (Michigan State University)14/02/2006, 16:40Online Computingoral presentationThis paper descibes an analysis and conceptual design for the steering of the ATLAS High Level Trigger (HLT). The steering is the framework that organises the event selection software. It implements the key event selection strategies of the ATLAS trigger, which are designed to minimise processing time and data transfers: reconstruction within regions of interest, menu-driven selection and...Go to contribution page
-
Kostas Kordas (Laboratori Nazionali di Frascati (LNF))14/02/2006, 17:00Online Computingoral presentationThe ATLAS experiment at the LHC will start taking data in 2007. Event data from protonโproton collisions will be selected in a three level trigger system which reduces the initial bunch crossing rate of 40 MHz at its first level trigger (LVL1) to 75 kHz with a fixed latency of 2.5 ฮผs. The second level trigger (LVL2) collects and analyses Regions of Interest (RoI) identified by LVL1 and...Go to contribution page
-
Dr Eric van Herwijnen (CERN)15/02/2006, 14:00Online Computingoral presentationLHCb has an integrated Experiment Control System (ECS), based on the commercial SCADA system PVSS. The novelty of this control system is that, in addition to the usual control and monitoring of all experimental equipment, it also provides control and monitoring for software processes, namely the on-line trigger algorithms. The trigger decisions are computed by algorithms on an event...Go to contribution page
-
Dr Wainer Vandelli (Universitร and INFN Pavia)15/02/2006, 14:20Online Computingoral presentationATLAS is one of the four experiments under construction along the Large Hadron Collider (LHC) ring at CERN. The LHC will produce interactions at a center of mass energy equal to $\sqrt s~=~14~TeV$ at a $40~MHz$ rate. The detector consists of more than 140 million electronic channels. The challenging experimental environment and the extreme detector complexity impose the necessity of a...Go to contribution page
-
Mrs Doris Burckhart (CERN)15/02/2006, 14:40Online Computingoral presentationThe Atlas Data Acquisition (DAQ) and High Level Trigger (HLT) software system will be comprised initially of 2000 PC nodes which take part in the control, event readout, second level trigger and event filter operations. This high number of PCs will only be purchased before data taking in 2007. The large CERN IT lxbatch facility provided the opportunity to run in July 2005 online...Go to contribution page
-
Dr Christos Leonidopoulos (CERN)15/02/2006, 15:00Online Computingoral presentationThe Physics and Data Quality Monitoring framework (DQM) aims at providing a homogeneous monitoring environment across various applications related to data taking at the CMS experiment. Initially developed as a monitoring application for the 1000 dual-CPU box (High-Level) Trigger Farm, it quickly expanded its scope to accommodate different groups across the experiment. The DQM organizes the...Go to contribution page
-
Andrei Kazarov (Petersburg Nuclear Physics Institute (PNPI))15/02/2006, 16:00Online Computingoral presentationIn order to meet the requirements of ATLAS data taking, the ATLAS Trigger-DAQ system is composed of O(1000) of applications running on more than 2000 computers in a network. With such system size, s/w and h/w failures are quite often. To minimize system downtime, the Trigger-DAQ control system shall include advanced verification and diagnostics facilities. The operator should use tests and...Go to contribution page
-
Dr Benedetto Gorini (CERN)15/02/2006, 16:20Online Computingoral presentationThis paper introduces the Log Service, developed at CERN within the ATLAS TDAQ/DCS framework. This package remedies the long standing problem of attempting to direct messages to the standard output and/or error in diskless nodes with no terminal. The Log Service provides a centralized mechanism for archiving and retrieving qualified information (Log Messages) created by TDAQ applications...Go to contribution page
-
Dr Christopher Pinkenburg (BROOKHAVEN NATIONAL LABORATORY)15/02/2006, 16:40Online Computingoral presentationThe PHENIX experiment took 2*10^9 CuCu events and more than 7*10^9 pp events during Run5 of RHIC. The total stored raw data volume was close to 500 TB. Since our DAQ bandwidth allowed us to store all events selected by the low level triggers, we did not filter events with an online processor farm which we refer to as level 2 trigger. Instead we ran the level 2 triggers offline in the...Go to contribution page
-
Dr Robert Bainbridge (Imperial College London)15/02/2006, 17:00Online Computingoral presentationThe CMS silicon strip tracker (SST), comprising a sensitive area of over 200m2 and 10M readout channels, is unprecedented in its size and complexity. The readout system is based on a 128-channel analogue front-end ASIC, optical readout and an off-detector VME board, using FPGA technology, that performs digitization, zero suppression and data formatting before forwarding the detector data...Go to contribution page
-
Dr Catalin Meirosu (CERN and "Politehnica" Bucharest)16/02/2006, 14:00Online Computingoral presentationThe Trigger and Data Acquisition System of the ATLAS experiment is currently being installed at CERN. A significant amount of computing resources will be deployed in the Online computing system, in the close proximity of the ATLAS detector. More than 3000 high-performance computers will be supported by networks composed of about 200 Ethernet switches. The architecture of the networks was...Go to contribution page
-
Mr Michael DePhillips (BROOKHAVEN NATIONAL LABORATORY)16/02/2006, 14:20Online Computingoral presentationThe STAR experiment at Brookhaven National Laboratory's Relativistic Heavy-Ion Collider (RHIC) has been accumulating 100's of millions events over its already 5 years running program. Within a growing Physics demand for statistics, STAR has more than doubled the events taken each year and is planning to increase its capability by an order of magnitude to reach billion events capabilities...Go to contribution page
-
Piotr Golonka (CERN, IT/CO-BE)16/02/2006, 14:40Online Computingoral presentationThe control systems of the LHC experiments are built using the common commercial product: PVSS II (from the ETM company). The JCOP Framework Project delivers a set of common tools built on top of, or extending the functionality of, PVSS (such as the control for widely used hardware, a Finite State Machine (FSM) toolkit, access control management, cooling and ventilation application)...Go to contribution page
Choose timezone
Your profile timezone: