Alice off-line week

Europe/Zurich
160/R-009 (CERN)

160/R-009

CERN

20
Show room on map
Carminati, F.
    • 09:00 18:00
      Monday

      HLT Session

      • 09:00
        Introduction to the week 30m
        Speaker: Carminati,F. (CERN)
      • 09:30
        DAQ-HLT software interface 30m
        Speaker: Divia, R. (CERN)
        Roberto presented the DAQ architecture and the interface to the HLT. After the presentation there was a discussion on the possibility to have the HLT algorithms off-line. One can work on the events before they are assembled in the GDC, or after GDC with a full event. The second possibility is simpler from the software point of view, but depends on the speed and memory requirements of the program. It has to be investigated in details. The compression is another important question. Even in TPC there could be a possibility to send data without compression. The event size is not well known, there might be a factor of 3, and this is the reason to need possibility to compress. From the other hand, if we can avoid the compression, we should do it. This will speed up the processing, but for the moment the default is compressed data. Many things will depend on the noise level, etc.
      • 10:00
        Fast seeding for TPC 30m
        Speaker: Ajit, M. (CERN)
        Ajit described the Independent Component Analysis for HLT track finding. The principal component analysis has two phases: linear transformation and chi2 minimization. It is very fast. Is the performance of the Independent Component Analysis comparable? Yes, it is more or less the same. The same technique could be used for seeds in the normal track reconstruction. Could this method be used in another place apart from TPC? Probably in MUON detector. The slow tracks in ITS are more difficult question
      • 10:30
        Coffee Break 30m
      • 11:00
        Production of Raw data for TPC and ITS 20m
        Speaker: Favretto,D. (CERN)
        Dario showed a lossless compression based on the Hofman coding. There was a discussion on the performance of the code. The important question is how fast will be the decoding, because the HLT needs decompressed data to work. A similar investigation was done in the Bergen group, and the results were published. A FPGA chip will do the coding in the on-line. The present work was done in order to provide a realistic data source for the data challenges.
      • 11:20
        HLT tracking in the TPC 20m
        Speaker: Sandoval, A. (GSI)
        Andres presented the TPC fast on-line reconstruction based on the Hough transformation. A discussion on the efficiency concluded that the important number is which part of the good tracks is found. The number of fake tracks is not an issue, because one wants to keep as much good tracks as possible. The implementation of HLT (which goes to the hardware) has to be put also in the off-line framework.
      • 11:40
        Discussion on future activities in HLT 50m
        Speaker: Carminati,F. (CERN)
        Discussion on the HLT issues: Federico: first question - what can go to the data challenges, second question - how do we coordinate the future work. The next ADC is in October, so we haven't much time to work on. The DAQ people need the raw data (pure binary files), and then they can use them correctly. One doesn't need all the detectors in, for example ITS and TPC would be enough. This afternoon a small work-group can decide how to formate the raw data and provide it to the DAQ. - raw data arrangement for the data challenge 4 (Roberto, Ajit, Dario, Marian, Sandro, Bjorn, Mariana, Karel, Andres, Marek) A HLT day should be planned for each Alice off-line day. Each detector should revise the status of the HLT code. ITS and TPC have prepared already something, TRD is also following the question. There is some code from MUON detector.
      • 12:30
        Lunch 1h 30m
      • 15:00
        Working session on the raw data arangement for the data challenge 2h
        Speaker: Roberto, Ajit, Dario, Marian, Sandro, Bjorn, Mariana, Karel, Andres, Marek
    • 09:00 18:00
      Tuesday

      Virtual MC

      • 09:00
        Report from the previous day 15m
        Speaker: Masera, M. (INFN)
        more information
        transparencies
      • 09:15
        Status of Virtual MonteCarlo branch 15m
        Speaker: Hristov, P. (CERN)
        transparencies
      • 09:30
        Discussion on Virtual MonteCarlo 1h
        Speaker: Carminati, F. (CERN)
        Discussion on the compilers. One minimal requirement for RH 7.2,7.3 would be gcc 3.1. Federico explained the reason for the change he proposes. The code we deliver to the users should be easily installable. The Geant4 visualization is not a critical issue. Ivana. We should not disconnect possibilities without providing a replacement. The case of the main program has to be corrected, but we need people to work on. For the moment the backup solution is working and should be kept. Federico. If we have a bug in Geant4 libraries and Can we access G4 visualization vie VMC interface? Yes, you have access all the visualization modules provided in G4. It depends on the choice, so lets make one of them default and don't bother about the others. Marek. Rene could suggest the best visualization. Rene. The various visualization modules have different purpose, so it is difficult to select one of them with universal scope. Isidro. OpenGL is not sufficient. You need OpenGL+Motif. For VRML and DON we need external packages. VRL is very good. Federico. Remember, that for some users it is difficult. Ivana. We are using reflections and many of the visualizations don't support it. The choice has to take it in the account. Rene. The problem with the dynamical linking is probably in the order of the class initialization. It has to be fixed, either by G4 people or by us. Fed. Proposes to use OpenGL. We need the simplified version with the sufficient minimum of features. The full version also has to be kept and we will be happy to have it on the CVS server. Ivana. All you require can be achieved easily by changing the initialization script without touching the structure. Fed. It is much easier to debug the new package, it makes the maintenance easier. Ivana. The changes in the makefiles will be done very often in the geant4vmc, which is not the case in geant4_mc. We always will need some env. variables to configure. Marek It must be something wrong with the G4 if we are not able to configure easily. Federico. The main decision has to be that the G4 main program is removed, and we support one visualization module. Then we ship it to the users and if the installation passes the tests, everything is OK. Rene summarized the discussion between the experts. OpenScientist is difficult to install. OpenGL has good navigation, but no picking. It supports reflections. The OGLIXm version needs Motif or similar package, but some problems with the dynamical linking are expected. Probably we need to investigate it. DOWN is developed by Japan group 15 years ago for SSC. It supports no reflections. Raytracer has nice features, but is too slow. GraXML interface in geant4_mc. It is probably the best for 3D pictures. It looks simple for installation. VRML has no support for reflections, but it is going probably to appear. Federico proposes to start with OGLIX.
      • 10:30
        Coffee Break 30m
      • 11:00
        Report from the algorithm review 30m
        Speaker: Morsch, A. (CERN)
        transparencies
        Andreas presented the report from the algorithm review. Marek. The MUON resolution is too good, but probably the reason is that the fluctuations of the gas gain are not taken into account. It leads to an additional factor of Sqrt(2). PG. Simulation of SPD:The timing error was introduced and the comparison with the test beam date is better. The work is going on. Boris. SSD double peak problem - the double peak is seen also in the test beam data. Qualitatively the pictures are similar. It probably is related to the low energy electrons. The effect is also presented in the pixels. There it is masked by the low number of double clusters. Is the neutron capture in Xe put in the TRD simulation? The effect is less then 5% and there is a person working on the simulation.
      • 11:30
        Status of the signal reconstruction and PID with HMPID 30m
        Speaker: DiBari, D. (INFN)
        Nico presented the status of signal reconstruction and PID with HMPID. The backward track propagation (in TPC) includes the possibility to add a new point to the track. This is important and can give 50% better reconstruction (CLEO report). Since the yield of the particles depends on the centrality, the method works on a given class of events. 20% of the particles produce no MIP in the RICH. Guy. General question - when you talk about the Lambdas, could you think also about embedding protons? The result should be the same. Using the Lambda has the advantage that the tracks are reconstructed in the experiment. Question about the calibration and alignment are taken into account automatically. The problem is in the limited statistics.
      • 12:00
        Status of the fast simulation in the muon arm 20m
        Speaker: De Falco, A. (INFN)
        Alessando showed the fast tracking approach to the MUON detector. The resolutions in MC and test data are very different. One could use the Andrea Dainese parametrization with corrected TPC track parameters according to the test beam. This way the contribution from the track smearing is taken into account better. Boris. Is the noise simulation OK? It could be the reason for the difference. Andreas. The reconstruction of the test beam data is completely different from the one in AliRoot. They have to be made coherent for the comparison. Guy. The mass resolution to find the upsilon is about 100MeV. The realistic point gives 115MeV, which is not so bad, but is higher then the needed one.
      • 12:20
        Status of FMD 10m
        Speaker: Maevskaia, A. (INR)
        transparencies
      • 12:30
        Lunch 1h 30m
    • 09:00 18:00
      Wednesday

      GRID & Distributed computing

      • 09:30
        AliEn/Grid/Computing status in the US 15m
        Speaker: Nilsen, B. (OSU)
      • 09:45
        AliEn/Grid status at OSC 20m
        Speaker: Johnson, D. (OSC)
      • 10:05
        Update on Grid activities in Germany 10m
        Speaker: Schwarz, K. (GSI)
      • 10:15
        GRID status in YerPhi 15m
        Speaker: A. Grigorian (Yerevan)
      • 10:30
        Coffee Break 30m
      • 10:50
        EDG status 15m
        Speaker: Cerello, P. (INFN)
        transparencies
      • 11:05
        EDG 1.2: present GDMP functionality 20m
        Speaker: Barbera, R. (INFN)
      • 11:25
        Configuration of the Torino Farm as EDG UI, Ce and SE 20m
        Speaker: Sitta, M. (INFN)
        transparencies
      • 11:55
        Discussion 45m
      • 12:30
        Lunch 1h 30m
      • 14:30
        AliEn status: introduction 10m
        Speaker: P.Buncic (CERN)
        more information
        transparencies
      • 14:40
        New configuration class for AliEn, disk pool manager 10m
        Speaker: M. Horner (CERN)
        transparencies
      • 14:50
        Web based accounting and documentation tools, WDSL and UDDI for AliEn services 10m
        Speaker: T. Reisinger (CERN)
      • 15:00
        Bootstrapping the AliEn server; checking the file catalogue consistency 10m
        Speaker: Y. Bugaenko (JINR)
      • 15:10
        Discussion 30m
      • 15:40
        Coffee Break 20m
      • 16:00
        AliEn tutorial 1h
        Speaker: P. Buncic, P. Saiz (CERN)
        document
      • 17:00
        Discussion 1h
    • 09:00 18:00
      Thursday

      AliRoot developments

      • 09:00
        New IO a la PHOS/EMCAL 30m
        Speaker: Peressounko, D.
        Bjorn. Concern about the number of files. The 2Gb issue is not anymore relevant, probably one can keep everything in one event. Rene is in favor of the split file system. The only concern is when we have small events, they should be correctly put in groups. Fed. We take as order of magnitude a central event. Then the time to access files is negligible. If the dir. contains range of small events, we will keep the contain consistent with a central event. The avant age of the split files is that it minimizes possibility to make error and to corrupt the data. Jiri. Which class is responsible for the writing of general data? It is the main point for the discussion after the presentation. Piotr. Question about the situation when we have more then one event in the memory, for example during the Merging. Yuri - we need only the sDigits of the two events. By default we read the header of the signal event. For the merging I read also the sDigits. Marian. Is it possible to open the sDigits trees from the signal and the background event at the same time? Yes, it is. The name of the sDigits folder is taken from the full path name. There is one background event and several signals which are merged to, but they are not considered as events. For the signals one needs only sDigits.
      • 09:30
        Discussion on new I/O 1h
        Speaker: TBD
        Discussion: Piotr. How do you use your approach for several detectors? Yves. For every detector you can use a task What is the main difference: we have only one event (one gAlice object in the memory) We have only one getter. The runLoader is unique, and the DetectorLoaders are provided by each detector. The two approaches are similar, but the NewIO is somehow redundant. There is some things which cannot be done easily in the Piotr's approach. The problem was when the hits want to access the geometry. Poitr tried to use a fixed names in order to solve it (a folder name where you can get the information). If we fix the folder names and use only one gAlice, then the two approaches are very similar. In this case probably one can avoid the storage of folders. What is wrong with the storage of folders? It gives you a possibility to change for example the simulation. The overhead is negligible. Yves. The schema should do only the minimal set of required things, and everything one might imagine. Jiri. If the only question is the redundancy, then it is not very important, we already have a lot of redundant information. The Piotr's approach allows to change of the folder names, but the default system is similar. The NewIO has possibility for many event headers in the folder, which is not the case in the PHOS approach. PHOS NewIO EV: EV1: - Eheader - Eheader - SDigits1 - SDigits - SDigits2 - Hits - Hits - Digits - Digits EV2: - Eheader - SDigits - Digits - Hits Marian: In case of NewIO it is easier to access events and the question of cleaning is easier. Predrag: The NewIO also has better separation between the data structure and algorithm. Mixing is when you take tracks from several events to produce an uncorrelated background event. It is need for the analysis of HBT and fluctuations. When you do physics, you know what you mix with what you mix. Merging is the adding of signal events to a background event. At the level of merging both approaches have the same features. The HBT analysis is independent from the IO structure. The mixing probably will be done at the level ESD. The PHOS approach is you load data from the first event and select what you need, then Do we need two complete event structure (not full events) in the memory? Andreas. Having two headers is good for sanity checks, for the TPC you can compare the time structure, combine beam halo and end central event, etc. Nico. Analysis point of view. The first (PHOS) schema is more clear, even if the second (Piotr) has probably some advantages in certain cases. Yves: the analysis and the IO should be separated. The whole complexity has to be avoided, and the PHOS approach is simpler. Federico. Proposed a working session between the experts in order to converge on the NewIO pending questions. The work group decided to produce the requirements and the use cases for the IO: Requirements: - single and split files (trees/files) - single and multiple detector - access more than one event - access to all levels of the data structure - load data on demand (file -> memory) Use cases - Digitization (needs Header, Geometry/Field, Hit data) - Merging - Reconstruction - Analysis - Mixing New IO approach File Structure <-> RunLoader <-> Digitizer Both approaches can meet the requirements. The PHOS approach is still to be generalized, while the NewIO is already general enough. From a design point of view the hiding of the way we use the file structure is good. The work of the interface should continue.
      • 10:30
        Coffee Break 30m
      • 11:00
        Status of PHOS 30m
        Speaker: Kharlov, Y. (IN2P3)
      • 11:45
        New IO session 2h
      • 12:30
        Lunch 1h 30m
      • 14:30
        Discussion about EMCAL simulation, reconstruction, and analysis for the EMCAL proposal 1h
        Speaker: Nilsen, B. (OSU)
      • 15:30
        Status of geometrical modeler 30m
        Speaker: Gheata, A. (CERN)
        transparencies
      • 16:00
        Virtual MC session 1h
    • 09:00 18:00
      Friday

      Wrap-up session

      • 09:00
        Status of SPD simulation 20m
        Speaker: Nilsen, B. (OSU)
      • 09:20
        News about Virtual MC 10m
        Speaker: Hrivnacova, I. (Orsay)
      • 09:30
        Discussion on the calibrartion DB and calibrarion strategy 1h
      • 10:30
        Coffee Break 30m
      • 11:00
        Off-line board 1h 30m
        Computing and physics data challenges are distinguished only in Alice. DC 10/2003 - 09/2004 Classic one: simulation, reconstruction DC 10/2004 - 09/2006 Initially we supposed 10% reconstructed data Tier0 data ----> Tier1's It has disadvantage that supposes a huge Tier0 (financial problems, etc.) Now we propose to ship the raw data and make the reconstruction in Tier1's. This will require very high network bandwidth. The size of Alice Computing should not exceed those of ATLAS and CMS. The size for DC should lead to the final system. The planning exercise was done by Alberto. 2006 - bye equipment for 2007 20% 2007 mostly proton beam, less data => no need for the full system 2007 -//- 2008 30% Estimate the multiplicity of the heavy ion events and the need 2008 -//- 2009 100% Yves presented the document The physics objectivities were not completely met: - simulation->reconstruction chain (not for all detectors) - simple analysis tools - not available for most of the tasks - digits in Root format (done) The number from the last production: - the size of the simulate events is much bigger than what we reported in the Hoffman review. The reason is that we keep hits, clusters, etc., which will not be there for the real experiment. During the next data challenge we have to decide what do we keep. The simplest possibility it to use temporary hits and then remove it. To be done (for the next DC): - introduce VirtualMC - move to NewIO - introduce the calibration - implement the ESD classes - add trigger classes - make internal planning POOL - RTag on the persistency (IO) - the basic idea was to use Root, but hiding it via abstract interfaces. The first version should be released... BLUEPRINT - RTag on the global architecture of the LCG (LHC Computing Grid) software. Proposes the following: - Python as a scripting language - ANAPHE as an abstract interface to Root (and to Lite histograming package, using ZEBRA) - Root for IO There was a discussion on the Alice position. Simulation RTag: VirtualMC and the geom. modeler were presented. ATLAS is interested in the virtual MC.