Alice Weekly Meeting: Software for Hardware Accelerators / PDP-SRC - MINUTES ONLY
-
-
10:00 AM
→
10:20 AM
Discussion 20mSpeakers: David Rohr (CERN), Giulio Eulisse (CERN)
-
10:20 AM
→
10:25 AM
Following up JIRA tickets 5mSpeaker: Ernst Hellbar (CERN)
-
10:25 AM
→
10:30 AM
TPC ML Clustering 5mSpeaker: Christian Sonnabend (CERN, Heidelberg University (DE))
- O2 & alidist are settled for ONNX execution. Both build- & runtime variables are now there, all checks passed, ready to merge from my side: https://github.com/alisw/alidist/pull/5622
- Had some issues with GSI (cluster reservation went end-of-life, had to be restored for O2 builds), redoing simulations / evaluations for different collision systems and interaction rates now
- Started working on separation power and Bethe-Bloch fit (left: GPU CF, right: NN)
- Caveats: Way to few electrons in either case to make a reliable fit: Need to generate more data
- Plot 1: dE/dx spectra with default LHC18b parameterization. Only MIP scaling changed from 50 (GPU CF) to 60 (NN), otherwise the parameters are the same.
- Plot 2: Separation power at ~MIP (p = 0.3 to 0.4 GeV/c). Gaussian fits were constrained (at least the electron fit, red). But pion band width (relative dE/dx resolution) is improved from 12.6% (GPU CF) to 11.5% (NN), to be confirmed with more data.
- Plot 3: Default parameterization still looks a lot healthier than the fit since electrons cannot yet constrain the Fermi plateau (more data).
Planning ahead
- Currently working on V0 reco: K0S and Lambda mass peaks
- Soon checking again the dE/dx calibration of the MIP -> There were some commits by Jens on O2 about this
- In the meantime: Evaluate NN on the pp and PbPb simulations
- Today / tomorrow: Preparing GSI presentation for friday and next tuesday
-
10:30 AM
→
10:35 AM
ITS Tracking 5mSpeaker: Matteo Concas (CERN)
-
10:35 AM
→
10:45 AM
TPC Track Model Decoding on GPU 10mSpeaker: Gabriele Cimador (Universita e INFN Trieste (IT))
Global Parameter Optimisation
Input dataset simulation
Simulated several timeframes:
- pp: 100kHz, 200kHz, 500kHz, 1MHz, 2MHz
- PbPb: 10kHz, 15kHz, 20kHz, 27kHz, 35kHz, 42kHz, 47kHz, 50kHz
Every timeframe simulated twice, one for 32 orbits timeframe and one for 128 orbits timeframe
For the moment just one simulation per configuration (beam type - interaction rate - timeframe length)
GPU Parameters study
Focusing on grid and block size. Analysed the GPU workflow of the sync/async TPC processing. Image below is the workflow of two HIP streams of the sync TPC processing:
By looking at the tracefile:
- Clusterizer chain:
- small concurrent kernels
- overlap during execution
- overall taking considerable time
- --> dependent parameters, global optimisation
- SliceTracker chain:
- medium concurrent kernels
- all streams used
-
main kernel is TrackletConstructor
- trace file outputs that CreateSliceData takes a lot of time, however --debug does not say so, still investigating
- trace file outputs "Marker" which is not present in nvidia trace files, still investigating
- --> dependent parameters, global optimisation
- Merger chain:
-
MergeBorders_step2: lots of small concurrent kernels, concurrent to a limited set of other one stream kernels --> dependent parameters, global optimisation (within set)
-
SliceRefit: lots of small one stream kernels --> independent parameters, local optimisation
- MergerTrackFit: one stream long kernel --> independent parameters, local optimisation (maybe limited since values dependent also on number of tracks)
-
MergerFollowLoopers: one stream medium kernel --> independent parameters, local optimisation
-
- Compression/Decompression chain:
- One stream kernels --> independent parameters, local optimisation
- Multiple stream kernels, not overlapping --> independent parameters, local optimisation
Optimisation strategy
- For the moment just a "Manual Trial-and-Error" using observations from the output
- Started from MergerTrackFit, why:
- Long kernel
- One stream
- Not concurrent to any other kernels
- Caveat: grid size dependent on number of tracks
- Changing values in GPUDefGPUParameters.h takes a loooong time to compile, even with standalone benchmark
- Currently forcing custom krnlExec object in kernel calls, e.g.:
runKernel<KernelClass, KernelClass::step>({{n_blocks,n_threads,stream}});
- Not handy, but way faster
- Currently forcing custom krnlExec object in kernel calls, e.g.:
- Created script that automatically fetches grid and block size for all the kernels, useful for runtime grid/block numbers like GetGrid(Merger.NOutputTracks(), 0, deviceType)
Possible bug spotted
HIP_AMDGPUTARGET set to "default" in GPU/GPUTracking/Standalone/cmake/config.cmake translates in HIP_AMDGPUTARGET=gfx906;gfx908 and forces to use MI50 params
Basically here HIP_AMDGPUTARGET=gfx906;gfx908 enters the first if clause for MI50 even if I am compiling for MI100. Commented set(HIP_AMDGPUTARGET "default") on the config.cmake of the standalone benchmark and forced usage of MI100 parameters via
cmake -DCMAKE_INSTALL_PREFIX=../ -DHIP_AMDGPUTARGET="gfx908" ~/alice/O2/GPU/GPUTracking/Standalone/
Did not investigate further on this.
-
10:45 AM
→
10:55 AM
Efficient Data Structures 10mSpeaker: Dr Oliver Gregor Rietmann (CERN)
-
10:00 AM
→
10:20 AM