The CERN ATLAS experiment successfully uses a worldwide
computing infrastructure to support the physics program during LHC
Run 2. The grid workflow system PanDA routinely manages 250 to
500 thousand concurrently running production and analysis jobs
to process simulation and detector data. In total more than 300 PB
of data is distributed over more than 150 sites in the WLCG and
handled by the ATLAS data management system Rucio. To prepare
for the ever growing LHC luminosity in future runs new
developments are underway to even more efficiently use
opportunistic resources such as HPCs and utilize new technologies.
This presentation will review and explain the outline and the
performance of the ATLAS distributed computing system and give an
outlook to new workflow and data management ideas for the
beginning of the LHC Run 3.