Speaker
Description
Large-scale scientific computing relies on cost-effective, high-capacity storage systems to support data-intensive workloads , such as those from the Worldwide LHC Computing Grid and future data-intensive sciences like the Square Kilometre Array Observatory. At STFC, we evaluated three Ceph-based storage configurations – 8TB HDD, 22TB HDD, and 15TB TLC NVMe flash. Using low level benchmarks across varied erasure coding and replicated pool layouts, we measured performance and power efficiency under varied workloads. Results show NVMe delivers superior small I/O performance and better IOPS-per-watt, while dense HDD remains cost and power-effective for capacity-driven workloads. However, idle power dominates overall energy consumption, making device density a key factor in reducing operational costs. We discuss trade-offs between cost, performance, and sustainability, highlighting erasure coding layout impacts and the emerging viability of QLC flash. We also draw comparisons with the at-rest power consumption of our tape storage. These findings could guide procurement strategies for large-scale scientific infrastructures seeking to optimize performance and energy efficiency.