论文标题
高性能计算时代的HEP实验数据存储
Data Storage for HEP Experiments in the Era of High-Performance Computing
论文作者
论文摘要
随着粒子物理实验将其限制在能量和强度边界上都限制时,所产生的数据的数量和复杂性也将相应增加。借助如此庞大的数据量,HL-LHC和Dune等下一代工作将更多地依赖于高通量(HTC)和高性能(HPC)计算簇。 HPC资源的全面利用需要可扩展有效的数据处理以及I/O。在过去的几十年中,大多数HEP实验都使用了根来存储数据。但是,其他存储技术(例如HDF5)在HPC环境中的性能可能更好。 HDF5的初始探索已经开始使用Atlas,CMS和Dune数据。 Dune实验还采用了HDF5来进行数据收购系统。本文介绍了HEP计算的未来前景和HPC的作用,以及正在进行的和未来的工作摘要,将HDF5用作HEP实验可能在HPC环境中使用的数据存储技术。
As particle physics experiments push their limits on both the energy and the intensity frontiers, the amount and complexity of the produced data are also expected to increase accordingly. With such large data volumes, next-generation efforts like the HL-LHC and DUNE will rely even more on both high-throughput (HTC) and high-performance (HPC) computing clusters. Full utilization of HPC resources requires scalable and efficient data-handling and I/O. For the last few decades, ROOT has been used by most HEP experiments to store data. However, other storage technologies like HDF5 may perform better in HPC environments. Initial explorations with HDF5 have begun using ATLAS, CMS and DUNE data; the DUNE experiment has also adopted HDF5 for its data-acquisition system. This paper presents the future outlook of the HEP computing and the role of HPC, and a summary of ongoing and future works to use HDF5 as a possible data storage technology for the HEP experiments to use in HPC environments.