Unique scientific instruments designed and operated by large global collaborations are expected to produce Exabyte-scale data volumes per year by 2030. These collaborations depend on globally distributed storage and compute to turn raw data into science. While all of these infrastructures have batch scheduling capabilities to share compute, Research and Education networks lack those capabilities. There is thus uncontrolled competition for bandwidth between and within collaborations. As a result, data "hogs" disk space at processing facilities for much longer than it takes to process, leading to vastly over-provisioned storage infrastructures. Integrated co-scheduling of networks as part of high-level managed workflows might reduce these storage needs by more than an order of magnitude. This paper describes such a solution, demonstrates its functionality in the context of the Large Hadron Collider (LHC) at CERN, and presents the next-steps towards its use in production.
翻译:由大型全球协作设计和操作的独特科学仪器预计将在2030年之前每年产生Exabbyte规模的数据量,这些协作取决于全球分布的储存量和计算将原始数据转化为科学。所有这些基础设施都具有分批列表能力,可以共同计算,但研究和教育网络缺乏这些能力。因此,在协作之间和协作内部对带宽的竞争不受控制。结果,加工设施的数据“猪”磁盘空间比处理要长得多,导致大量过剩的储存基础设施。作为高层管理工作流程一部分的网络一体化联合安排可能会减少这些储存需求,而减少数量可能超过一个数量级。本文描述了这种解决方案,在法国核电离心电离心电流大型对撞机(LHC)背景下显示了其功能,并介绍了在生产中使用它的下一个步骤。