The problem of learning parallel computer performance is investigated in the context of multicore processors. Given a fixed workload, the effect of varying system configuration on performance is sought. Conventionally, the performance speedup due to a single resource enhancement is formulated using Amdahl's law. However, in case of multiple configurable resources the conventional formulation results in several disconnected speedup equations that cannot be combined together to determine the overall speedup. To solve this problem, we propose to (1) extend Amdahl's law to accommodate multiple configurable resources into the overall speedup equation, and (2) transform the speedup equation into a multivariable regression problem suitable for machine learning. Using experimental data from fifty-eight tests spanning two benchmarks (SPECCPU 2017 and PCMark 10) and four hardware platforms (Intel Xeon 8180M, AMD EPYC 7702P, Intel CoffeeLake 8700K, and AMD Ryzen 3900X), analytical models are developed and cross-validated. Findings indicate that in most cases, the models result in an average cross-validated accuracy higher than 95%, thereby validating the proposed extension of Amdahl's law. The proposed methodology enables rapid generation of multivariable analytical models to support future industrial development, optimization, and simulation needs.
翻译:学习平行计算机性能的问题由多核心处理器来调查。在固定工作量的情况下,将寻求不同系统配置对性能的影响。在公约中,利用Amdahl的法律来制定单一资源增强的加速性能。但是,在多种可配置资源的情况下,常规配制产生若干不相连的加速等式,无法结合来决定总体速度。为了解决这个问题,我们提议:(1) 扩大Amdahl的法律,将多种可配置资源纳入全面加速方程式,(2) 将加速方程式转化为适合机器学习的多变回归问题。利用涵盖两个基准(SPECCPU-2017和PCmark 10)和四个硬件平台(Intel Xeon 8180M、AMDYC 7702P、Intel CoffoLake 8700K和AMD Ryzen 3900X)的实验数据,开发并交叉验证分析模型。调查结果表明,在大多数情况下,模型的结果是平均交叉精确度高于95%的精确度,而PCM 10) 和四个硬件平台(Intel Xe Xe Xe,从而实现了拟议的快速分析模型的扩展模式。