We devise a performance model for GPU training of Deep Learning Recommendation Models (DLRM), whose GPU utilization is low compared to other well-optimized CV and NLP models. We show that both the device active time (the sum of kernel runtimes) and the device idle time are important components of the overall device time. We therefore tackle them separately by (1) flexibly adopting heuristic-based and ML-based kernel performance models for operators that dominate the device active time, and (2) categorizing operator overheads into five types to determine quantitatively their contribution to the device active time. Combining these two parts, we propose a critical-path-based algorithm to predict the per-batch training time of DLRM by traversing its execution graph. We achieve less than 10% geometric mean average error (GMAE) in all kernel performance modeling, and 5.23% and 7.96% geomean errors for GPU active time and overall end-to-end per-batch training time prediction, respectively. We show that our general performance model not only achieves low prediction error on DLRM, which has highly customized configurations and is dominated by multiple factors, but also yields comparable accuracy on other compute-bound ML models targeted by most previous methods. Using this performance model and graph-level data and task dependency analyses, we show our system can provide more general model-system co-design than previous methods.
翻译:我们为深学习建议模型(DLRM)的GPU培训设计了一个性能模型,其GPU利用率比其他优化的CV和NLP模型低,我们为深学习建议模型(DLRM)的GPU培训设计了一个性能模型。我们显示,设备运行时间(内核运行时间总和)和设备闲置时间都是整个设备运行时间的重要组成部分。因此,我们分别处理这些模型的办法是:(1) 灵活地为控制设备运行时间的操作者采用基于超光速和基于ML的内核性能模型,(2) 将操作者间接费用分为五种类型,以便量化地确定其对设备运行时间的贡献。结合这两部分,我们提出了基于关键路径的算法,以预测DLRM的每批培训时间(内核运行时间总和内核运行时间总和7.96%的地缘差培训时间预测),我们提出的一般性能模型不仅在数量上达到低预测错误,而且通过测试性能模型(MRMM)也采用高定制的多度分析方法。