Machine learning (ML) is successful in achieving human-level performance in various fields. However, it lacks the ability to explain an outcome due to its black-box nature. While existing explainable ML is promising, almost all of these methods focus on formatting interpretability as an optimization problem. Such a mapping leads to numerous iterations of time-consuming complex computations, which limits their applicability in real-time applications. In this paper, we propose a novel framework for accelerating explainable ML using Tensor Processing Units (TPUs). The proposed framework exploits the synergy between matrix convolution and Fourier transform, and takes full advantage of TPU's natural ability in accelerating matrix computations. Specifically, this paper makes three important contributions. (1) To the best of our knowledge, our proposed work is the first attempt in enabling hardware acceleration of explainable ML using TPUs. (2) Our proposed approach is applicable across a wide variety of ML algorithms, and effective utilization of TPU-based acceleration can lead to real-time outcome interpretation. (3) Extensive experimental results demonstrate that our proposed approach can provide an order-of-magnitude speedup in both classification time (25x on average) and interpretation time (13x on average) compared to state-of-the-art techniques.
翻译:机器学习(ML)在各个领域都成功地实现了人文水平的绩效。然而,由于黑盒的性质,它缺乏解释结果的能力。虽然现有的可解释ML很有希望,但几乎所有这些方法都侧重于将可解释性格式化为优化问题。这种绘图导致大量重复耗时的复杂计算,限制了其在实时应用中的适用性。在本文件中,我们提出了一个新的框架,以加速使用Tensor处理器(TPU)解释可解释的 ML。拟议框架利用矩阵变换和Fourierer变换之间的协同作用,充分利用TPU的自然能力加速矩阵计算。具体地说,本文件作出了三项重要的贡献。 (1) 就我们所知的最佳而言,我们拟议的工作是首次尝试使可解释ML的硬件加速使用TPU。 (2) 我们的拟议方法适用于范围广泛的多种ML算法,有效利用基于TPU的加速可以导致实时的结果解释。(3)广泛的实验结果表明,我们拟议的方法可以在两个分类(13x)平均和平均时间判读(13x)平均时间和平均判读(13x)平均判时速技术。