Sparse tensors are prevalent in many data-intensive applications, yet existing differentiable programming frameworks are tailored towards dense tensors. This presents a significant challenge for efficiently computing gradients through sparse tensor operations, as their irregular sparsity patterns can result in substantial memory and computational overheads. In this work, we introduce a novel framework that enables the efficient and automatic differentiation of sparse tensors, addressing this fundamental issue. Our experiments demonstrate the effectiveness of the proposed framework in terms of performance and scalability, outperforming state-of-the-art frameworks across a range of synthetic and real-world datasets. Our approach offers a promising direction for enabling efficient and scalable differentiable programming with sparse tensors, which has significant implications for numerous applications in machine learning, natural language processing, and scientific computing.
翻译:许多数据密集型应用中普遍存在粗散的沙粒体,但现有的可区别的编程框架是针对密集的沙粒体设计的,这对通过稀疏的沙粒体操作高效计算梯度提出了重大挑战,因为其不规则的聚变模式可能导致大量的记忆和计算间接费用。在这项工作中,我们引入了一个新颖的框架,使稀散的沙粒体能够高效和自动地区分,解决这一根本问题。我们的实验表明,拟议框架在一系列合成和现实世界数据集的性能和可扩展性方面是有效的,优于最先进的框架。我们的方法为与稀散的沙粒体进行高效和可扩展的不同编程提供了有希望的方向,这对机器学习、自然语言处理和科学计算方面的许多应用具有重大影响。</s>