Graphics Processing Units (GPUs) were once used solely for graphical computation tasks but with the increase in the use of machine learning applications, the use of GPUs to perform general-purpose computing has increased in the last few years. GPUs employ a massive amount of threads, that in turn achieve a high amount of parallelism, to perform tasks. Though GPUs have a high amount of computation power, they face the problem of cache contention due to the SIMT model that they use. A solution to this problem is called "cache bypassing". This paper presents a predictive model that analyzes the access patterns of various machine learning algorithms and determines whether certain data should be stored in the cache or not. It presents insights on how well each model performs on different datasets and also shows how minimizing the size of each model will affect its performance The performance of most of the models were found to be around 90% with KNN performing the best but not with the smallest size. We further increase the features by splitting the addresses into chunks of 4 bytes. We observe that this increased the performance of the neural network substantially and increased the accuracy to 99.9% with three neurons.
翻译:图形处理器( GPUs) 曾经仅用于图形化计算任务, 但随着机器学习应用程序的使用增加, 使用 GPUs 进行一般用途计算的情况在过去几年中有所增加。 GPUs 使用大量线条, 从而实现大量的平行, 执行任务。 虽然 GPUs 具有很高的计算能力, 但是由于他们使用的 SIMT 模型, 他们面临着缓存争议问题。 解决这个问题的解决方案被称为“ 缓冲绕行 ” 。 本文展示了一个预测模型, 分析各种机器学习算法的存取模式, 并确定某些数据是否应该存储在缓存中。 它揭示了每个模型在不同数据集上的表现有多好, 并展示了如何将每个模型的大小最小化影响其性能。 发现大多数模型的性能大约为90%, KNNN 运行最佳, 但没有最小的大小。 我们通过将地址拆分成4 字块来进一步增加特性。 我们观察到, 神经网络的性能大幅提高, 并将精度提高到99. 99% 和3 神经 。