In recent years, the growing size of neural networks has led to a vast amount of research concerning compression techniques to mitigate the drawbacks of such large sizes. Most of these research works can be categorized into three broad families : Knowledge Distillation, Pruning, and Quantization. While there has been steady research in this domain, adoption and commercial usage of the proposed techniques has not quite progressed at the rate. We present KD-Lib, an open-source PyTorch based library, which contains state-of-the-art modular implementations of algorithms from the three families on top of multiple abstraction layers. KD-Lib is model and algorithm-agnostic, with extended support for hyperparameter tuning using Optuna and Tensorboard for logging and monitoring. The library can be found at - https://github.com/SforAiDl/KD_Lib.
翻译:近年来,神经网络规模不断扩大,导致大量关于压缩技术的研究,以减轻如此庞大规模的缺陷。这些研究大部分可以分为三大类:知识蒸馏、普林宁和量化。虽然在这一领域进行了稳定研究,但采用和商业使用拟议技术的情况并没有按这个速度取得相当进展。我们介绍了基于开放源码的PyTorch图书馆KD-Lib,该图书馆包括三个家庭在多个抽象层之上最先进的算法模块实施。KD-Lib是模型和算法-算法-算术,对利用Optuna和Tensor板进行超参数调整进行伐木和监测提供了广泛的支持。图书馆可在https://github.com/SforAiDl/KD_Lib找到。