The task of compressing pre-trained Deep Neural Networks has attracted wide interest of the research community due to its great benefits in freeing practitioners from data access requirements. In this domain, low-rank approximation is a promising method, but existing solutions considered a restricted number of design choices and failed to efficiently explore the design space, which lead to severe accuracy degradation and limited compression ratio achieved. To address the above limitations, this work proposes the SVD-NAS framework that couples the domains of low-rank approximation and neural architecture search. SVD-NAS generalises and expands the design choices of previous works by introducing the Low-Rank architecture space, LR-space, which is a more fine-grained design space of low-rank approximation. Afterwards, this work proposes a gradient-descent-based search for efficiently traversing the LR-space. This finer and more thorough exploration of the possible design choices results in improved accuracy as well as reduction in parameters, FLOPS, and latency of a CNN model. Results demonstrate that the SVD-NAS achieves 2.06-12.85pp higher accuracy on ImageNet than state-of-the-art methods under the data-limited problem setting. SVD-NAS is open-sourced at https://github.com/Yu-Zhewen/SVD-NAS.
翻译:压缩经过培训的深神经网络的任务已吸引了研究界的广泛兴趣,因为它在使开业者摆脱数据访问要求方面大有裨益。在这方面,低级近似是一个很有希望的方法,但现有解决方案考虑了有限的设计选择,未能有效地探索设计空间,从而导致严重精度退化和压缩率的提高。为解决上述局限性,这项工作提议SVD-NAS框架,将低级近距离和神经结构搜索领域结合起来。SVD-NAS通俗化并扩大了以前作品的设计选择,引入了低兰克建筑空间、LR-空间,这是一个更精细的低级近似设计空间。随后,这项工作提议以梯度白度为基础搜索,以高效地完成LLR-空间的穿孔。对可能的设计选择进行精度和降低参数、FLOPS和CNNM模型的Lative。结果显示,SVD-NAS-NAS达到2.06-12/D85pp-LOFS-FIFS-FAL-S-S-ODS-OFS-FS-FS-FS-FIGS-S-S-S-OD-OD-OD-OD-S-S-S-SOD-S-SOD-SOD-OD-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-SD-SD-SD-SD-SD-SD-S-S-S-S-S-SD-SD-S-S-S-S-S-S-S-S-S-S-S-ID-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-