The existing tensor networks adopt conventional matrix product for connection. The classical matrix product requires strict dimensionality consistency between factors, which can result in redundancy in data representation. In this paper, the semi-tensor product is used to generalize classical matrix product-based mode product to semi-tensor mode product. As it permits the connection of two factors with different dimensionality, more flexible and compact tensor decompositions can be obtained with smaller sizes of factors. Tucker decomposition, Tensor Train (TT) and Tensor Ring (TR) are common decomposition for low rank compression of deep neural networks. The semi-tensor product is applied to these tensor decompositions to obtained their generalized versions, i.e., semi-tensor Tucker decomposition (STTu), semi-tensor train(STT) and semi-tensor ring (STR). Experimental results show the STTu, STT and STR achieve higher compression factors than the conventional tensor decompositions with the same accuracy but less training times in ResNet and WideResNetcompression. With 2% accuracy degradation, the TT-RN (rank = 14) and the TR-WRN (rank = 16) only obtain 3 times and99t times compression factors while the STT-RN (rank = 14) and the STR-WRN (rank = 16) achieve 9 times and 179 times compression factors, respectively.
翻译:古典矩阵产品要求各要素之间严格的维维度一致性,这可能导致数据表达方式的冗余。在本文中,半超度产品用于将古典矩阵基于产品的模式产品推广到半蒸汽模式产品。由于它允许将两个因素与不同的维度连接,因此可以以较小大小的因素获得更灵活和紧凑的超强分解。塔克分解、Tensor tra(TTT)和Tensor Rring(TR)是深神经网络低级压缩的常见分解。半超度产品用于这些高压分解状态,以获得其普遍版本,即半超度或塔克分解(STT)、半强度列列(STT)和半分解(ST)。实验结果显示,STT、STT和STR(T)的压缩系数高于常规的调解压缩系数,在ResNet和广度内核网络网络网络网络的低级压缩(NR)中,半超度产品应用到这些高分解状态,即:精确度为2倍(TB-R)仅达到S-RS-rmax 和16次的递解(TB)的递解(S-r),只有2次。