In this paper, we propose the Redundancy Reduction Twins Network (RRTN), a redundancy reduction training framework that minimizes redundancy by measuring the cross-correlation matrix between the outputs of the same network fed with distorted versions of a sample and bringing it as close to the identity matrix as possible. RRTN also applies a new loss function, the Barlow Twins loss function, to help maximize the similarity of representations obtained from different distorted versions of a sample. However, as the distribution of losses can cause performance fluctuations in the network, we also propose the use of a Restrained Uncertainty Weight Loss (RUWL) or joint training to identify the best weights for the loss function. Our best approach on CNN14 with the proposed methodology obtains a CCC over emotion regression of 0.678 on the ExVo Multi-task dev set, a 4.8% increase over a vanilla CNN 14 CCC of 0.647, which achieves a significant difference at the 95% confidence interval (2-tailed).
翻译:在本文中,我们建议采用裁员后双胞胎网络(裁员后双胞胎网络),这是一个裁员后培训框架,通过测量同一网络以扭曲的样本版本提供的产出之间的交叉关系矩阵,尽可能接近身份矩阵,从而最大限度地减少冗余冗余。 RRTN还运用了新的损失功能,即巴洛双胞胎损失功能,帮助尽量扩大不同扭曲样本的表述的相似性。然而,由于损失分配可能导致网络的性能波动,我们还建议使用弹性不确定不确定性能损失(RUWL)或联合培训,以确定损失函数的最佳重量。我们在CNN14上采用拟议方法的最佳方法,在ExVo多任务模式上获得了0.678的情感回归,比Vanilla CNN 14 CCC0.647增加了4.8%,这在95%的置信度间隔(2号)上有很大差异。