We develop biologically plausible training mechanisms for self-supervised learning (SSL) in deep networks. SSL, with a contrastive loss, is more natural as it does not require labelled data and its robustness to perturbations yields more adaptable embeddings. Moreover the perturbation of data required to create positive pairs for SSL is easily produced in a natural environment by observing objects in motion and with variable lighting over time. We propose a contrastive hinge based loss whose error involves simple local computations as opposed to the standard contrastive losses employed in the literature, which do not lend themselves easily to implementation in a network architecture due to complex computations involving ratios and inner products. Furthermore we show that learning can be performed with one of two more plausible alternatives to backpropagation. The first is difference target propagation (DTP), which trains network parameters using target-based local losses and employs a Hebbian learning rule, thus overcoming the biologically implausible symmetric weight problem in backpropagation. The second is simply layer-wise learning, where each layer is directly connected to a layer computing the loss error. The layers are either updated sequentially in a greedy fashion (GLL) or in random order (RLL), and each training stage involves a single hidden layer network. The one step backpropagation needed for each such network can either be altered with fixed random feedback weights as proposed in Lillicrap et al. (2016), or using updated random feedback as in Amit (2019). Both methods represent alternatives to the symmetric weight issue of backpropagation. By training convolutional neural networks (CNNs) with SSL and DTP, GLL or RLL, we find that our proposed framework achieves comparable performance to its implausible counterparts in both linear evaluation and transfer learning tasks.
翻译:我们为深层网络的自我监督学习(SSL)开发了在生物上看似可信的培训机制(SSL ) 。 SSL, 具有对比性损失, 比较自然, 因为它不需要贴贴标签的数据, 而它对于扰动的坚固性可以产生更适应性化的嵌入。 此外, 创建 sLS 的正配所需的数据在自然环境中很容易被干扰, 方法是在运动中观测物体, 并随着时间的变化, 使用不同光线的光线规则。 我们提出一个对比性基点损失, 其错误涉及简单的本地计算, 而不是文献中使用的标准对比性反馈。 由于涉及比率和内产产品的复杂计算, 它不容易在网络中实施网络结构。 此外, 我们显示, 学习可以通过两种更合理的替代方法来进行。 不同的目标传播( DTP) 利用基于目标的本地损失来培训参数来培训网络参数, 从而克服反正向反正调的生理不光度重量问题。 第二, 简单的是层次学习, 每个层直接连接到计算CN 损耗损耗损率的网络, 在每一阶段里程中, 需要不断更新 Sralal- 的网络 。 。 水平 需要的运行的运行的运行的运行的运行的运行的运行的 。 需要的运行的网络的运行的运行的 。 。 在每平级的运行的运行的运行的运行的运行的运行的运行的运行的运行的运行的运行的运行的系统 。 在每级 需要的运行的网络 。