Long-tail learning is the problem of learning under skewed label distributions, which pose a challenge for standard learners. Several recent approaches for the problem have proposed enforcing a suitable margin in logit space. Such techniques are intuitive analogues of the guiding principle behind SVMs, and are equally applicable to linear models and neural models. However, when applied to neural models, such techniques do not explicitly control the geometry of the learned embeddings. This can be potentially sub-optimal, since embeddings for tail classes may be diffuse, resulting in poor generalization for these classes. We present Embedding and Logit Margins (ELM), a unified approach to enforce margins in logit space, and regularize the distribution of embeddings. This connects losses for long-tail learning to proposals in the literature on metric embedding, and contrastive learning. We theoretically show that minimising the proposed ELM objective helps reduce the generalisation gap. The ELM method is shown to perform well empirically, and results in tighter tail class embeddings.
翻译:长尾学习是使用扭曲标签分布的学习问题,这对标准学习者构成挑战。最近针对这一问题的几种方法都提议在登录空间中实施适当的边距。这些技术是SVM 背后指导原则的直观类比,同样适用于线性模型和神经模型。但是,当应用到神经模型时,这些技术并不明确控制所学嵌入内容的几何。这可能是次优的,因为对尾巴类的嵌入可能分散,导致这些类的概括性差。我们展示了嵌入和Logit Margins(ELM),这是在登录空间中执行边际的一致方法,并规范嵌入的分布。这把长尾学习损失与关于指标嵌入和对比学习的文献中的建议联系起来。我们从理论上表明,将拟议的ELM目标最小化有助于缩小总体差距。ELM方法在经验上表现良好,并导致更紧密的尾层嵌入。