Learning binary representations of instances and classes is a classical problem with several high potential applications. In modern settings, the compression of high-dimensional neural representations to low-dimensional binary codes is a challenging task and often require large bit-codes to be accurate. In this work, we propose a novel method for Learning Low-dimensional binary Codes (LLC) for instances as well as classes. Our method does not require any side-information, like annotated attributes or label meta-data, and learns extremely low-dimensional binary codes (~20 bits for ImageNet-1K). The learnt codes are super-efficient while still ensuring nearly optimal classification accuracy for ResNet50 on ImageNet-1K. We demonstrate that the learnt codes capture intrinsically important features in the data, by discovering an intuitive taxonomy over classes. We further quantitatively measure the quality of our codes by applying it to the efficient image retrieval as well as out-of-distribution (OOD) detection problems. For ImageNet-100 retrieval problem, our learnt binary codes outperform 16 bit HashNet using only 10 bits and also are as accurate as 10 dimensional real representations. Finally, our learnt binary codes can perform OOD detection, out-of-the-box, as accurately as a baseline that needs ~3000 samples to tune its threshold, while we require none. Code and pre-trained models are available at https://github.com/RAIVNLab/LLC.
翻译:在现代环境中,将高维神经表示器压缩为低维二进制代码是一项艰巨的任务,往往需要大量的位码才能准确。在这项工作中,我们提出了一种创新的方法,用于学习低维二进制代码(LLC),用于实例和类别。我们的方法不需要任何侧端信息,如附加说明的属性或标签元数据,并学习极低维的二进制代码(图像Net-1K 的 ~20 比特)。在现代环境中,将高维神经表示器压缩为低维二进制代码是一项挑战性的任务,而且往往需要大量的位代码才能准确无误地分类。我们通过在课堂上发现一个直观的二进制代码(LL),进一步量化地测量我们的代码的质量,将它应用到高效的图像检索以及调控前(OOODD)的检测问题。对于图像Net-100检索问题,我们学习的二进制代码超越了16比哈斯网,仅使用10比特,并且也作为10比特的精确的分类,我们用OD-40的样本进行精确的检测。最后,我们可以将代码作为OD-bis-bas-bas-bormormex 需要。