Learning binary representations of instances and classes is a classical problem with several high potential applications. In modern settings, the compression of high-dimensional neural representations to low-dimensional binary codes is a challenging task and often require large bit-codes to be accurate. In this work, we propose a novel method for Learning Low-dimensional binary Codes (LLC) for instances as well as classes. Our method does not require any side-information, like annotated attributes or label meta-data, and learns extremely low-dimensional binary codes (~20 bits for ImageNet-1K). The learnt codes are super-efficient while still ensuring nearly optimal classification accuracy for ResNet50 on ImageNet-1K. We demonstrate that the learnt codes capture intrinsically important features in the data, by discovering an intuitive taxonomy over classes. We further quantitatively measure the quality of our codes by applying it to the efficient image retrieval as well as out-of-distribution (OOD) detection problems. For ImageNet-100 retrieval problem, our learnt binary codes outperform 16 bit HashNet using only 10 bits and also are as accurate as 10 dimensional real representations. Finally, our learnt binary codes can perform OOD detection, out-of-the-box, as accurately as a baseline that needs ~3000 samples to tune its threshold, while we require none. Code is open-sourced at https://github.com/RAIVNLab/LLC.
翻译:在现代环境中,将高度神经神经表示压缩到低维二进码的低维二进码中是一项艰巨的任务,往往需要大量的位码才能准确。在这项工作中,我们建议了一种创新方法,用于学习低维二进码(LLC),用于实例和类别。我们的方法不需要任何侧信息,例如附加说明的属性或标签的元数据,并学习极低维的二进码(图像Net-1K ) 。在现代环境中,将高维神经神经显示压缩到低维二进码中是一个典型的问题。在现代环境中,将高维神经显示到低维的神经神经在低维码中是一个挑战性的任务,同时仍然确保低维的低维神经神经神经显示低维的低维码,同时仍确保图像Net50 的ResNet50 在图像Net-1K 上,将高维神经神经神经代码包含数据中固有的重要特征。 最后,我们进一步量化测量我们的代码的质量,将它应用到高效的开放图像检索检索和外部分配(OOOD)的检测,我們只用10BLIM 的標頭的標碼,我們可以用16BRBRBRB 的標關需要。