We consider the problem of learning a neural network classifier. Under the information bottleneck (IB) principle, we associate with this classification problem a representation learning problem, which we call "IB learning". We show that IB learning is, in fact, equivalent to a special class of the quantization problem. The classical results in rate-distortion theory then suggest that IB learning can benefit from a "vector quantization" approach, namely, simultaneously learning the representations of multiple input objects. Such an approach assisted with some variational techniques, result in a novel learning framework, "Aggregated Learning", for classification with neural network models. In this framework, several objects are jointly classified by a single neural network. The effectiveness of this framework is verified through extensive experiments on standard image recognition and text classification tasks.
翻译:我们考虑的是学习神经网络分类的问题。根据信息瓶颈(IB)原则,我们把这个分类问题与代表学习问题联系起来,我们称之为“IB学习”。我们表明,IB学习实际上相当于量化问题的一个特殊类别。典型的扭曲率理论结果表明,IB学习可以受益于“Victor量化”方法,即同时学习多个输入对象的表达方式。这种方法借助一些变异技术,形成了与神经网络模型分类的新学习框架,即“聚合学习”。在这个框架内,几个对象由单一神经网络联合分类。通过对标准图像识别和文本分类任务的广泛实验,验证这一框架的有效性。