The design of robust learning systems that offer stable performance under a wide range of supervision degrees is investigated in this work. We choose the image classification problem as an illustrative example and focus on the design of modularized systems that consist of three learning modules: representation learning, feature learning and decision learning. We discuss ways to adjust each module so that the design is robust with respect to different training sample numbers. Based on these ideas, we propose two families of learning systems. One adopts the classical histogram of oriented gradients (HOG) features while the other uses successive-subspace-learning (SSL) features. We test their performance against LeNet-5, which is an end-to-end optimized neural network, for MNIST and Fashion-MNIST datasets. The number of training samples per image class goes from the extremely weak supervision condition (i.e., 1 labeled sample per class) to the strong supervision condition (i.e., 4096 labeled sample per class) with gradual transition in between (i.e., $2^n$, $n=0, 1, \cdots, 12$). Experimental results show that the two families of modularized learning systems have more robust performance than LeNet-5. They both outperform LeNet-5 by a large margin for small $n$ and have performance comparable with that of LeNet-5 for large $n$.
翻译:在这项工作中,我们选择图像分类问题作为示例,并侧重于模块化系统的设计,该系统由三个学习模块组成:代表性学习、特征学习和决策学习;我们讨论如何调整每个模块,以使每个模块的设计与不同的培训抽样数字相符;根据这些想法,我们建议两个学习系统组。一个采用面向梯度特点的传统直观图,而另一个使用连续子空间学习特征。我们用LeNet-5测试其性能,这是一个端到端优化的神经网络,用于MNIST和Fashon-MNIST数据集。每个图像单元的培训样本数量从极弱的监督条件(即每类1个贴标签样本)到强的监管条件(即每类4096个标签样本)到在连续的子空间学习中逐步转换(即2美元,$=0,1美元/dots,12美元)。实验结果显示,Le-5模型组的大型性能表现比Le-5模型大,两个模型都比Le-5模型的软性能更强。