We propose to employ the hierarchical coarse-grained structure in the artificial neural networks explicitly to improve the interpretability without degrading performance. The idea has been applied in two situations. One is a neural network called TaylorNet, which aims to approximate the general mapping from input data to output result in terms of Taylor series directly, without resorting to any magic nonlinear activations. The other is a new setup for data distillation, which can perform multi-level abstraction of the input dataset and generate new data that possesses the relevant features of the original dataset and can be used as references for classification. In both cases, the coarse-grained structure plays an important role in simplifying the network and improving both the interpretability and efficiency. The validity has been domonstrated on MNIST and CIFAR-10 datasets. Further improvement and some open questions related are also discussed.
翻译:我们提议在人工神经网络中采用粗粗粗的等级结构,明确改善可解释性,而不降低性能。这一想法适用于两种情况。一种是称为TaylorNet的神经网络,目的是直接将从输入数据到Taylor系列输出结果的总图绘制相近,而不必诉诸任何非线性神奇的激活。另一种是数据蒸馏的新结构,它可以对输入数据集进行多层次的抽取,产生具有原始数据集相关特征的新数据,并可用于作为分类参考。在这两种情况下,粗皮结构在简化网络和改善可解释性和效率方面都发挥着重要作用。对MNIST和CIFAR-10数据集进行了调整,还讨论了进一步改进和一些公开问题。