Many machine learning techniques incorporate identity-preserving transformations into their models to generalize their performance to previously unseen data. These transformations are typically selected from a set of functions that are known to maintain the identity of an input when applied (e.g., rotation, translation, flipping, and scaling). However, there are many natural variations that cannot be labeled for supervision or defined through examination of the data. As suggested by the manifold hypothesis, many of these natural variations live on or near a low-dimensional, nonlinear manifold. Several techniques represent manifold variations through a set of learned Lie group operators that define directions of motion on the manifold. However theses approaches are limited because they require transformation labels when training their models and they lack a method for determining which regions of the manifold are appropriate for applying each specific operator. We address these limitations by introducing a learning strategy that does not require transformation labels and developing a method that learns the local regions where each operator is likely to be used while preserving the identity of inputs. Experiments on MNIST and Fashion MNIST highlight our model's ability to learn identity-preserving transformations on multi-class datasets. Additionally, we train on CelebA to showcase our model's ability to learn semantically meaningful transformations on complex datasets in an unsupervised manner.
翻译:许多机器学习技术将身份保留转换纳入模型,以便将其性能推广到先前的不为人知的数据中。这些转换通常从一系列已知的功能中选择,这些功能在应用时(例如轮用、翻译、翻转和缩放)保持输入的特性。然而,许多自然变异不能在监督上贴标签,或通过对数据的审查来界定。如多重假设所示,许多自然变异都存在于或接近于一个低维、非线性多元的本地变异中。一些技术通过一组确定多元体运动方向的有学识的Lie组操作员代表多种变异。但是这些变异是有限的,因为它们在培训模型时需要变异标签以保持输入的特性特性,而它们缺乏一种方法来确定每个特定操作员都适合应用该输入的区域。我们通过引入一个不需要变异标签的学习战略来解决这些局限性,并开发一种方法来学习每个操作员在保存投入特性的同时可能使用的本地区域。关于MNIST和Fashon MNIST的实验突出了我们的模型在多级数据变异中学习身份模型在多级系统上进行有意义的数据展示的能力。