We propose a few-shot learning method for unsupervised feature selection, which is a task to select a subset of relevant features in unlabeled data. Existing methods usually require many instances for feature selection. However, sufficient instances are often unavailable in practice. The proposed method can select a subset of relevant features in a target task given a few unlabeled target instances by training with unlabeled instances in multiple source tasks. Our model consists of a feature selector and decoder. The feature selector outputs a subset of relevant features taking a few unlabeled instances as input such that the decoder can reconstruct the original features of unseen instances from the selected ones. The feature selector uses the Concrete random variables to select features via gradient descent. To encode task-specific properties from a few unlabeled instances to the model, the Concrete random variables and decoder are modeled using permutation-invariant neural networks that take a few unlabeled instances as input. Our model is trained by minimizing the expected test reconstruction error given a few unlabeled instances that is calculated with datasets in source tasks. We experimentally demonstrate that the proposed method outperforms existing feature selection methods.
翻译:我们为未受监督的特性选择建议了一个微小的学习方法, 这是在未加标签的数据中选择一组相关特性的任务。 现有方法通常需要许多特性选择实例。 但是, 实际中往往没有足够的例子。 拟议的方法可以在目标任务中选择一组相关特性, 给一些未加标签的目标实例提供一些未加标签的多个源任务。 我们的模型包含一个特性选择器和解码器。 特性选择器输出一个相关特性的子集, 将几个未加标签的例子作为输入, 例如, 解码器可以从选定的例子中重建未见实例的原始特征。 特性选择器使用 CED 随机变量来通过渐渐脱落来选择特性。 要将任务特定特性从几个未加标签的事例编码到模型, 具体随机变量和解码器将使用以少数未加标签的事例作为输入的变异体网络来建模。 我们的模型通过最大限度地减少预期的测试重建错误, 以少数未加标签的例子来计算出源任务中的数据设置。 我们实验性地证明拟议的方法超越了现有特性选择方法。