Zero-shot Learners are models capable of predicting unseen classes. In this work, we propose a Zero-shot Learning approach for text categorization. Our method involves training model on a large corpus of sentences to learn the relationship between a sentence and embedding of sentence's tags. Learning such relationship makes the model generalize to unseen sentences, tags, and even new datasets provided they can be put into same embedding space. The model learns to predict whether a given sentence is related to a tag or not; unlike other classifiers that learn to classify the sentence as one of the possible classes. We propose three different neural networks for the task and report their accuracy on the test set of the dataset used for training them as well as two other standard datasets for which no retraining was done. We show that our models generalize well across new unseen classes in both cases. Although the models do not achieve the accuracy level of the state of the art supervised models, yet it evidently is a step forward towards general intelligence in natural language processing.
翻译:零点学习者是能够预测隐蔽课程的模型。 在这项工作中, 我们提出一个零点学习方法, 用于文本分类。 我们的方法是用大量句子的培训模型来学习句子和嵌入句子标签之间的关系。 学习这种关系使模型概括为看不见的句子、 标记, 甚至新的数据集, 只要可以将其放入相同的嵌入空间。 模型学会预测某一句子是否与标签有关; 不同于学习将句子归类为可能课类之一的其他分类者。 我们建议三个不同的神经网络来进行任务分类, 并在用于培训的数据集测试集上报告其准确性, 以及另外两个没有再培训的标准数据集。 我们显示, 我们的模式在两种情况下都广泛分布在新的隐蔽类中。 虽然模型没有达到艺术监管模式的准确度, 但显然向自然语言处理中的一般情报迈进了一步。