Adversarial training provides a means of regularizing supervised learning algorithms while virtual adversarial training is able to extend supervised learning algorithms to the semi-supervised setting. However, both methods require making small perturbations to numerous entries of the input vector, which is inappropriate for sparse high-dimensional inputs such as one-hot word representations. We extend adversarial and virtual adversarial training to the text domain by applying perturbations to the word embeddings in a recurrent neural network rather than to the original input itself. The proposed method achieves state of the art results on multiple benchmark semi-supervised and purely supervised tasks. We provide visualizations and analysis showing that the learned word embeddings have improved in quality and that while training, the model is less prone to overfitting. Code is available at https://github.com/tensorflow/models/tree/master/research/adversarial_text.
翻译:双向培训提供了一种使受监督的学习算法正规化的手段,而虚拟对抗性培训则能够将受监督的学习算法扩大到半受监督的环境,然而,这两种方法都需要对输入矢量的许多条目进行小扰动,这对诸如一热单词表达式等稀疏的高维输入不合适。我们通过对嵌入经常性神经网络而非原始输入本身的词进行扰动,将对抗性和虚拟对抗性培训扩大到文本领域。拟议方法在多个基准半监督和纯受监督的任务中取得了最新的结果。我们提供了可视化和分析,表明学到的词嵌入质量有所提高,而且虽然培训过程不易过度适应。代码可在https://github.com/tensorflow/models/tree/master/research/research/abtyt_text查阅。