Recent advances in neural-based generative modeling have reignited the hopes of having computer systems capable of conversing with humans and able to understand natural language. The employment of deep neural architectures has been largely explored in a multitude of context and tasks to fulfill various user needs. On one hand, producing textual content that meets specific requirements is of priority for a model to seamlessly conduct conversations with different groups of people. On the other hand, latent variable models (LVM) such as variational auto-encoders (VAEs) as one of the most popular genres of generative models are designed to characterize the distributional pattern of textual data. Thus they are inherently capable of learning the integral textual features that are worth exploring for controllable pursuits. \noindent This overview gives an introduction to existing generation schemes, problems associated with text variational auto-encoders, and a review of several applications about the controllable generation that are instantiations of these general formulations,\footnote{A detailed paper list is available at \url{https://github.com/ImKeTT/CTG-latentAEs}} as well as related datasets, metrics and discussions for future researches. Hopefully, this overview will provide an overview of living questions, popular methodologies and raw thoughts for controllable language generation under the scope of variational auto-encoder.
翻译:最近基于神经的基因模型的进展重新燃起了使计算机系统能够与人类交流并能理解自然语言的希望。深层神经结构的使用在多种背景和任务中广泛探索,以满足各种用户的需要。一方面,产生符合具体要求的文本内容是同不同人群进行无缝对话的模式的优先。另一方面,潜伏变量模型(LVM),如变式自动编码器(VAE),作为最受欢迎的基因模型类型之一,旨在描述文本数据的分布模式。因此,深层神经结构的运用在很多方面得到了广泛的探讨,因此它们具有内在能力学习值得探索的综合性文本特征,以适应各种可控的追求。本概览介绍了现有一代计划、与文本变异自动编码器有关的问题,并审查了关于可控的一代的若干应用程序,这些通用配方的可即可调控的版本,\fofootnote{https://github.com/ILATT/G-LOVATS) 详细文件列表,作为可操作的动态分析方法,提供了当前数据分析的版本,作为相关的数据分析范围。