This work develops problem statements related to encoders and autoencoders with the goal of elucidating variational formulations and establishing clear connections to information-theoretic concepts. Specifically, four problems with varying levels of input are considered : a) The data, likelihood and prior distributions are given, b) The data and likelihood are given; c) The data and prior are given; d) the data and the dimensionality of the parameters is specified. The first two problems seek encoders (or the posterior) and the latter two seek autoencoders (i.e. the posterior and the likelihood). A variational Bayesian setting is pursued, and detailed derivations are provided for the resulting optimization problem. Following this, a linear Gaussian setting is adopted, and closed form solutions are derived. Numerical experiments are also performed to verify expected behavior and assess convergence properties. Explicit connections are made to rate-distortion theory, information bottleneck theory, and the related concept of sufficiency of statistics is also explored. One of the motivations of this work is to present the theory and learning dynamics associated with variational inference and autoencoders, and to expose information theoretic concepts from a computational science perspective.
翻译:这项工作开发了与编码器和自动编码器有关的问题说明,目的是澄清变异配方,并明确与信息理论概念的联系;具体地说,考虑了投入程度不同的四个问题:(a) 提供了数据、可能性和先前的分布情况;(b) 提供了数据和可能性;(c) 提供了数据和可能性;(c) 提供了数据和先前的数据;(d) 具体说明了参数的数据和维度;前两个问题寻求变异配方(或后方),后两个问题寻求自动编码器(即后方和可能性); 追求变异贝叶斯设置,并为由此产生的优化问题提供详细的衍生结果; 之后,采用了线性高斯设置,并提出了封闭式解决办法; 也进行了数值实验,以核实预期的行为并评估趋同特性; 与率扭曲理论、信息瓶颈理论和相关的统计充足性概念进行了外向联系; 也探讨了这项工作的动机之一,是提出理论和学习与变异学相关的动力,从理论和汽车学角度来了解与变异学相关的动态。