We analyze phase transitions in the conditional entropy of a sequence caused by a change in the conditional variables or input distribution. Such transitions happen, for example, when training to learn the parameters of a system, since the switch from training input to data input causes a discontinuous jump in the conditional entropy of the measured system response. We show that the size of the discontinuity is of particular interest in the computation of mutual information during data transmission, and that this mutual information can be calculated as the difference between two derivatives of a single function. For large-scale systems we present a method of computing the mutual information for a system model with one-shot learning that does not require Gaussianity or linearity in the model, and does not require worst-case noise approximations or explicit estimation of any unknown parameters. The model applies to a broad range of algorithms and methods in communication, signal processing, and machine learning that employ training as part of their operation.
翻译:我们分析由有条件变量或输入分布变化导致的序列的有条件变异或输入分布的有条件变异。 这种变异发生在学习系统参数的培训中,因为从培训输入到数据输入的转换导致测量系统响应的有条件变异的不连续跳跃。 我们表明,在计算数据传输过程中的相互信息时,不连续的大小特别有意义,这种相互信息可以计算为单一函数的两个衍生物之间的差别。 对于大型系统,我们提出了一个为系统模型计算相互信息的方法,其中一手学习不需要模型中的高斯性或线性,也不需要最坏的噪音近似值或任何未知参数的明确估计。 该模型适用于通信、信号处理和机器学习方面的多种算法和方法,这些算作其操作的一部分。