We analyze phase transitions in the conditional entropy of a sequence caused by a change in the conditional variables. Such transitions happen, for example, when training to learn the parameters of a system, since the transition from the training phase to the data phase causes a discontinuous jump in the conditional entropy of the measured system response. For large-scale systems, we present a method of computing a bound on the mutual information obtained with one-shot training, and show that this bound can be calculated using the difference between two derivatives of a conditional entropy. The system model does not require Gaussianity or linearity in the parameters, and does not require worst-case noise approximations or explicit estimation of any unknown parameters. The model applies to a broad range of algorithms and methods in communication, signal processing, and machine learning that employ training as part of their operation.
翻译:我们分析由条件变量变化导致的序列的有条件变换的有条件变换。这种变换发生在学习系统参数的培训时,因为从培训阶段向数据阶段的过渡导致测量系统响应的有条件变换的不连续跳跃。对于大型系统,我们提出一种方法,在通过一次性培训获得的相互信息上进行约束计算,并表明这一结合可以使用有条件变变换的两种衍生物之间的差别来计算。系统模型不需要参数中的高西尼或线性,也不需要最坏的噪音近似值或任何未知参数的明确估计。该模型适用于通信、信号处理和机器学习方面的多种算法和方法,这些算法和方法将培训作为其操作的一部分。