The success of machine learning stems from its structured data representation. Similar data have close representation as compressed codes for classification or emerged labels for clustering. We observe that the frequency of the internal representation follows power laws in both supervised and unsupervised learning. The scale-invariant distribution implies that machine learning largely compresses frequent typical data, and at the same time, differentiates many atypical data as outliers. In this study, we derive how the power laws can naturally arise in machine learning. In terms of information theory, the scale-invariant representation corresponds to a maximally uncertain data grouping among possible representations that guarantee pre-specified learning accuracy.
翻译:机器学习的成功源于其结构化的数据代表性。类似数据作为压缩分类代码或集群标签的出现,具有近似代表性。我们观察到内部代表性的频率遵循受监督和不受监督的学习中的权力法。比例差异分布意味着机器学习在很大程度上压缩了常见的典型数据,同时将许多非典型数据作为外端数据加以区分。在本研究中,我们得出了权力法如何在机器学习中自然产生。在信息理论中,规模差异性代表与保证事先确定学习准确性的各种可能表述之间的最不确定的数据组合相对应。