An important problem with machine learning is that when label number n>2, it is very difficult to construct and optimize a group of learning functions, and we wish that optimized learning functions are still useful when prior distribution P(x) (where x is an instance) is changed. To resolve this problem, the semantic information G theory, Logical Bayesian Inference (LBI), and a group of Channel Matching (CM) algorithms together form a systematic solution. A semantic channel in the G theory consists of a group of truth functions or membership functions. In comparison with likelihood functions, Bayesian posteriors, and Logistic functions used by popular methods, membership functions can be more conveniently used as learning functions without the above problem. In Logical Bayesian Inference (LBI), every label's learning is independent. For Multilabel learning, we can directly obtain a group of optimized membership functions from a big enough sample with labels, without preparing different samples for different labels. A group of Channel Matching (CM) algorithms is developed for machine learning. For the Maximum Mutual Information (MMI) classification of three classes with Gaussian distributions on a two-dimensional feature space, 2-3 iterations can make mutual information between three classes and three labels surpass 99% of the MMI for most initial partitions. For mixture models, the Expectation-Maximization (EM) algorithm is improved and becomes the CM-EM algorithm, which can outperform the EM algorithm when mixture ratios are imbalanced, or local convergence exists. The CM iteration algorithm needs to combine neural networks for MMI classifications on high-dimensional feature spaces. LBI needs further studies for the unification of statistics and logic.
翻译:机器学习的一个重要问题是,当标签编号 n>2 时,很难构建和优化一组学习功能,因此我们希望,当先前的分布 P (x) (在 x 是一例的情况下) 改变时,优化学习功能仍然有用。要解决这个问题,语义信息 G 理论、 逻辑贝氏推断(LBI) 以及一组频道匹配逻辑(CM) 算法将形成一个系统性的解决方案。 G 理论中的语义频道由一组真理趋同功能或会籍函数组成。与流行方法使用的可能性函数、 巴耶西亚的箭头和物流函数相比,最优化的学习功能仍然有用。在Logical Bayesian Inference (LBIBI) 中,每个标签的学习都是独立的。多标签学习中,我们可以直接从一个足够大的样本中获取一组最优化的会籍函数,而不必为不同的标签准备不同的样本。 频道匹配(CMCM) 的算法可以用来进行机器学习。对于三个类的最小相互信息化信息(MMI) 和MIM 3 的初始分析,可以将三个类的MEM 的MLlevormal 。