In this paper, online linear regression in environments corrupted by non-Gaussian noise (especially heavy-tailed noise) is addressed. In such environments, the error between the system output and the label also does not follow a Gaussian distribution and there might exist abnormally large error samples (or outliers) which mislead the learning process. The main challenge is how to keep the supervised learning problem least affected by these unwanted and misleading outliers. In recent years, an information theoretic algorithm based on Renyi's entropy, called minimum error entropy (MEE), has been employed to take on this issue. However, this minimization might not result in a desired estimator inasmuch as entropy is shift-invariant, i.e., by minimizing the error entropy, error samples may not be necessarily concentrated around zero. In this paper, a quantization technique is proposed by which not only aforementioned need of setting errors around the origin in MEE is addressed, but also major outliers are rejected from MEE-based learning and MEE performance is improved from convergence rate, steady state misalignment, and testing error points of view.
翻译:在本文中,解决了非加西噪音(特别是重尾噪声)腐蚀环境中的在线线性回归问题。在这种环境中,系统输出和标签之间的错误也并不随高斯分布而变化,而且可能存在异常大的错误样本(或离子),误导学习过程。主要的挑战是如何使受监督的学习问题受到这些不想要的和误导的离子的影响最小。近年来,采用了基于Renyi的酶(称为最小误差激(MEE))的信息理论算法来处理这一问题。然而,这种最小化可能不会导致预期的测算器,因为恒温是变异的,也就是说,通过最小化误差样本不一定集中在零点左右。在本文中,提出了一种昆虫化技术,不仅解决了上述需要确定MEE来源的错误,而且还拒绝了MEE学习中的主要离子,MEE的性能从趋同率、稳定状态误差、测试和误差点上改进了MEE的趋同率。