Inherent in virtually every iterative machine learning algorithm is the problem of hyper-parameter tuning, which includes three major design parameters: (a) the complexity of the model, e.g., the number of neurons in a neural network, (b) the initial conditions, which heavily affect the behavior of the algorithm, and (c) the dissimilarity measure used to quantify its performance. We introduce an online prototype-based learning algorithm that can be viewed as a progressively growing competitive-learning neural network architecture for classification and clustering. The learning rule of the proposed approach is formulated as an online gradient-free stochastic approximation algorithm that solves a sequence of appropriately defined optimization problems, simulating an annealing process. The annealing nature of the algorithm contributes to avoiding poor local minima, offers robustness with respect to the initial conditions, and provides a means to progressively increase the complexity of the learning model, through an intuitive bifurcation phenomenon. The proposed approach is interpretable, requires minimal hyper-parameter tuning, and allows online control over the performance-complexity trade-off. Finally, we show that Bregman divergences appear naturally as a family of dissimilarity measures that play a central role in both the performance and the computational complexity of the learning algorithm.
翻译:在几乎所有迭代机学习算法中,几乎每个迭代机的内在都是超参数调制问题,其中包括三大设计参数:(a) 模型的复杂性,例如神经网络中的神经神经元数量;(b) 初始条件,这些条件严重影响了算法的行为,以及(c) 用于量化其性能的不相似性衡量标准。我们引入了在线原型学习算法,这种原型学习算法可被视为一种不断增长的竞争性学习神经网络分类和集群结构。拟议方法的学习规则是设计成一种在线的无梯度梯度随机近似算法,它解决了定义得当的优化问题序列,模拟了厌食过程。算法的隐性性质有助于避免当地微弱的微弱行为,提供了对初始条件的稳健性,并提供了一种手段,通过直觉的两极分化现象逐渐增加学习模式的复杂性。拟议方法可以解释,需要最低限度的超度调制,并允许对性能兼容性交易进行在线控制。最后,我们展示的是,算法的反变法的功能是自然地演算法。