Hierarchical learning algorithms that gradually approximate a solution to a data-driven optimization problem are essential to decision-making systems, especially under limitations on time and computational resources. In this study, we introduce a general-purpose hierarchical learning architecture that is based on the progressive partitioning of a possibly multi-resolution data space. The optimal partition is gradually approximated by solving a sequence of optimization sub-problems that yield a sequence of partitions with increasing number of subsets. We show that the solution of each optimization problem can be estimated online using gradient-free stochastic approximation updates. As a consequence, a function approximation problem can be defined within each subset of the partition and solved using the theory of two-timescale stochastic approximation algorithms. This simulates an annealing process and defines a robust and interpretable heuristic method to gradually increase the complexity of the learning architecture in a task-agnostic manner, giving emphasis to regions of the data space that are considered more important according to a predefined criterion. Finally, by imposing a tree structure in the progression of the partitions, we provide a means to incorporate potential multi-resolution structure of the data space into this approach, significantly reducing its complexity, while introducing hierarchical variable-rate feature extraction properties similar to certain classes of deep learning architectures. Asymptotic convergence analysis and experimental results are provided for supervised and unsupervised learning problems.
翻译:分层学习算法可逐步逼近基于数据的优化问题的解决方案,对于决策制定系统特别重要,尤其是在时间和计算资源限制下。本研究介绍了一种基于可能具有多分辨率空间的分层和渐进划分的通用分层学习架构。最优划分通过解决一系列优化子问题逐步逼近,得到一系列数量递增的子集划分。我们证明可以使用无梯度随机逼近更新在线估计每个优化问题的解,从而定义了一个任务不可知的,强调被认为更重要的数据空间区域的鲁棒且可解释的启发式方法,逐渐增加学习架构的复杂性。最后,通过在划分的进程中引入树形结构,我们提供了一种将数据空间的潜在多分辨率结构纳入该方法的手段,显著减少其复杂性,同时引入与某些类别的深度学习架构相似的分层可变速率特征提取属性。针对监督和无监督学习问题提供了渐进收敛分析和实验结果。