Machine learning is the dominant approach to artificial intelligence, through which computers learn from data and experience. In the framework of supervised learning, for a computer to learn from data accurately and efficiently, some auxiliary information about the data distribution and target function should be provided to it through the learning model. This notion of auxiliary information relates to the concept of regularization in statistical learning theory. A common feature among real-world datasets is that data domains are multiscale and target functions are well-behaved and smooth. In this paper, we propose a learning model that exploits this multiscale data structure and discuss its statistical and computational benefits. The hierarchical learning model is inspired by the logical and progressive easy-to-hard learning mechanism of human beings and has interpretable levels. The model apportions computational resources according to the complexity of data instances and target functions. This property can have multiple benefits, including higher inference speed and computational savings in training a model for many users or when training is interrupted. We provide a statistical analysis of the learning mechanism using multiscale entropies and show that it can yield significantly stronger guarantees than uniform convergence bounds.
翻译:机器学习是人工智能的主要方法,计算机通过它从数据和经验中学习。在监督学习的框架内,为了使计算机能够准确和有效地从数据中学习,应当通过学习模式向计算机提供关于数据分布和目标功能的一些辅助信息。辅助信息的概念与统计学习理论的正规化概念有关。现实世界数据集的一个共同特点是数据领域是多尺度的,目标功能是妥善和顺畅的。在本文中,我们提出了一个学习模式,利用这一多尺度的数据结构并讨论其统计和计算效益。等级学习模式受人类逻辑和渐进的容易硬件学习机制的启发,具有可解释的水平。模型根据数据实例和目标功能的复杂程度分配计算资源。这一属性可以产生多种好处,包括许多用户培训模式或培训中断时的推论速度和计算节约。我们用多尺度的编织对学习机制进行统计分析,并表明它能够产生比统一汇合界限更强得多的保障。