We consider the problem of data classification where the training set consists of just a few data points. We explore this phenomenon mathematically and reveal key relationships between the geometry of an AI model's feature space, the structure of the underlying data distributions, and the model's generalisation capabilities. The main thrust of our analysis is to reveal the influence on the model's generalisation capabilities of nonlinear feature transformations mapping the original data into high, and possibly infinite, dimensional spaces.
翻译:我们从数学角度来探讨这个现象,并揭示AI模型特征空间的几何、基础数据分布结构和模型的概括能力之间的关键关系。我们分析的主旨是揭示模型非线性特征转换对模型一般化能力的影响,将原始数据映射为高、甚至无限的宇宙空间。