A recent paper (Neural Networks, {\bf 132} (2020), 253-268) introduces a straightforward and simple kernel based approximation for manifold learning that does not require the knowledge of anything about the manifold, except for its dimension. In this paper, we examine the pointwise error in approximation using least squares optimization based on this kernel, in particular, how the error depends upon the data characteristics and deteriorates as one goes away from the training data. The theory is presented with an abstract localized kernel, which can utilize any prior knowledge about the data being located on an unknown sub-manifold of a known manifold. We demonstrate the performance of our approach using a publicly available micro-Doppler data set investigating the use of different pre-processing measures, kernels, and manifold dimension. Specifically, it is shown that the Gaussian kernel introduced in the above mentioned paper leads to a near-competitive performance to deep neural networks, and offers significant improvements in speed and memory requirements. Similarly, a kernel based on treating the feature space as a submanifold of the Grassman manifold outperforms conventional hand-crafted features. To demonstrate the fact that our methods are agnostic to the domain knowledge, we examine the classification problem in a simple video data set.
翻译:最近的一份论文(Neural Networks, {bf 132}(2020), 253-268) 提出了一个简单、简单、以内核为基础的多元学习近似,除了其尺寸外,不需要了解任何关于元体的知识。 在本文中,我们研究了使用基于此内核的最小正方形优化的近似中点错误,特别是错误如何取决于数据特性和随着与培训数据脱节而恶化。该理论以一个抽象的局部内核来展示,它能够利用关于数据位于已知元体的未知次层的任何先前知识。我们用一个公开的微多普勒数据集来展示我们的方法的性能,该数据集调查不同预处理措施、内核和多维度的使用情况。具体地说,上述文件中引入的高斯内核内核内核导致深度神经网络的接近竞争性性能,并大大改进速度和记忆要求。 同样,一个基于将地域空间特征作为格拉斯曼元外部下层外部的下层部分。 我们用一个常规域域知识来研究一个事实。