Let a labeled dataset be given with scattered samples and consider the hypothesis of the ground-truth belonging to the reproducing kernel Hilbert space (RKHS) of a known positive-definite kernel. It is known that out-of-sample bounds can be established at unseen input locations, thus limiting the risk associated with learning this function. We show how computing tight, finite-sample uncertainty bounds amounts to solving parametric quadratically constrained linear programs. In our setting, the outputs are assumed to be contaminated by bounded measurement noise that can otherwise originate from any compactly supported distribution. No independence assumptions are made on the available data. Numerical experiments are presented to compare the present results with other closed-form alternatives.
翻译:使用分散的样本提供贴有标签的数据集,并考虑Hilbert空间(RKHS)复制核心空间(RKHS)的地面真实性假设,即已知的正定内核。已知可以在未知输入地点建立外标界限,从而限制与学习此功能相关的风险。我们展示了计算紧凑、有限抽样的不确定性界限如何相当于解决参数四边限制线性程序。在我们的设置中,产出假定受到来自任何压缩支持分布的封闭测量噪音的污染。对现有数据没有独立假设。我们用数字实验将现有结果与其他封闭式的替代数据进行比较。