We introduce the notion of weak convexity in metric spaces, a generalization of ordinary convexity commonly used in machine learning. It is shown that weakly convex sets can be characterized by a closure operator and have a unique decomposition into a set of pairwise disjoint connected blocks. We give two generic efficient algorithms, an extensional and an intensional one for learning weakly convex concepts and study their formal properties. Our experimental results concerning vertex classification clearly demonstrate the excellent predictive performance of the extensional algorithm. Two non-trivial applications of the intensional algorithm to polynomial PAC-learnability are presented. The first one deals with learning $k$-convex Boolean functions, which are already known to be efficiently PAC-learnable. It is shown how to derive this positive result in a fairly easy way by the generic intensional algorithm. The second one is concerned with the Euclidean space equipped with the Manhattan distance. For this metric space, weakly convex sets are a union of pairwise disjoint axis-aligned hyperrectangles. We show that a weakly convex set that is consistent with a set of examples and contains a minimum number of hyperrectangles can be found in polynomial time. In contrast, this problem is known to be NP-complete if the hyperrectangles may be overlapping.
翻译:我们引入了测量空间的弱共性概念, 这是一种在机器学习中常用的普通共共性的一般共性的一般概念。 事实证明, 弱共性组可以由关闭操作员来描述, 并具有独特的分解功能, 形成一组双向互不连接的块块。 我们给出了两种通用有效算法, 一种扩展法和强化法, 用于学习弱共性概念, 并研究其正式属性。 我们关于垂直分类的实验结果, 清楚地表明扩展算法的预测性表现良好。 提供了两种非三维的机器学习中常见共性算法的非三端应用。 演示了两个超高端的加固算算法对多式 PAC- learnable 。 第一个是学习 $k$k$- convex Boolean 函数, 这些函数已经众所周知是高效的 PAC- learnable 。 我们展示了如何以较易交错的极性轴式模型, 能够显示一个固定的比对齐性点。 我们发现一个固定的极性轴式的比重的比重模型。