Modelling functions of sets, or equivalently, permutation-invariant functions, is a long-standing challenge in machine learning. Deep Sets is a popular method which is known to be a universal approximator for continuous set functions. We provide a theoretical analysis of Deep Sets which shows that this universal approximation property is only guaranteed if the model's latent space is sufficiently high-dimensional. If the latent space is even one dimension lower than necessary, there exist piecewise-affine functions for which Deep Sets performs no better than a na\"ive constant baseline, as judged by worst-case error. Deep Sets may be viewed as the most efficient incarnation of the Janossy pooling paradigm. We identify this paradigm as encompassing most currently popular set-learning methods. Based on this connection, we discuss the implications of our results for set learning more broadly, and identify some open questions on the universality of Janossy pooling in general.
翻译:数据集或等效的变异函数的模型功能是机器学习中长期存在的挑战。深数据集是一种流行的方法,已知是连续设定函数的通用近似符。我们对深数据集进行理论分析,表明只有模型的潜层空间足够高的高度才能保证这种通用近似属性。如果潜伏空间甚至比必要的一个维度低一点,那么深数据集在机器学习中不会比“恒定基线”更好。深套功能可以被视为Janossy集合模式中最高效的化身。我们把这一模式确定为包括目前最受欢迎的成套学习方法。基于这一联系,我们讨论我们的结果对更广泛学习的影响,并找出关于Janossy集合普遍性的一些未决问题。