Learning on sets is increasingly gaining attention in the machine learning community, due to its widespread applicability. Typically, representations over sets are computed by using fixed aggregation functions such as sum or maximum. However, recent results showed that universal function representation by sum- (or max-) decomposition requires either highly discontinuous (and thus poorly learnable) mappings, or a latent dimension equal to the maximum number of elements in the set. To mitigate this problem, we introduce a learnable aggregation function (LAF) for sets of arbitrary cardinality. LAF can approximate several extensively used aggregators (such as average, sum, maximum) as well as more complex functions (e.g., variance and skewness). We report experiments on semi-synthetic and real data showing that LAF outperforms state-of-the-art sum- (max-) decomposition architectures such as DeepSets and library-based architectures like Principal Neighborhood Aggregation, and can be effectively combined with attention-based architectures.
翻译:机组学习由于其广泛适用性,在机组学习中日益受到注意。通常,通过使用固定集合功能(如总和或最大)来计算各组的表示方式,但最近的结果显示,以总和(或最大)分解法表示的通用功能要求高度不连续(因而学习不力)绘图,或者要求具有与集集中最大元素数量相等的潜在维度。为了缓解这一问题,我们为各组任意的基点引入了可学习的集合功能(LAF ) 。LAF可以估计几个广泛使用的聚合物(如平均、总和、最大)以及更复杂的功能(如差异和偏差 ) 。我们报告关于半合成和真实数据的实验表明,LAF优于诸如DeepSet和以图书馆为基础的建筑(如Gent Neighborhood Agregation)等最先进的合成(最大)分解结构,并且可以有效地与以注意力为基础的建筑相结合。