For graph learning tasks, many existing methods utilize a message-passing mechanism where vertex features are updated iteratively by aggregation of neighbor information. This strategy provides an efficient means for graph features extraction, but obtained features after many iterations might contain too much information from other vertices, and tend to be similar to each other. This makes their representations less expressive. Learning graphs using paths, on the other hand, can be less adversely affected by this problem because it does not involve all vertex neighbors. However, most of them can only compare paths with the same length, which might engender information loss. To resolve this difficulty, we propose a new Graph Kernel based on a Longest Common Subsequence (LCS) similarity. Moreover, we found that the widely-used R-convolution framework is unsuitable for path-based Graph Kernel because a huge number of comparisons between dissimilar paths might deteriorate graph distances calculation. Therefore, we propose a novel metric space by exploiting the proposed LCS-based similarity, and compute a new Wasserstein-based graph distance in this metric space, which emphasizes more the comparison between similar paths. Furthermore, to reduce the computational cost, we propose an adjacent point merging operation to sparsify point clouds in the metric space.
翻译:对于图形学习任务,许多现有方法都使用一个信息传递机制,让顶点特征通过相邻信息汇总来迭接更新。这一战略为图形特征提取提供了有效的手段,但在许多迭代后获得的特征可能包含来自其他脊椎的信息过多,而且往往彼此相似。这使得其表达方式不那么直观。另一方面,使用路径的学习图形可能受到这一问题的影响较小,因为它不涉及所有顶点邻居。然而,其中多数方法只能将路径与同一长度比较,从而可能造成信息损失。为了解决这一困难,我们提议在“长期共同后序”(LCS)相似之处的基础上建立一个新的“内核图 ” 。 此外,我们发现,广泛使用的R- 革命框架不适合基于路径的“ 图形内核”, 因为不同路径之间的大量比较可能会降低图形距离的计算。 因此,我们提出一个新的计量空间, 办法是利用基于 LCS 的相近点的路径, 并计算出一个新的基于瓦瑟斯坦 图形的距离, 从而更加强调在相邻的轨道上进行对比。