We target open-world feature extrapolation problem where the feature space of input data goes through expansion and a model trained on partially observed features needs to handle new features in test data without further retraining. The problem is of much significance for dealing with features incrementally collected from different fields. To this end, we propose a new learning paradigm with graph representation and learning. Our framework contains two modules: 1) a backbone network (e.g., feedforward neural nets) as a lower model takes features as input and outputs predicted labels; 2) a graph neural network as an upper model learns to extrapolate embeddings for new features via message passing over a feature-data graph built from observed data. Based on our framework, we design two training strategies, a self-supervised approach and an inductive learning approach, to endow the model with extrapolation ability and alleviate feature-level over-fitting. We also provide theoretical analysis on the generalization error on test data with new features, which dissects the impact of training features and algorithms on generalization performance. Our experiments over several classification datasets and large-scale advertisement click prediction datasets demonstrate that our model can produce effective embeddings for unseen features and significantly outperforms baseline methods that adopt KNN and local aggregation.
翻译:我们针对的是开放世界特征外推问题,输入数据的特点空间是通过扩展而来的,对部分观察到的特征进行训练的模型需要处理测试数据中的新特征,而无需进一步再培训。这个问题对于处理从不同领域逐渐收集的特征非常重要。为此,我们提出一个新的学习范式,配有图形表达和学习。我们的框架包含两个模块:1)主干网络(例如饲料向前神经网),作为较低模型,具有作为投入和产出预测标签的特征;2)一个图形神经网络,作为顶级模型,学会通过传递来自观测数据的信息,将新特征外嵌入外嵌。我们根据我们的框架,设计了两种培训战略、一种自我监督的方法和一种感性学习方法,用外推法将模型置于外推能力之下,并减轻特征水平上的过度。我们还对带有新特征的测试数据的一般错误进行了理论分析,这种错误分解了培训特征和算法对一般化绩效的影响。我们通过一些分类数据集和大型广告根据从观察到的数据图表的图表进行实验,大大地点击了模型的预测模型,从而得出了本地的模型和模型,从而得出了可靠的模型。