Recently, the study of graph neural network (GNN) has attracted much attention and achieved promising performance in molecular property prediction. Most GNNs for molecular property prediction are proposed based on the idea of learning the representations for the nodes by aggregating the information of their neighbor nodes (e.g. atoms). Then, the representations can be passed to subsequent layers to deal with individual downstream tasks. Therefore, the architectures of GNNs can be considered as being composed of two core parts: graph-related layers and task-specific layers. Facing real-world molecular problems, the hyperparameter optimization for those layers are vital. Hyperparameter optimization (HPO) becomes expensive in this situation because evaluating candidate solutions requires massive computational resources to train and validate models. Furthermore, a larger search space often makes the HPO problems more challenging. In this research, we focus on the impact of selecting two types of GNN hyperparameters, those belonging to graph-related layers and those of task-specific layers, on the performance of GNN for molecular property prediction. In our experiments. we employed a state-of-the-art evolutionary algorithm (i.e., CMA-ES) for HPO. The results reveal that optimizing the two types of hyperparameters separately can gain the improvements on GNNs' performance, but optimising both types of hyperparameters simultaneously will lead to predominant improvements. Meanwhile, our study also further confirms the importance of HPO for GNNs in molecular property prediction problems.
翻译:最近,对图形神经网络(GNN)的研究在分子属性预测方面吸引了许多关注,并取得了有希望的绩效。大多数用于分子属性预测的GNN(GNN)是根据通过汇总其相邻节点(如原子)的信息来学习节点的表示方式的想法提出的。然后,这些表示方式可以传递到随后的层次,以处理个别下游任务。因此,GNN的结构可以被视为由两个核心部分组成:与图形有关的层次和任务特定的层次。面对现实世界分子问题,这些层次的超参数优化至关重要。超参数优化在此情况下变得昂贵,因为评估候选解决方案需要大量的计算资源来培训和验证模型。此外,更大的搜索空间往往使HPO问题更具挑战性。在这项研究中,我们侧重于选择两种GNNN超参数,即与图形有关的层次和任务特定层次的架构,对GNNNM的绩效的绩效表现至关重要。在实验中,超参数的超参数优化的状态分析将使得GPO(GPO) 两种水平的升级结果(GPO) 的升级结果将同时用于HMA的双重的升级。