Generative Flow Networks (GFlowNets) have demonstrated significant performance improvements for generating diverse discrete objects $x$ given a reward function $R(x)$, indicating the utility of the object and trained independently from the GFlowNet by supervised learning to predict a desirable property $y$ given $x$. We hypothesize that this can lead to incompatibility between the inductive optimization biases in training $R$ and in training the GFlowNet, potentially leading to worse samples and slow adaptation to changes in the distribution. In this work, we build upon recent work on jointly learning energy-based models with GFlowNets and extend it to learn the joint over multiple variables, which we call Joint Energy-Based GFlowNets (JEBGFNs), such as peptide sequences and their antimicrobial activity. Joint learning of the energy-based model, used as a reward for the GFlowNet, can resolve the issues of incompatibility since both the reward function $R$ and the GFlowNet sampler are trained jointly. We find that this joint training or joint energy-based formulation leads to significant improvements in generating anti-microbial peptides. As the training sequences arose out of evolutionary or artificial selection for high antibiotic activity, there is presumably some structure in the distribution of sequences that reveals information about the antibiotic activity. This results in an advantage to modeling their joint generatively vs. pure discriminative modeling. We also evaluate JEBGFN in an active learning setting for discovering anti-microbial peptides.
翻译:源源网络(GFlowNets)在产生不同离散对象方面表现显著改善, 以美元xx美元作为奖励功能, 显示该对象的效用, 并且通过监督学习, 预测一个理想的属性, 以美元x美元为单位对GFlowNet进行训练, 从而与GFlowNet独立培训。 我们假设,这可能导致在培训美元和培训GFlowNet时的暗示优化偏差不相容, 可能导致更差的样本和对分配变化的适应缓慢。 在这项工作中, 我们以最近的工作为基础, 与GFlowNets联合学习基于能源的模型, 以美元xxx美元为单位, 表明该对象的用途的效用和训练与GFlowNet公司独立, 表明该对象在多个变量上是联合学习, 我们称之为联合基于能源的虚拟网络(JEBGFW)网络(JEBGN)网络(JEBGGN)的效用, 我们称之为联合培训或联合的主动性变量, 也导致在BREB 选择结构中, 联合进行一项活动, 指向BEBEBEB的升级的升级活动, 显示, 高的排序。