We present a comparison of several online machine learning techniques for tactical learning and proving in the Coq proof assistant. This work builds on top of Tactician, a plugin for Coq that learns from proofs written by the user to synthesize new proofs. This learning happens in an online manner -- meaning that Tactician's machine learning model is updated immediately every time the user performs a step in an interactive proof. This has important advantages compared to the more studied offline learning systems: (1) it provides the user with a seamless, interactive experience with Tactician and, (2) it takes advantage of locality of proof similarity, which means that proofs similar to the current proof are likely to be found close by. We implement two online methods, namely approximate $k$-nearest neighbors based on locality sensitive hashing forests and random decision forests. Additionally, we conduct experiments with gradient boosted trees in an offline setting using XGBoost. We compare the relative performance of Tactician using these three learning methods on Coq's standard library.
翻译:我们比较了用于战术学习的几种在线机器学习技术,并在 Coq 校对助理中进行验证。 这项工作以Tactician (Coq 的插件) 为基础, 以Tactician 为基础, 从用户撰写的证明中学习, 合成新的证明。 这种学习以在线方式进行, 这意味着Tactician 的机器学习模式在用户在互动证明中迈出一步时会立即更新。 与更多研究的离线学习系统相比,这具有重要的优势:(1) 它为用户提供了与Tactician 的无缝互动经验,(2) 它利用了相似的证明地点, 这就意味着与当前证明相近的证明很可能在附近找到。 我们实施了两种在线方法, 即基于地方敏感散雨林和随机决定森林的近邻。 此外, 我们用 XGBoost 进行离线环境中梯度增生树实验。 我们比较了Tactician 的相对表现, 使用Coq 标准图书馆的这三个学习方法。