Deep learning has had a significant impact on many fields. Recently, code-to-code neural models have been used in code translation, code refinement and decompilation. However, the question of whether these models can automate compilation has yet to be investigated. In this work, we explore neural compilation, building and evaluating Transformer models that learn how to produce x86 assembler from C code. Although preliminary results are relatively weak, we make our data, models and code publicly available to encourage further research in this area.
翻译:深层学习对许多领域产生了重大影响。 最近,代码对代码神经模型被用于代码翻译、代码完善和解密。 但是,这些模型能否自动汇编的问题还有待调查。 在这项工作中,我们探索神经汇编、构建和评估能够学习如何从 C 代码中生成 x86 集成器的变换器模型。尽管初步结果相对薄弱,但我们公开提供我们的数据、模型和代码,以鼓励在这一领域进行进一步的研究。