Due to the ever-increasing threat of cyber-attacks to critical cyber infrastructure, organizations are focusing on building their cybersecurity knowledge base. A salient list of cybersecurity knowledge is the Common Vulnerabilities and Exposures (CVE) list, which details vulnerabilities found in a wide range of software and hardware. However, these vulnerabilities often do not have a mitigation strategy to prevent an attacker from exploiting them. A well-known cybersecurity risk management framework, MITRE ATT&CK, offers mitigation techniques for many malicious tactics. Despite the tremendous benefits that both CVEs and the ATT&CK framework can provide for key cybersecurity stakeholders (e.g., analysts, educators, and managers), the two entities are currently separate. We propose a model, named the CVE Transformer (CVET), to label CVEs with one of ten MITRE ATT&CK tactics. The CVET model contains a fine-tuning and self-knowledge distillation design applied to the state-of-the-art pre-trained language model RoBERTa. Empirical results on a gold-standard dataset suggest that our proposed novelties can increase model performance in F1-score. The results of this research can allow cybersecurity stakeholders to add preliminary MITRE ATT&CK information to their collected CVEs.
翻译:由于网络攻击对关键网络基础设施的威胁日益增大,各组织正在集中精力建立网络安全知识库,网络安全知识的显著清单是共同脆弱性和暴露(CVE)清单,其中详细列出了在各种软件和硬件中发现的弱点。然而,这些弱点往往没有防止攻击者利用这些弱点的缓解战略。众所周知的网络安全风险管理框架MITRE ATT&CK为许多恶意策略提供了缓解技术。尽管CVes和ATT&CK框架可为关键的网络安全利益攸关方(例如分析师、教育工作者和管理人员)提供巨大好处,但这两个实体目前是分开的。我们提议了一个名为CVE变换器(CVET)的模型,给CVE(CVE)贴上10MTRET&C策略之一的标签。CVET模型包含一个微调和自学蒸馏设计,适用于最先进的预先培训语言模型RoBERTA。在黄金标准数据集上取得的经验表明,我们提议的CVERC数据库初步研究成果可以增加FREC数据库收集的成绩。