Probing is a popular method to discern what linguistic information is contained in the representations of pre-trained language models. However, the mechanism of selecting the probe model has recently been subject to intense debate, as it is not clear if the probes are merely extracting information or modeling the linguistic property themselves. To address this challenge, this paper introduces a novel model-free approach to probing, by formulating probing as a prompting task. We conduct experiments on five probing tasks and show that our approach is comparable or better at extracting information than diagnostic probes while learning much less on its own. We further combine the probing via prompting approach with attention head pruning to analyze where the model stores the linguistic information in its architecture. We then examine the usefulness of a specific linguistic property for pre-training by removing the heads that are essential to that property and evaluating the resulting model's performance on language modeling.
翻译:试探是一种常见的方法,用以辨别在经过训练的语文模型中包含的语言信息。然而,最近对选择探测模型的机制进行了激烈的辩论,因为不清楚探探是否仅仅是提取信息或模拟语言属性本身。为了应对这一挑战,本文提出了一种新的无示范的探究方法,将探究方法作为一种快速任务来拟订。我们在五项勘测任务上进行实验,并表明我们的方法在提取信息方面比诊断探测器要相似或更好,而自己学习得少得多。我们进一步将探索方法与注意力头划线结合起来,分析模型在其结构中储存语言信息的地点。然后我们研究特定语言属性在培训前的用处,方法是将对于该属性至关重要的负责人除名,并评估由此得出的模型在语言模型方面的表现。