Large-scale pre-training has recently emerged as a technique for creating capable, general purpose, generative models such as GPT-3, Megatron-Turing NLG, Gopher, and many others. In this paper, we highlight a counterintuitive property of such models and discuss the policy implications of this property. Namely, these generative models have an unusual combination of predictable loss on a broad training distribution (as embodied in their "scaling laws"), and unpredictable specific capabilities, inputs, and outputs. We believe that the high-level predictability and appearance of useful capabilities drives rapid development of such models, while the unpredictable qualities make it difficult to anticipate the consequences of model deployment. We go through examples of how this combination can lead to socially harmful behavior with examples from the literature and real world observations, and we also perform two novel experiments to illustrate our point about harms from unpredictability. Furthermore, we analyze how these conflicting properties combine to give model developers various motivations for deploying these models, and challenges that can hinder deployment. We conclude with a list of possible interventions the AI community may take to increase the chance of these models having a beneficial impact. We intend this paper to be useful to policymakers who want to understand and regulate AI systems, technologists who care about the potential policy impact of their work, and academics who want to analyze, critique, and potentially develop large generative models.
翻译:最近出现了大规模培训前的大规模培训,作为创造有能力的、通用的、通用的、实用能力外观的模型的技术,例如GPT-3、Megatron-Trining NLG、Gopher和其他许多模型。在本文件中,我们强调这类模型的反直觉特性,并讨论这种财产的政策影响。也就是说,这些基因模型在广泛的培训分布(体现在其“规模化法”中)和不可预测的具体能力、投入和产出方面有着不同寻常的、可预见的损失组合。我们认为,高层次的可预测性和显示的有用能力推动这些模型的迅速发展,而不可预测的品质使得难以预测模型部署的后果。我们通过文献和现实世界观察的范例来举例说明这种组合如何导致对社会有害的行为。我们还进行两项新的实验,以说明我们关于不可预测性的伤害。此外,我们分析这些相互冲突的特性如何结合给模型开发者以各种动机,以及可能阻碍部署的挑战。我们最后指出,AI界可能采取的干预清单,以增加这些模型产生有益影响的机会。我们打算用这一文件来说明这种组合如何引领悟、分析其潜在的决策者和研究。我们想要分析他们的能力。