AI is undergoing a paradigm shift with the rise of models (e.g., BERT, DALL-E, GPT-3) that are trained on broad data at scale and are adaptable to a wide range of downstream tasks. We call these models foundation models to underscore their critically central yet incomplete character. This report provides a thorough account of the opportunities and risks of foundation models, ranging from their capabilities (e.g., language, vision, robotics, reasoning, human interaction) and technical principles(e.g., model architectures, training procedures, data, systems, security, evaluation, theory) to their applications (e.g., law, healthcare, education) and societal impact (e.g., inequity, misuse, economic and environmental impact, legal and ethical considerations). Though foundation models are based on standard deep learning and transfer learning, their scale results in new emergent capabilities,and their effectiveness across so many tasks incentivizes homogenization. Homogenization provides powerful leverage but demands caution, as the defects of the foundation model are inherited by all the adapted models downstream. Despite the impending widespread deployment of foundation models, we currently lack a clear understanding of how they work, when they fail, and what they are even capable of due to their emergent properties. To tackle these questions, we believe much of the critical research on foundation models will require deep interdisciplinary collaboration commensurate with their fundamentally sociotechnical nature.
翻译:大赦国际正在经历一种范式的转变,其模式(如BERT、DALL-E、GPT-3)的兴起,这些模型经过了广泛的大规模数据培训,适应了广泛的下游任务。我们将这些模型的基础模型称为基础模型,以强调其关键的核心但不完整的性质。本报告透彻地描述了基础模型的机会和风险,这些模型有其能力(如语言、愿景、机器人、推理、人际互动)和技术原则(如模型结构、培训程序、数据、系统、安全、评价、理论),适用于其应用(如法律、保健、教育)和社会影响(如不平等、滥用、经济和环境影响、法律和道德因素)。虽然基础模型基于标准的深层次学习和转移学习,其规模导致新的出现能力,以及其在许多任务中的有效性都鼓励同质化。调和基因化提供了强大的杠杆作用,但需要谨慎,因为基础模型的缺陷已被所有调整模式的下游继承。尽管基础模型即将广泛部署,但我们目前缺乏一个深刻的深刻的基础模型,我们无法理解这些研究基础基础的基础,我们需要如何理解这些研究的深刻基础。我们如何理解这些基础基础基础,我们如何理解这些基础基础基础基础基础基础基础基础,我们需要深入地理解这些基础基础基础基础基础基础基础基础基础基础基础基础,我们需要在如何发展。我们如何深入地理解。