With the development of artificial intelligence, dialogue systems have been endowed with amazing chit-chat capabilities, and there is widespread interest and discussion about whether the generated contents are socially beneficial. In this paper, we present a new perspective of research scope towards building a safe, responsible, and modal dialogue system, including 1) abusive and toxic contents, 2) unfairness and discrimination, 3) ethics and morality issues, and 4) risk of misleading and privacy information. Besides, we review the mainstream methods for evaluating the safety of large models from the perspectives of exposure and detection of safety issues. The recent advances in methodologies for the safety improvement of both end-to-end dialogue systems and pipeline-based models are further introduced. Finally, we discussed six existing challenges towards responsible AI: explainable safety monitoring, continuous learning of safety issues, robustness against malicious attacks, multimodal information processing, unified research framework, and multidisciplinary theory integration. We hope this survey will inspire further research toward safer dialogue systems.
翻译:随着人工智能的发展,对话系统具有惊人的聊天能力,人们广泛关注和讨论所产生的内容是否对社会有益。在本文件中,我们提出了一个新的研究范围,以建立一个安全、负责和模式的对话系统,包括:(1) 滥用和有毒的内容,(2) 不公平和歧视,(3) 伦理和道德问题,(4) 误导和隐私信息的风险。此外,我们还从接触和发现安全问题的角度,审查评估大型模型安全性的主流方法。还进一步介绍了在改进终端至终端对话系统和管道模式的安全性方法方面的最新进展。最后,我们讨论了对负责任的AI的六项现有挑战:可解释的安全监测、持续学习安全问题、抵御恶意攻击的稳健性、多式信息处理、统一研究框架和多学科理论整合。我们希望这一调查将鼓励进一步研究更安全的对话系统。</s>