This work explores the impact of moderation on users' enjoyment of conversational AI systems. While recent advancements in Large Language Models (LLMs) have led to highly capable conversational AIs that are increasingly deployed in real-world settings, there is a growing concern over AI safety and the need to moderate systems to encourage safe language and prevent harm. However, some users argue that current approaches to moderation limit the technology, compromise free expression, and limit the value delivered by the technology. This study takes an unbiased stance and shows that moderation does not necessarily detract from user enjoyment. Heavy handed moderation does seem to have a nefarious effect, but models that are moderated to be safer can lead to a better user experience. By deploying various conversational AIs in the Chai platform, the study finds that user retention can increase with a level of moderation and safe system design. These results demonstrate the importance of appropriately defining safety in models in a way that is both responsible and focused on serving users.
翻译:本文探讨了系统审查对用户享受对话AI系统的影响。虽然LLM(巨型语言模型)的最近进展已经使得对话AI变得越来越有能力,并且在现实世界中被广泛部署,但越来越多的人对人工智能的安全性和需要对系统进行审查以鼓励安全语言并防止伤害的问题表示担忧。然而,一些用户认为当前的审查方法限制了技术的发展,侵犯了表达自由,并且限制了这项技术的价值。这项研究采取中立立场,表明审查并不一定会减少用户的享受。过度的审查似乎是有潜在风险的,但通过适当的审查和安全系统设计,模型可以实现更大的用户体验。通过在Chai平台上部署各种对话AI,研究发现,在进行一定程度的审查和安全设计的情况下,用户的留存率可以提高。这些结果展示了适当地定义在模型中的安全性是多么重要,它既负责又专注于服务用户。