We describe cases where real recommender systems were modified in the service of various human values such as diversity, fairness, well-being, time well spent, and factual accuracy. From this we identify the current practice of values engineering: the creation of classifiers from human-created data with value-based labels. This has worked in practice for a variety of issues, but problems are addressed one at a time, and users and other stakeholders have seldom been involved. Instead, we look to AI alignment work for approaches that could learn complex values directly from stakeholders, and identify four major directions: useful measures of alignment, participatory design and operation, interactive value learning, and informed deliberative judgments.
翻译:我们描述的是,在为多样性、公平性、福祉、美好时间和事实准确性等各种人类价值观服务方面,真正的建议系统被修改的案例。我们从中确定当前价值观工程的做法:利用基于价值的标签从人造数据中创建分类器。这在实践中对各种问题发挥了作用,但问题一个一个一个地得到解决,用户和其他利益攸关方很少参与。相反,我们期待AI调整工作能够直接从利益攸关方那里学习复杂的价值观,并确定四个主要方向:调整、参与性设计和操作、互动价值学习和知情的审议判断等有用措施。