Machine learning models are often personalized based on information that is protected, sensitive, self-reported, or costly to acquire. These models use information about people, but do not facilitate nor inform their \emph{consent}. Individuals cannot opt out of reporting information that a model needs to personalize their predictions, nor tell if they would benefit from personalization in the first place. In this work, we introduce a new family of prediction models, called \emph{participatory systems}, that allow individuals to opt into personalization at prediction time. We present a model-agnostic algorithm to learn participatory systems for supervised learning tasks where models are personalized with categorical group attributes. We conduct a comprehensive empirical study of participatory systems in clinical prediction tasks, comparing them to common approaches for personalization and imputation. Our results demonstrate that participatory systems can facilitate and inform consent in a way that improves performance and privacy across all groups who report personal data.
翻译:机器学习模式往往基于受保护、敏感、自我报告或昂贵的信息而个性化。这些模式使用关于人的信息,但却不为其提供方便或信息。个人不能选择不报告模型需要使其预测个人化的信息,也不能首先说明他们是否会从个性化中受益。在这项工作中,我们引入了一个新的预测模型系列,称为\emph{参与性系统},允许个人在预测时选择个性化。我们提出了一个模型-不可知性算法,以学习参与系统,在监督的学习任务中,模型具有绝对群体属性。我们对临床预测任务中的参与性系统进行了全面的经验研究,将其与个人化和估算的共同方法进行比较。我们的结果表明,参与系统可以促进和告知同意,从而改善所有报告个人数据的群体的业绩和隐私。