Decisions such as which movie to watch next, which song to listen to, or which product to buy online, are increasingly influenced by recommender systems and user models that incorporate information on users' past behaviours, preferences, and digitally created content. Machine learning models that enable recommendations and that are trained on user data may unintentionally leverage information on human characteristics that are considered vulnerabilities, such as depression, young age, or gambling addiction. The use of algorithmic decisions based on latent vulnerable state representations could be considered manipulative and could have a deteriorating impact on the condition of vulnerable individuals. In this paper, we are concerned with the problem of machine learning models inadvertently modelling vulnerabilities, and want to raise awareness for this issue to be considered in legislation and AI ethics. Hence, we define and describe common vulnerabilities, and illustrate cases where they are likely to play a role in algorithmic decision-making. We propose a set of requirements for methods to detect the potential for vulnerability modelling, detect whether vulnerable groups are treated differently by a model, and detect whether a model has created an internal representation of vulnerability. We conclude that explainable artificial intelligence methods may be necessary for detecting vulnerability exploitation by machine learning-based recommendation systems.
翻译:在本文中,我们关注机器学习模型无意间造成脆弱性的问题,并想提高对这一问题的认识,以便在立法和AI道德中加以考虑。 因此,我们界定和描述共同的脆弱性,并举例说明它们有可能在算法决策中发挥作用的案例。 我们提出一套要求,以各种方法查明脆弱性建模的潜力,查明脆弱群体是否受到模型的不同待遇,并查明模型是否造成了脆弱性的内部代表性。