Throughout application domains, we now rely extensively on algorithmic systems to engage with ever-expanding datasets of information. Despite their benefits, these systems are often complex (comprising of many intricate tools, e.g., moderation, recommender systems, prediction models), of unknown structure (due to the lack of accompanying documentation), and having hard-to-predict yet potentially severe downstream consequences (due to the extensive use, systematic enactment of existing errors, and many comprising feedback loops). As such, understanding and evaluating these systems as a whole remains a challenge for both researchers and legislators. To aid ongoing efforts, we introduce a formal framework for such visibility allocation systems (VASs) which we define as (semi-)automated systems deciding which (processed) data to present a human user with. We review typical tools comprising VASs and define the associated computational problems they solve. By doing so, VASs can be decomposed into sub-processes and illustrated via data flow diagrams. Moreover, we survey metrics for evaluating VASs throughout the pipeline, thus aiding system diagnostics. Using forecasting-based recommendations in school choice as a case study, we demonstrate how our framework can support VAS evaluation. We also discuss how our framework can support ongoing AI-legislative efforts to locate obligations, quantify systemic risks, and enable adaptive compliance.
翻译:在各个应用领域中,我们现在广泛依赖算法系统来处理日益扩大的信息数据集。尽管这些系统带来了诸多益处,但它们通常具有复杂性(包含许多精细工具,如内容审核、推荐系统、预测模型)、结构不透明(由于缺乏配套文档),且可能产生难以预测但潜在严重的下游后果(源于广泛使用、既有错误的系统性实施以及众多反馈循环的构成)。因此,对研究者与立法者而言,整体理解和评估这些系统仍是一项挑战。为助力当前研究,我们为此类可见性分配系统引入了一个形式化框架,并将其定义为(半)自动化系统,其核心功能是决定向人类用户呈现哪些(经处理的)数据。我们梳理了构成VAS的典型工具,并定义了它们所解决的相关计算问题。通过这一方式,VAS可被分解为多个子过程,并通过数据流图加以阐释。此外,我们系统综述了贯穿整个流程的VAS评估指标,从而助力系统诊断。以择校场景中基于预测的推荐系统为案例研究,我们展示了该框架如何支持VAS评估。最后,我们探讨了该框架如何协助当前的人工智能立法工作,包括定位责任义务、量化系统性风险以及实现适应性合规。