The recent advancements in machine learning (ML) have demonstrated the potential for providing a powerful solution to build complex prediction systems in a short time. However, in highly regulated industries, such as the financial technology (Fintech), people have raised concerns about the risk of ML systems discriminating against specific protected groups or individuals. To address these concerns, researchers have introduced various mathematical fairness metrics and bias mitigation algorithms. This paper discusses hidden technical debts and challenges of building fair ML systems in a production environment for Fintech. We explore various stages that require attention for fairness in the ML system development and deployment life cycle. To identify hidden technical debts that exist in building fair ML system for Fintech, we focus on key pipeline stages including data preparation, model development, system monitoring and integration in production. Our analysis shows that enforcing fairness for production-ready ML systems in Fintech requires specific engineering commitments at different stages of ML system life cycle. We also propose several initial starting points to mitigate these technical debts for deploying fair ML systems in production.
翻译:近来在机器学习方面的进步表明,有可能为短期内建立复杂的预测系统提供强有力的解决办法,然而,在金融技术(Fintech)等高度管制的行业中,人们提出了对歧视特定受保护群体或个人的ML系统的风险的关切;为解决这些关切,研究人员采用了各种数学公平度指标和减少偏见的算法;本文件讨论了在Fintech生产环境中建立公平的ML系统的隐性技术债务和挑战;我们探讨了在ML系统开发和部署生命周期中需要注意的公平性的各个阶段;为了查明在为Fintech建立公平的ML系统方面存在的隐性技术债务,我们把重点放在关键管道阶段,包括数据编制、模型开发、系统监测和生产一体化。我们的分析表明,在Fintech实施适合生产的ML系统,需要在ML系统生命周期的不同阶段执行具体的工程承诺。我们还提出几个初步起点,以减轻这些技术债务,以便在生产中部署公平的ML系统。