A common privacy issue in traditional machine learning is that data needs to be disclosed for the training procedures. In situations with highly sensitive data such as healthcare records, accessing this information is challenging and often prohibited. Luckily, privacy-preserving technologies have been developed to overcome this hurdle by distributing the computation of the training and ensuring the data privacy to their owners. The distribution of the computation to multiple participating entities introduces new privacy complications and risks. In this paper, we present a privacy-preserving decentralised workflow that facilitates trusted federated learning among participants. Our proof-of-concept defines a trust framework instantiated using decentralised identity technologies being developed under Hyperledger projects Aries/Indy/Ursa. Only entities in possession of Verifiable Credentials issued from the appropriate authorities are able to establish secure, authenticated communication channels authorised to participate in a federated learning workflow related to mental health data.
翻译:传统机器学习中常见的隐私问题是,需要为培训程序披露数据;在保健记录等高度敏感数据的情况下,获取这种信息具有挑战性,而且往往受到禁止;幸运的是,已经开发了保护隐私的技术,通过将培训的计算方法进行分配,并确保其所有者享有数据隐私,以克服这一障碍;将计算方法分发给多个参与实体,带来了新的隐私问题和风险;在本文件中,我们提出了一个保护隐私的分散工作流程,为参与者之间可信赖的联合会式学习提供便利;我们的概念证明界定了一个利用超光速项目Aries/Indy/Ursa开发的分散身份技术即时利用的分散身份技术的信托框架;只有拥有主管当局颁发的可验证证书的实体才能建立安全、经认证的通信渠道,授权参与与心理健康数据有关的联合学习工作流程。