Chatbots are mainly data-driven and usually based on utterances that might be sensitive. However, training deep learning models on shared data can violate user privacy. Such issues have commonly existed in chatbots since their inception. In the literature, there have been many approaches to deal with privacy, such as differential privacy and secure multi-party computation, but most of them need to have access to users' data. In this context, Federated Learning (FL) aims to protect data privacy through distributed learning methods that keep the data in its location. This paper presents Fedbot, a proof-of-concept (POC) privacy-preserving chatbot that leverages large-scale customer support data. The POC combines Deep Bidirectional Transformer models and federated learning algorithms to protect customer data privacy during collaborative model training. The results of the proof-of-concept showcase the potential for privacy-preserving chatbots to transform the customer support industry by delivering personalized and efficient customer service that meets data privacy regulations and legal requirements. Furthermore, the system is specifically designed to improve its performance and accuracy over time by leveraging its ability to learn from previous interactions.
翻译:聊天机器人主要基于敏感话语的数据驱动,然而,联合数据训练深度学习模型可能会侵犯用户隐私。这种问题自聊天机器人诞生以来就一直存在于文献中。在文献中,有许多方法来处理隐私问题,例如差分隐私和安全多方计算,但大多数方法需要访问用户的数据。在这种情况下,联邦学习旨在通过分布式学习方法保护数据隐私,使数据保持在其位置。本文介绍了Fedbot——一个概念验证(POC)隐私保护聊天机器人,利用大规模客户支持数据。POC结合深度双向变换器模型和联邦学习算法,在协作模型训练期间保护客户数据隐私。概念验证的结果展示了隐私保护聊天机器人的潜力,它能通过提供符合数据隐私法规和法律要求的个性化高效客户服务,改变客户支持行业。此外,该系统专门设计,通过利用其从先前互动中学习的能力,不断提高其性能和准确性。