MLOps is about taking experimental ML models to production, i.e., serving the models to actual users. Unfortunately, existing ML serving systems do not adequately handle the dynamic environments in which online data diverges from offline training data, resulting in tedious model updating and deployment works. This paper implements a lightweight MLOps plugin, termed ModelCI-e (continuous integration and evolution), to address the issue. Specifically, it embraces continual learning (CL) and ML deployment techniques, providing end-to-end supports for model updating and validation without serving engine customization. ModelCI-e includes 1) a model factory that allows CL researchers to prototype and benchmark CL models with ease, 2) a CL backend to automate and orchestrate the model updating efficiently, and 3) a web interface for an ML team to manage CL service collaboratively. Our preliminary results demonstrate the usability of ModelCI-e, and indicate that eliminating the interference between model updating and inference workloads is crucial for higher system efficiency.
翻译:不幸的是,现有的ML服务系统无法充分处理在线数据与离线培训数据不一致的动态环境,导致模式更新和部署工作乏味。本文使用一个轻巧的MLOPs插件,称为ModelCI-e(持续整合和演进),以解决这一问题。具体地说,它包含持续学习(CL)和ML部署技术,为模型更新和验证提供端对端支持,而不为引擎定制。模型CI-e包括:1)一个模型工厂,使CL研究人员能够轻松地原型和基准CL模型;2) CL后端到自动更新并高效调整模型更新;3)一个网络界面,供ML团队协作管理CL服务。我们的初步结果显示模型-e的可用性,并表明消除模型更新和推断工作量之间的干扰对于提高系统效率至关重要。