As the use of AI-powered applications widens across multiple domains, so do increase the computational demands. Primary driver of AI technology are the deep neural networks (DNNs). When focusing either on cloud-based systems that serve multiple AI queries from different users each with their own DNN model, or on mobile robots and smartphones employing pipelines of various models or parallel DNNs for the concurrent processing of multi-modal data, the next generation of AI systems will have multi-DNN workloads at their core. Large-scale deployment of AI services and integration across mobile and embedded systems require additional breakthroughs in the computer architecture front, with processors that can maintain high performance as the number of DNNs increases while meeting the quality-of-service requirements, giving rise to the topic of multi-DNN accelerator design.
翻译:随着AI动力应用的使用在多个领域的扩大,也增加了计算需求。AI型技术的主要驱动力是深神经网络。当侧重于为不同用户以自己的DNN模式提供多个AI查询服务的云基系统时,或者侧重于使用各种模型的管道或平行的DNN的移动机器人和智能手机以同时处理多模式数据时,下一代AI系统的核心工作将是多DNN工作量。大规模部署AI服务和在移动和嵌入系统中的整合需要计算机结构方面的更多突破,随着DNN数量增加,处理器能够保持高性能,同时满足服务质量要求,从而产生了多DNNNC加速器设计的主题。