In a real-world setting, object instances from new classes can be continuously encountered by object detectors. When existing object detectors are applied to such scenarios, their performance on old classes deteriorates significantly. A few efforts have been reported to address this limitation, all of which apply variants of knowledge distillation to avoid catastrophic forgetting. We note that although distillation helps to retain previous learning, it obstructs fast adaptability to new tasks, which is a critical requirement for incremental learning. In this pursuit, we propose a meta-learning approach that learns to reshape model gradients, such that information across incremental tasks is optimally shared. This ensures a seamless information transfer via a meta-learned gradient preconditioning that minimizes forgetting and maximizes knowledge transfer. In comparison to existing meta-learning methods, our approach is task-agnostic, allows incremental addition of new-classes and scales to high-capacity models for object detection. We evaluate our approach on a variety of incremental learning settings defined on PASCAL-VOC and MS COCO datasets, where our approach performs favourably well against state-of-the-art methods.
翻译:在现实世界环境中,物体探测器可以不断发现新类别中的物体实例。当现有物体探测器应用到这些假设情况时,其老类的性能会大大恶化。据报告,为解决这一局限性,已作出一些努力,所有这些努力都应用了知识蒸馏的变异,以避免灾难性的遗忘。我们注意到,尽管蒸馏有助于保留以前的学习,但它阻碍着对新任务的快速适应,而新任务是渐进学习的关键要求。在进行这项工作时,我们建议采用一种元学习方法,学会重塑模型梯度,以便最佳地分享跨增量任务的信息。这确保了通过元学梯度先决条件无缝地传递信息,最大限度地减少遗忘和最大限度地实现知识转让。与现有的元学习方法相比,我们的方法是任务性,允许逐步增加新级和尺度,成为目标探测的高能力模型。我们评估了我们在PACAL-VOC和MS COCO数据集中界定的各种增量学习环境的方法,我们的方法与最新方法相比表现良好。