When we fine-tune a well-trained deep learning model for a new set of classes, the network learns new concepts but gradually forgets the knowledge of old training. In some real-life applications, we may be interested in learning new classes without forgetting the capability of previous experience. Such learning without forgetting problem is often investigated using 2D image recognition tasks. In this paper, considering the growth of depth camera technology, we address the same problem for the 3D point cloud object data. This problem becomes more challenging in the 3D domain than 2D because of the unavailability of large datasets and powerful pretrained backbone models. We investigate knowledge distillation techniques on 3D data to reduce catastrophic forgetting of the previous training. Moreover, we improve the distillation process by using semantic word vectors of object classes. We observe that exploring the interrelation of old and new knowledge during training helps to learn new concepts without forgetting old ones. Experimenting on three 3D point cloud recognition backbones (PointNet, DGCNN, and PointConv) and synthetic (ModelNet40, ModelNet10) and real scanned (ScanObjectNN) datasets, we establish new baseline results on learning without forgetting for 3D data. This research will instigate many future works in this area.
翻译:当我们微调一套新班级经过良好训练的深层次学习模式时,网络会学习新概念,但逐渐忘记旧培训的知识。在某些现实应用中,我们可能有兴趣在不忘记以往经验的能力的情况下学习新课程。这种不忘记问题的学习常常使用2D图像识别任务来调查。在本文中,考虑到深度摄像技术的增长,我们处理3D点云天数据同样的问题。由于缺少大型数据集和强大的预先训练的骨干模型,这个问题在3D领域比2D更具有挑战性。我们调查3D数据的知识蒸馏技术,以减少对以往培训的灾难性遗忘。此外,我们通过使用对象课程的语义语言矢量来改进蒸馏过程。我们观察到,在培训过程中探索新老知识的相互关系有助于学习新概念,而不会忘记旧概念。实验3D点云识别支柱(PointNet、DGCNN和Point Conv)和合成的(ModelNet40、ModelNet10)和真实扫描(ScentrentNNN),我们将在将来的研究工作中建立新的基线结果,而不会忘记这一数据领域。