The 3D deep learning community has seen significant strides in pointcloud processing over the last few years. However, the datasets on which deep models have been trained have largely remained the same. Most datasets comprise clean, clutter-free pointclouds canonicalized for pose. Models trained on these datasets fail in uninterpretible and unintuitive ways when presented with data that contains transformations "unseen" at train time. While data augmentation enables models to be robust to "previously seen" input transformations, 1) we show that this does not work for unseen transformations during inference, and 2) data augmentation makes it difficult to analyze a model's inherent robustness to transformations. To this end, we create a publicly available dataset for robustness analysis of point cloud classification models (independent of data augmentation) to input transformations, called RobustPointSet. Our experiments indicate that despite all the progress in the point cloud classification, there is no single architecture that consistently performs better -- several fail drastically -- when evaluated on transformed test sets. We also find that robustness to unseen transformations cannot be brought about merely by extensive data augmentation. RobustPointSet can be accessed through https://github.com/AutodeskAILab/RobustPointSet.
翻译:在过去几年里,3D深层学习社区在点球处理方面取得了长足的进步。 然而, 深模型所培训的数据集基本上没有变化。 大多数数据集都是干净的, 零碎的点球球, 用来装扮。 这些数据集所训练的模型在显示含有“ 看不见” 变异的数据时, 无法解释和不直观地失败。 虽然数据增强使模型能够稳健到“ 先前看到” 输入转换, 1 我们显示, 这在推断期间对看不见的变异没有作用, 2 数据增强使得很难分析模型对变异固有的稳健性。 为此,我们创建了一个公开可用的数据集,用于对点云分类模型(依靠数据增强)进行稳健性分析,称为 RobustPointSet。 我们的实验表明,尽管在点云分类中取得了所有进展,但在对变换的测试机组进行评估时,没有一种持续更好的结构 -- 有几个失败的。 我们还发现, 无法通过广泛的数据放大系统/ MABODS。