Progress in machine learning is typically measured by training and testing a model on the same distribution of data, i.e., the same domain. This over-estimates future accuracy on out-of-distribution data. The Visual Domain Adaptation (VisDA) 2021 competition tests models' ability to adapt to novel test distributions and handle distributional shift. We set up unsupervised domain adaptation challenges for image classifiers and will evaluate adaptation to novel viewpoints, backgrounds, modalities and degradation in quality. Our challenge draws on large-scale publicly available datasets but constructs the evaluation across domains, rather that the traditional in-domain bench-marking. Furthermore, we focus on the difficult "universal" setting where, in addition to input distribution drift, methods may encounter missing and/or novel classes in the target dataset. Performance will be measured using a rigorous protocol, comparing to state-of-the-art domain adaptation methods with the help of established metrics. We believe that the competition will encourage further improvement in machine learning methods' ability to handle realistic data in many deployment scenarios.
翻译:计算机学习的进展通常通过培训和测试数据相同分布模式来衡量,即同一领域。高估了未来分配外数据的准确性。2021年视觉域适应(VisDA)竞争测试模型适应新测试分布和处理分布转移的能力。我们为图像分类者设置了不受监督的域适应挑战,并将评估适应新观点、背景、模式和质量退化的情况。我们的挑战来自大规模公开可得到的数据集,但构建了跨领域的评价,而不是传统的域域域域域域域标记。此外,我们侧重于困难的“通用”设置,其中除了输入分布流外,在目标数据集中可能遇到缺失和/或新的类的方法。绩效将使用严格的协议加以衡量,与最先进的域适应方法相比,并借助既定的衡量尺度。我们认为,竞争将鼓励进一步提高机器学习方法在许多部署情景中处理现实数据的能力。