It is very challenging for various visual tasks such as image fusion, pedestrian detection and image-to-image translation in low light conditions due to the loss of effective target areas. In this case, infrared and visible images can be used together to provide both rich detail information and effective target areas. In this paper, we present LLVIP, a visible-infrared paired dataset for low-light vision. This dataset contains 33672 images, or 16836 pairs, most of which were taken at very dark scenes, and all of the images are strictly aligned in time and space. Pedestrians in the dataset are labeled. We compare the dataset with other visible-infrared datasets and evaluate the performance of some popular visual algorithms including image fusion, pedestrian detection and image-to-image translation on the dataset. The experimental results demonstrate the complementary effect of fusion on image information, and find the deficiency of existing algorithms of the three visual tasks in very low-light conditions. We believe the LLVIP dataset will contribute to the community of computer vision by promoting image fusion, pedestrian detection and image-to-image translation in very low-light applications. The dataset is being released in https://bupt-ai-cz.github.io/LLVIP.
翻译:对于各种视觉任务,例如图像聚合、行人探测和图像到图像翻译等,由于丧失有效目标区域,在低光条件下的低光条件下的图像光化、行人探测和图像到图像的翻译非常困难。在此情况下,红红外图像和可见图像可以一起使用,以提供丰富的详细信息和有效的目标区域。在本文中,我们介绍了一个可见的红外对齐的低光视觉数据集LLVIP。该数据集包含33672张图像,或16836对,其中多数是在非常暗的场景拍摄的,所有图像都严格在时间和空间上对齐。数据集中贴有污点标签。我们将数据集与其他可见红外数据集进行比较,并评价一些流行的视觉算法的性能,包括图像混集、行人探测和图像到图像的翻译。实验结果显示了对图像信息的补充效应,并在非常低光的场景条件下发现三种视觉任务的现有算法的缺陷。我们认为LLLVIP数据集将促进图像聚合、行人探测和图像显示/图像到LLLU的低光度应用。在 MAs-lishup-LSet中,这是低光/LUDADD。在低光/MDDDDDDD中,这是非常低的低的低的翻译。