Volumetric imaging by fluorescence microscopy is often limited by anisotropic spatial resolution from inferior axial resolution compared to the lateral resolution. To address this problem, here we present a deep-learning-enabled unsupervised super-resolution technique that enhances anisotropic images in volumetric fluorescence microscopy. In contrast to the existing deep learning approaches that require matched high-resolution target volume images, our method greatly reduces the effort to put into practice as the training of a network requires as little as a single 3D image stack, without a priori knowledge of the image formation process, registration of training data, or separate acquisition of target data. This is achieved based on the optimal transport driven cycle-consistent generative adversarial network that learns from an unpaired matching between high-resolution 2D images in lateral image plane and low-resolution 2D images in the other planes. Using fluorescence confocal microscopy and light-sheet microscopy, we demonstrate that the trained network not only enhances axial resolution beyond the diffraction limit, but also enhances suppressed visual details between the imaging planes and removes imaging artifacts.
翻译:荧光显微镜的体积成像往往受到与横向分辨率相比较低的轴分辨率低的厌食空间分辨率的厌食空间分辨率的限制。为了解决这一问题,我们在这里展示了一种深学、不受监督的超分辨率技术,在体积荧光显微镜中增强厌食性成像。与现有的深学方法相比,需要高分辨率目标体积图像,我们的方法大大减少了投入实践的努力,因为培训一个网络需要的只有一个仅3D图像堆,而不需要事先了解图像形成过程、培训数据的登记或单独获取目标数据。这是以最佳运输驱动周期性循环一致的遗传对抗网络为基础实现的,这种网络从横向图像平面高分辨率2D图像与其他平面低分辨率2D图像之间不匹配中学习。我们使用荧光镜的显微镜和光谱显微镜,我们证明经过培训的网络不仅能增强超过分解极限的氧化分辨率,而且还能加强图像和图像的压制性成像成像之间的图像。