Convolutional neural networks (CNN) have demonstrated outstanding Compressed Sensing (CS) performance compared to traditional, hand-crafted methods. However, they are broadly limited in terms of generalisability, inductive bias and difficulty to model long distance relationships. Transformer neural networks (TNN) overcome such issues by implementing an attention mechanism designed to capture dependencies between inputs. However, high-resolution tasks typically require vision Transformers (ViT) to decompose an image into patch-based tokens, limiting inputs to inherently local contexts. We propose a novel image decomposition that naturally embeds images into low-resolution inputs. These Kaleidoscope tokens (KD) provide a mechanism for global attention, at the same computational cost as a patch-based approach. To showcase this development, we replace CNN components in a well-known CS-MRI neural network with TNN blocks and demonstrate the improvements afforded by KD. We also propose an ensemble of image tokens, which enhance overall image quality and reduces model size. Supplementary material is available: https://github.com/uqmarlonbran/TCS.git
翻译:与传统的手工制作方法相比,电动神经网络(CNN)表现出了杰出的压缩感应性能,但是,在一般性、感应偏差和模拟长距离关系的困难方面,它们普遍有限。变形神经网络(TNN)通过实施一种旨在捕捉投入之间依赖性的注意机制克服了这些问题。然而,高分辨率任务通常要求视觉变异器(VT)将图像分解成基于补丁的标志,将输入限于内在的当地环境。我们建议一种新颖的图像分解,将图像自然地嵌入低分辨率投入中。这些Kaleidorocos标志(KD)提供了一种全球关注机制,其计算成本与基于宽度的方法相同。为展示这一发展,我们用广受欢迎的CS-MRI神经网络中的部件以TNNNU块取代,并展示KD提供的改进。我们还提议了一套图像标记,以提高总体图像质量并降低模型大小。有补充材料:https://github.com/uqmarbran/Sgistratt.