Common fully glazed facades and transparent objects present architectural barriers and impede the mobility of people with low vision or blindness, for instance, a path detected behind a glass door is inaccessible unless it is correctly perceived and reacted. However, segmenting these safety-critical objects is rarely covered by conventional assistive technologies. To tackle this issue, we construct a wearable system with a novel dual-head Transformer for Transparency (Trans4Trans) model, which is capable of segmenting general and transparent objects and performing real-time wayfinding to assist people walking alone more safely. Especially, both decoders created by our proposed Transformer Parsing Module (TPM) enable effective joint learning from different datasets. Besides, the efficient Trans4Trans model composed of symmetric transformer-based encoder and decoder, requires little computational expenses and is readily deployed on portable GPUs. Our Trans4Trans model outperforms state-of-the-art methods on the test sets of Stanford2D3D and Trans10K-v2 datasets and obtains mIoU of 45.13% and 75.14%, respectively. Through various pre-tests and a user study conducted in indoor and outdoor scenarios, the usability and reliability of our assistive system have been extensively verified.
翻译:例如,在玻璃门后面探测到的一条路径无法进入,除非得到正确的认识和反应。然而,传统的辅助技术很少覆盖这些安全关键物体的分割。为了解决这一问题,我们建造了一个可磨损的系统,配有新型的双头透明变换器(Trans4Trans)模型(Trans4Trans),它能够将普通的和透明的对象分割开来,并进行实时探测,帮助人们更安全地单独行走。特别是,由我们拟议的变换器剖析模块(TPM)创建的解码器能够从不同的数据集中有效地联合学习。此外,由以对称变换器为基础的编码和解码器组成的高效的Trans4Transer模型不需要多少计算费用,并且可以随时安装在便携式GPUP上。我们的Trans4转式模型在斯坦福2D3D和 Trans10K-V2测试集的测试集中能够分解通用的状态方法,并且获得45.13%和75.14% mIOU的有效联合学习。通过各种测试前和可变式的系统,对各种可变式的可变式和可变式系统进行了广泛的研究。