Computational sign language research lacks the large-scale datasets that enables the creation of useful reallife applications. To date, most research has been limited to prototype systems on small domains of discourse, e.g. weather forecasts. To address this issue and to push the field forward, we release six datasets comprised of 190 hours of footage on the larger domain of news. From this, 20 hours of footage have been annotated by Deaf experts and interpreters and is made publicly available for research purposes. In this paper, we share the dataset collection process and tools developed to enable the alignment of sign language video and subtitles, as well as baseline translation results to underpin future research.
翻译:计算手语研究缺乏能够产生有用的现实应用的大规模数据集,迄今为止,大多数研究仅限于小讨论领域(如天气预报)的原型系统,为了解决这一问题和推动实地向前推进,我们发布了六套数据集,其中包括190小时新闻领域的录像;从中,20小时的录像由聋人专家和口译员加注,供公众查阅,供研究之用;在本文件中,我们分享了数据集的收集过程和开发工具,以便协调手语视频和字幕,以及基线翻译结果,以支持今后的研究。