With the recent advances in technology, automatic speech recognition (ASR) has been widely used in real-world applications. The efficiency of converting large amounts of speech into text accurately with limited resources has become more important than ever. This paper proposes a method to rapidly recognize a large speech database via a Transformer-based end-to-end model. Transformers have improved the state-of-the-art performance in many fields. However, they are not easy to use for long sequences. In this paper, various techniques to speed up the recognition of real-world speeches are proposed and tested, including decoding via multiple-utterance batched beam search, detecting end-of-speech based on a connectionist temporal classification (CTC), restricting the CTC prefix score, and splitting long speeches into short segments. Experiments are conducted with the Librispeech English and the real-world Korean ASR tasks to verify the proposed methods. From the experiments, the proposed system can convert 8 hours of speeches spoken at real-world meetings into text in less than 3 minutes with a 10.73% character error rate, which is 27.1% relatively lower than that of conventional systems.
翻译:随着技术的最新进步,自动语音识别(ASR)在现实世界应用中被广泛使用。用有限资源准确地将大量语音转换成文本的效率比以往更加重要。本文件提出了一个通过基于变异器的端对端模式快速识别大型语音数据库的方法。变异器改善了许多领域的最先进性能。但是,它们不容易用于长的顺序。在本文中,提出并测试了加速承认真实世界演讲的各种技术,包括通过多发量的分批搜索解码,根据连接时间分类(CTC)探测终端语音,限制CTC前缀分数,将长话分解成短段。与Librispeech英语和真实世界韩国ASR任务一起进行实验,以核实拟议方法。在实验中,拟议的系统可以将现实世界会议上8小时的演讲转换为不超过3分钟的文本,使用10.73%的性格误差率,比常规系统低27.1%。