In light of the COVID-19 pandemic, patients were required to manually input their daily oxygen saturation (SpO2) and pulse rate (PR) values into a health monitoring system-unfortunately, such a process trend to be an error in typing. Several studies attempted to detect the physiological value from the captured image using optical character recognition (OCR). However, the technology has limited availability with high cost. Thus, this study aimed to propose a novel framework called PACMAN (Pandemic Accelerated Human-Machine Collaboration) with a low-resource deep learning-based computer vision. We compared state-of-the-art object detection algorithms (scaled YOLOv4, YOLOv5, and YOLOR), including the commercial OCR tools for digit recognition on the captured images from pulse oximeter display. All images were derived from crowdsourced data collection with varying quality and alignment. YOLOv5 was the best-performing model against the given model comparison across all datasets, notably the correctly orientated image dataset. We further improved the model performance with the digits auto-orientation algorithm and applied a clustering algorithm to extract SpO2 and PR values. The accuracy performance of YOLOv5 with the implementations was approximately 81.0-89.5%, which was enhanced compared to without any additional implementation. Accordingly, this study highlighted the completion of PACMAN framework to detect and read digits in real-world datasets. The proposed framework has been currently integrated into the patient monitoring system utilized by hospitals nationwide.
翻译:鉴于COVID-19大流行,病人必须将其日常氧饱和(SpO2)和脉搏率(PR)值人工输入一个健康监测系统 -- -- 不幸地,这种过程趋势在打字方面是一个错误。一些研究试图利用光学字符识别(OCR)从所摄图像中检测到生理价值。然而,技术的可用性有限,成本高。因此,这项研究的目的是提出一个名为PACMAN(加速人类-机器协作)的新框架,以低资源深层次学习的计算机视觉为基础。我们比较了最先进的病人检测算法(标为YOLOv4、YOLOv5和YOLOR),包括用于通过光学字符识别(OCR)所摄取的图像的商用OCR工具。所有图像都来自质量和一致性不等的众包数据收集。YOLV是相对于所有数据集的拟议模型比较的最佳表现模型模型,特别是正确或更新的图像数据集。我们进一步改进了模型的性能,将数字模型的性能与目前自动定位算法的自动定位算法比起来,并且将改进了SPRO5的性数据采集系统。