The analysis of electrocardiogram (ECG) signals can be time consuming as it is performed manually by cardiologists. Therefore, automation through machine learning (ML) classification is being increasingly proposed which would allow ML models to learn the features of a heartbeat and detect abnormalities. The lack of interpretability hinders the application of Deep Learning in healthcare. Through interpretability of these models, we would understand how a machine learning algorithm makes its decisions and what patterns are being followed for classification. This thesis builds Convolutional Neural Network (CNN) and Long Short-Term Memory (LSTM) classifiers based on state-of-the-art models and compares their performance and interpretability to shallow classifiers. Here, both global and local interpretability methods are exploited to understand the interaction between dependent and independent variables across the entire dataset and to examine model decisions in each sample, respectively. Partial Dependence Plots, Shapley Additive Explanations, Permutation Feature Importance, and Gradient Weighted Class Activation Maps (Grad-Cam) are the four interpretability techniques implemented on time-series ML models classifying ECG rhythms. In particular, we exploit Grad-Cam, which is a local interpretability technique and examine whether its interpretability varies between correctly and incorrectly classified ECG beats within each class. Furthermore, the classifiers are evaluated using K-Fold cross-validation and Leave Groups Out techniques, and we use non-parametric statistical testing to examine whether differences are significant. It was found that Grad-CAM was the most effective interpretability technique at explaining predictions of proposed CNN and LSTM models. We concluded that all high performing classifiers looked at the QRS complex of the ECG rhythm when making predictions.
翻译:心电图(ECG)信号分析由于由心脏病学家手工进行,可能会耗时。 因此,越来越多地提议通过机器学习(ML)分类实现自动化,使ML模型能够学习心跳特征并检测异常。 缺乏解释性会妨碍深层学习在保健中的应用。 通过这些模型的可理解性,我们会理解机器学习算法如何做出其决定,以及遵循何种模式进行分类。 这篇论文建构了以最新模型为基础的CulvialQal网络(CNN)和长期短期内存(LSTM)分类,并将其性能和可解释性与浅层分类器进行比较。 这里,全球和地方的可解释性模型都被用来了解整个数据集中依赖性和独立变量之间的相互作用,并分别检查每个样本中的模式决定。 部分依赖性 Plots, Shapley Additite解释, Permotionation Reformate ECTAVAL 和SLLLA(GAGA)的精度评估发现(GRA-C-C)是用于时间序列的四种可解释性技术,我们不断解读的精选的精选的精选性计算。