Abstract
Deep learning models have demonstrated strong performance in electrocardiogram (ECG) arrhythmia classification. However, their lack of interpretability limits clinical trust and adoption. By adopting an explainable artificial intelligence (XAI) technique, this study aims to enhance the interpretability of a convolutional neural network (CNN) model. More specifically, the Local Interpretable Model-Agnostic Explanations (LIME) technique is utilized to interpret the CNN model used to classify 17 classes of ECG arrhythmias. The CNN model was developed using a five-stage framework. The study uses the MIT-BIH Arrhythmia database to evaluate the performance of the CNN model. Results indicate that the model was able to accomplish precision of 97.00%, recall of 97.00%, F1-score of 97.00%, and overall accuracy of 99.00%. In addition, the LIME technique provides local explanations that help in the understanding of the decision-making process of the CNN model in classifying the 17 classes of ECG arrhythmias.
Metadata
| Item Type: | Article |
|---|---|
| Creators: | Creators Email / ID Num. Mohd Khairuddin, Adam adam.mk@utm.my Mohd Aris, Siti Armiza UNSPECIFIED Azizan, Azizul UNSPECIFIED Zakaria, Noor Jannah UNSPECIFIED |
| Subjects: | L Education > LG Individual institutions > Asia > Malaysia > Universiti Teknologi MARA > Perak Q Science > QA Mathematics |
| Divisions: | Universiti Teknologi MARA, Perak > Tapah Campus > Faculty of Computer and Mathematical Sciences |
| Journal or Publication Title: | Mathematical Sciences and Informatics Journal (MIJ) |
| UiTM Journal Collections: | UiTM Journals > Mathematical Science and Information Journal (MIJ) |
| ISSN: | 2735-0703 |
| Volume: | 6 |
| Number: | 2 |
| Page Range: | pp. 247-260 |
| Keywords: | Interpretable, Arrhythmia, Classification, Electrocardiography, Convolutional neural network |
| Date: | October 2025 |
| URI: | https://ir.uitm.edu.my/id/eprint/128980 |
