This paper investigates Electrocardiogram (ECG) rhythm classification using a progressive deep learning framework that combines time–frequency representations with complementary hand-crafted features. In the first stage, ECG signals from the PhysioNet Challenge 2017 dataset are transformed into scalograms and input to diverse architectures, including Simple Convolutional Neural Network (SimpleCNN), Residual Network with 18 Layers (ResNet-18), Convolutional Neural Network-Transformer (CNNTransformer), and Vision Transformer (ViT). ViT achieved the highest accuracy (0.8590) and F1-score (0.8524), demonstrating the feasibility of pure image-based ECG analysis, although scalograms alone showed variability across folds. In the second stage, scalograms were fused with scattering and statistical features, enhancing robustness and interpretability. FusionViT without dimensionality reduction achieved the best performance (accuracy = 0.8623, F1-score = 0.8528), while Fusion ResNet-18 offered a favorable trade-off between accuracy (0.8321) and inference efficiency (0.016 s per sample). The application of Principal Component Analysis (PCA) reduced the dimensionality of the feature from 509 to 27, reducing the computational cost while maintaining competitive performance (FusionViT precision = 0.8590). The results highlight a trade-off between efficiency and fine-grained temporal resolution. Training-time augmentations mitigated class imbalance, enabling lightweight inference (0.006–0.043 s per sample). For real-world use, the framework can run on wearable ECG devices or mobile health apps. Scalogram transformation and feature extraction occur on-device or at the edge, with efficient models like ResNet-18 enabling near real-time monitoring. Abnormal rhythm alerts can be sent instantly to users or clinicians. By combining visual and statistical signal features, optionally reduced with PCA, the framework achieves high accuracy, robustness, and efficiency for practical deployment.

Generalizable Hybrid Wavelet–Deep Learning Architecture for Robust Arrhythmia Detection in Wearable ECG Monitoring / Thapa, Ukesh; Pati, Bipun Man; Taparugssanagorn, Attaphongse; Mucchi, Lorenzo. - In: SENSORS. - ISSN 1424-8220. - ELETTRONICO. - 25:(2025), pp. 6590.1-6590.26. [10.3390/s25216590]

Generalizable Hybrid Wavelet–Deep Learning Architecture for Robust Arrhythmia Detection in Wearable ECG Monitoring

Taparugssanagorn, Attaphongse;Mucchi, Lorenzo
2025

Abstract

This paper investigates Electrocardiogram (ECG) rhythm classification using a progressive deep learning framework that combines time–frequency representations with complementary hand-crafted features. In the first stage, ECG signals from the PhysioNet Challenge 2017 dataset are transformed into scalograms and input to diverse architectures, including Simple Convolutional Neural Network (SimpleCNN), Residual Network with 18 Layers (ResNet-18), Convolutional Neural Network-Transformer (CNNTransformer), and Vision Transformer (ViT). ViT achieved the highest accuracy (0.8590) and F1-score (0.8524), demonstrating the feasibility of pure image-based ECG analysis, although scalograms alone showed variability across folds. In the second stage, scalograms were fused with scattering and statistical features, enhancing robustness and interpretability. FusionViT without dimensionality reduction achieved the best performance (accuracy = 0.8623, F1-score = 0.8528), while Fusion ResNet-18 offered a favorable trade-off between accuracy (0.8321) and inference efficiency (0.016 s per sample). The application of Principal Component Analysis (PCA) reduced the dimensionality of the feature from 509 to 27, reducing the computational cost while maintaining competitive performance (FusionViT precision = 0.8590). The results highlight a trade-off between efficiency and fine-grained temporal resolution. Training-time augmentations mitigated class imbalance, enabling lightweight inference (0.006–0.043 s per sample). For real-world use, the framework can run on wearable ECG devices or mobile health apps. Scalogram transformation and feature extraction occur on-device or at the edge, with efficient models like ResNet-18 enabling near real-time monitoring. Abnormal rhythm alerts can be sent instantly to users or clinicians. By combining visual and statistical signal features, optionally reduced with PCA, the framework achieves high accuracy, robustness, and efficiency for practical deployment.
2025
25
1
26
Goal 3: Good health and well-being
Thapa, Ukesh; Pati, Bipun Man; Taparugssanagorn, Attaphongse; Mucchi, Lorenzo
File in questo prodotto:
File Dimensione Formato  
sensors-25-06590-v2.pdf

Accesso chiuso

Tipologia: Pdf editoriale (Version of record)
Licenza: Open Access
Dimensione 1.43 MB
Formato Adobe PDF
1.43 MB Adobe PDF   Richiedi una copia

I documenti in FLORE sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificatore per citare o creare un link a questa risorsa: https://hdl.handle.net/2158/1449573
Citazioni
  • ???jsp.display-item.citation.pmc??? 1
  • Scopus 1
  • ???jsp.display-item.citation.isi??? 1
social impact