Naidoo, Preshen ORCID: https://orcid.org/0009-0004-8328-6194, Fernandes, Patricia
ORCID: https://orcid.org/0009-0000-9720-2829, Dadashi Serej, Nasim
ORCID: https://orcid.org/0000-0002-2898-1926, Stowell, Catherine C
ORCID: https://orcid.org/0000-0003-3005-8636, Manisty, Charlotte H
ORCID: https://orcid.org/0000-0003-0245-7090, Francis, Darrel P
ORCID: https://orcid.org/0000-0002-3410-0814 and Zolgharni, Massoud
ORCID: https://orcid.org/0000-0003-0904-2904
(2026)
Robust fine-grained echocardiographic view classification with supervised contrastive learning.
Medical Image Analysis, 110.
p. 104006.
ISSN 13618415
Preview |
PDF/A
1-s2.0-S1361841526000757-main.pdf - Published Version Available under License Creative Commons Attribution No Derivatives. Download (11MB) | Preview |
Abstract
Accurate classification of echocardiographic views is fundamental for automated cardiac analysis. However, clinical practice relies on a large, heterogeneous set of fine-grained acquisitions that introduce substantial inter-observer variability. Existing studies have primarily focused on limited view sets, often collapsing specialised views into broad categories, which limits their clinical relevance. We address this limitation by introducing TTE47, the first publicly available benchmark comprising 47 clinically meaningful views annotated independently by three experts. This dataset enables the rigorous quantification of inter-observer agreement and establishes a foundation for reproducible, clinically relevant evaluation. To tackle the dual challenges of subtle inter-class distinctions and structured label variability, we propose a novel supervised contrastive learning framework incorporating a tailored loss function. Our method outperforms cross-entropy and standard supervised contrastive baselines, achieving leading performance among evaluated methods on TTE47 and surpassing prior work on TMED-2 without dataset-specific pretraining, using a model pretrained on TTE47. Beyond accuracy, we introduce clustering-based metrics, Detection Rate and Label Recovery Precision, that measure semantic coherence and the model’s ability to resist annotation variability. Results show that the learned feature space aligns more strongly with underlying anatomical structure than with any single annotator’s style, enabling resilience to label shifts and maintaining robustness comparable to human-level disagreement. By integrating multi-expert evaluation, robust representation learning, and interpretable feature-space analysis, this work establishes a scalable and clinically relevant framework for fine-grained echo view classification. Our findings highlight the potential of contrastive pretraining to standardise interpretation, mitigate subjectivity, and enhance the reliability of AI-assisted echocardiography in diverse clinical settings.
| Item Type: | Article |
|---|---|
| Identifier: | 10.1016/j.media.2026.104006 |
| Keywords: | Echocardiography; Contrastive learning; View; classification; Observer variability |
| Subjects: | Computing > Innovation and user experience > Computing interaction design Medicine and health |
| Date Deposited: | 04 Mar 2026 |
| URI: | https://repository.uwl.ac.uk/id/eprint/14691 |
Downloads
Downloads per month over past year
Actions (admin access)
![]() |
Lists
Lists