Robust fine-grained echocardiographic view classification with supervised contrastive learning

Naidoo, Preshen ORCID logoORCID: https://orcid.org/0009-0004-8328-6194, Fernandes, Patricia ORCID logoORCID: https://orcid.org/0009-0000-9720-2829, Dadashi Serej, Nasim ORCID logoORCID: https://orcid.org/0000-0002-2898-1926, Stowell, Catherine C ORCID logoORCID: https://orcid.org/0000-0003-3005-8636, Manisty, Charlotte H ORCID logoORCID: https://orcid.org/0000-0003-0245-7090, Francis, Darrel P ORCID logoORCID: https://orcid.org/0000-0002-3410-0814 and Zolgharni, Massoud ORCID logoORCID: https://orcid.org/0000-0003-0904-2904 (2026) Robust fine-grained echocardiographic view classification with supervised contrastive learning. Medical Image Analysis, 110. p. 104006. ISSN 13618415

[thumbnail of 1-s2.0-S1361841526000757-main.pdf]
Preview
PDF/A
1-s2.0-S1361841526000757-main.pdf - Published Version
Available under License Creative Commons Attribution No Derivatives.

Download (11MB) | Preview

Abstract

Accurate classification of echocardiographic views is fundamental for automated cardiac analysis. However, clinical practice relies on a large, heterogeneous set of fine-grained acquisitions that introduce substantial inter-observer variability. Existing studies have primarily focused on limited view sets, often collapsing specialised views into broad categories, which limits their clinical relevance. We address this limitation by introducing TTE47, the first publicly available benchmark comprising 47 clinically meaningful views annotated independently by three experts. This dataset enables the rigorous quantification of inter-observer agreement and establishes a foundation for reproducible, clinically relevant evaluation. To tackle the dual challenges of subtle inter-class distinctions and structured label variability, we propose a novel supervised contrastive learning framework incorporating a tailored loss function. Our method outperforms cross-entropy and standard supervised contrastive baselines, achieving leading performance among evaluated methods on TTE47 and surpassing prior work on TMED-2 without dataset-specific pretraining, using a model pretrained on TTE47. Beyond accuracy, we introduce clustering-based metrics, Detection Rate and Label Recovery Precision, that measure semantic coherence and the model’s ability to resist annotation variability. Results show that the learned feature space aligns more strongly with underlying anatomical structure than with any single annotator’s style, enabling resilience to label shifts and maintaining robustness comparable to human-level disagreement. By integrating multi-expert evaluation, robust representation learning, and interpretable feature-space analysis, this work establishes a scalable and clinically relevant framework for fine-grained echo view classification. Our findings highlight the potential of contrastive pretraining to standardise interpretation, mitigate subjectivity, and enhance the reliability of AI-assisted echocardiography in diverse clinical settings.

Item Type: Article
Identifier: 10.1016/j.media.2026.104006
Keywords: Echocardiography; Contrastive learning; View; classification; Observer variability
Subjects: Computing > Innovation and user experience > Computing interaction design
Medicine and health
Date Deposited: 04 Mar 2026
URI: https://repository.uwl.ac.uk/id/eprint/14691

Downloads

Downloads per month over past year

Actions (admin access)

View Item

Menu