Abstract

Purpose: Interpreting echocardiographic exams requires substantial manual interaction as videos lack scan-plane information and have inconsistent image quality, ranging from clinically relevant to unrecognizable. Thus, a manual prerequisite step for analysis is to select the appropriate views that showcase both the target anatomy and optimal image quality. To automate this selection process, we present a method for automatic classification of routine views, recognition of unknown views, and quality assessment of detected views. Approach: We train a neural network for view classification and employ the logit activations from the neural network for unknown view recognition. Subsequently, we train a linear regression algorithm that uses feature embeddings from the neural network to predict view quality scores. We evaluate the method on a clinical test set of 2466 echocardiography videos with expert-annotated view labels and a subset of 438 videos with expert-rated view quality scores. A second observer annotated a subset of 894 videos, including all quality-rated videos. Results: The proposed method achieved an accuracy of 84.9% ± 0.67 for the joint objective of routine view classification and unknown view recognition, whereas a second observer reached an accuracy of 87.6%. For view quality assessment, the method achieved a Spearman's rank correlation coefficient of 0.71, whereas a second observer reached a correlation coefficient of 0.62. Conclusion: The proposed method approaches expert-level performance, enabling fully automatic selection of the most appropriate views for manual or automatic downstream analysis.
Original languageEnglish
Pages (from-to)54002
Number of pages1
JournalJournal of Medical Imaging
Volume11
Issue number5
DOIs
Publication statusPublished - 1 Sept 2024

Keywords

  • cardiology
  • echocardiography
  • image quality assessment
  • open set recognition
  • view classification

Fingerprint

Dive into the research topics of 'Automated echocardiography view classification and quality assessment with recognition of unknown views'. Together they form a unique fingerprint.

Cite this