Steerable Pyramid Transform Enables Robust Left Ventricle Quantification

Xiangyang Zhu, Kede Ma*, Wufeng Xue*

*Corresponding author for this work

Research output: Chapters, Conference Papers, Creative and Literary WorksRGC 32 - Refereed conference paper (with host publication)peer-review

Abstract

Predicting cardiac indices has long been a focal point in the medical imaging community. While various deep learning models have demonstrated success in quantifying cardiac indices, they remain susceptible to mild input perturbations, e.g., spatial transformations, image distortions, and adversarial attacks. This vulnerability undermines confidence in using learning-based automated systems for diagnosing cardiovascular diseases. In this work, we describe a simple yet effective method to learn robust models for left ventricle (LV) quantification, encompassing cavity and myocardium areas, directional dimensions, and regional wall thicknesses. Our success hinges on employing the biologically inspired steerable pyramid transform (SPT) for fixed front-end processing, which offers three main benefits. First, the basis functions of SPT align with the anatomical structure of LV and the geometric features of the measured indices. Second, SPT facilitates weight sharing across different orientations as a form of parameter regularization and naturally captures the scale variations of LV. Third, the residual highpass subband can be conveniently discarded, promoting robust feature learning. Extensive experiments on the Cardiac-Dig benchmark show that our SPT-augmented model not only achieves reasonable prediction accuracy compared to state-of-the-art methods, but also exhibits significantly improved robustness against input perturbations. Code is available at https://github.com/yangyangyang127/RobustLV. © The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2025.
Original languageEnglish
Title of host publicationPattern Recognition and Computer Vision
Subtitle of host publication7th Chinese Conference, PRCV 2024, Urumqi, China, October 18–20, 2024, Proceedings, Part XIV
EditorsZhouchen Lin, Ming-Ming Cheng, Ran He, Kurban Ubul, Wushouer Silamu, Hongbin Zha, Jie Zhou, Cheng-Lin Liu
Place of PublicationSingapore
PublisherSpringer 
Pages32-45
ISBN (Electronic)978-981-97-8496-7
ISBN (Print) 978-981-97-8495-0
DOIs
Publication statusPublished - 2025
Event7th Chinese Conference on Pattern Recognition and Computer Vision (PRCV 2024) - Urumqi, China
Duration: 18 Oct 202420 Oct 2024
http://www.prcv.cn/

Publication series

NameLecture Notes in Computer Science
Volume15044
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349

Conference

Conference7th Chinese Conference on Pattern Recognition and Computer Vision (PRCV 2024)
Country/TerritoryChina
CityUrumqi
Period18/10/2420/10/24
Internet address

Bibliographical note

Full text of this publication does not contain sufficient affiliation information. With consent from the author(s) concerned, the Research Unit(s) information for this record is based on the existing academic department affiliation of the author(s).

Research Keywords

  • Left ventricle quantification
  • Robustness
  • Steerable pyramid

Fingerprint

Dive into the research topics of 'Steerable Pyramid Transform Enables Robust Left Ventricle Quantification'. Together they form a unique fingerprint.

Cite this