Enhancing Interpretability of Vertebrae Fracture Grading using Human-interpretable Prototypes
arxiv(2024)
摘要
Vertebral fracture grading classifies the severity of vertebral fractures,
which is a challenging task in medical imaging and has recently attracted Deep
Learning (DL) models. Only a few works attempted to make such models
human-interpretable despite the need for transparency and trustworthiness in
critical use cases like DL-assisted medical diagnosis. Moreover, such models
either rely on post-hoc methods or additional annotations. In this work, we
propose a novel interpretable-by-design method, ProtoVerse, to find relevant
sub-parts of vertebral fractures (prototypes) that reliably explain the model's
decision in a human-understandable way. Specifically, we introduce a novel
diversity-promoting loss to mitigate prototype repetitions in small datasets
with intricate semantics. We have experimented with the VerSe'19 dataset and
outperformed the existing prototype-based method. Further, our model provides
superior interpretability against the post-hoc method. Importantly, expert
radiologists validated the visual interpretability of our results, showing
clinical applicability.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要