少点错误 2024年10月07日
An X-Ray is Worth 15 Features: Sparse Autoencoders for Interpretable Radiology Report Generation
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

SAE-Rad 是一种基于稀疏自动编码器 (SAE) 的放射学报告生成框架,通过对放射图像编码器的类标记进行训练,提取可解释的特征,并利用这些特征生成临床相关的放射学报告。SAE-Rad 在准确率方面与最先进的医疗基础模型相比具有竞争力,同时参数数量和计算成本仅为其一小部分。该框架首次将 SAE 应用于非平凡的下游任务,即对医学图像进行多模态推理。

👨‍⚕️ **SAE-Rad 的核心是利用稀疏自动编码器 (SAE) 对放射图像编码器的类标记进行训练,提取可解释的特征。** SAE 的训练目标是学习一组稀疏激活的特征,这些特征能够捕获图像中的重要信息,如病理特征、器械特征等。这种稀疏性使得提取的特征具有可解释性,因为每个特征都对应于图像中的一个特定概念。

📊 **SAE-Rad 通过将 SAE 提取的特征与对应的文本描述相结合,生成详细的放射学报告。** 为了实现这一点,SAE-Rad 首先从训练数据集中检索与每个特征最高激活的图像,并使用这些图像的参考报告来生成该特征的文本描述。在生成报告时,SAE-Rad 会识别输入图像中激活的特征,并使用相应的文本描述来构建报告。

🔬 **SAE-Rad 通过对 SAE 特征进行干预,并使用扩散模型重建图像,来验证提取特征的可解释性和有效性。** 该框架可以通过控制 SAE 特征的激活状态,来观察图像的重建结果,从而验证特征的解释是否与实际图像变化相符。例如,通过增加“心肌肥大”特征的激活,可以观察到重建图像中心脏区域的扩大,这表明该特征确实对应于心肌肥大这一概念。

🖼️ **SAE-Rad 的优势在于其可解释性、高效性和准确性。** 与其他基于大型语言模型 (LLM) 的放射学报告生成方法相比,SAE-Rad 不需要对 LLM 进行微调,从而减少了计算成本和训练时间。同时,SAE-Rad 能够生成与图像内容相一致的报告,并且可以解释每个报告语句背后的图像特征,提高了报告的可信度。

💡 **SAE-Rad 为放射学报告生成提供了一种新的思路,并为医学图像理解和人工智能辅助诊断开辟了新的可能性。** 该框架可以有效地提取图像中的关键信息,并将其转化为人类可理解的语言描述,从而为临床医生提供更准确、更全面的诊断信息。

Published on October 7, 2024 8:53 AM GMT

This is an archival link-post for our preprint, which can be found here.

Figure 1: SAE-Rad identifies clinically relevant and interpretable features within radiological images. We illustrate a number of pathological and instrumentation features relevant for producing radiology reports. We add annotations (green arrows) to emphasize the presence of each feature.

Executive Summary

This post is a heavily slimmed down summary of our main paper, linked above. We have omitted all the technical details here. This post acts as a TL;DR archival link-post to the main paper.

We train Sparse Autoencoders (SAEs) on the class token of a radiology image encoder, on a dataset of chest x-rays. We use the trained SAE, in conjunction with automated interpretability, to generate radiology reports. The final radiology report represents a concatenation of the text descriptions of activate SAE features. We train a diffusion model to allow causal interventions on SAE features. This diffusion model enables us to highlight where in the chest x-ray each sentence in the radiology report comes from by localising changes in the image post-intervention. Our method achieves competitive accuracy in comparison to state of the art medical foundation models while using a fraction of the parameter count and compute costs. To the best of our knowledge, this is the first time SAEs have been used for a non-trivial downstream task - namely to perform multi-modal reasoning on medical images.

Of particular note to the mechanistic interpretability community, we demonstrate that SAEs extract sparse and interpretable features on a small dataset (240,000) of homogenous images (chest x-rays appear very homogenous), and that these features can be accurately labeled by means of automated interpretability to produce pathologically relevant findings.

Motivation

Radiological services are essential to modern clinical practice, with demand rising rapidly. In the UK, the NHS performs over 43 million radiological procedures annually, costing over £2 billion, and demand for scans more than doubled between 2012 and 2019. A significant portion of these costs addresses rising demand through agency, bank, and overtime staff, but a national imaging strategy notes this funding is unsustainable. Consequently, there’s growing interest in (semi)-automating tasks like radiology report generation, augmentation, and summarization to assist clinicians, spurred by advances in multimodal text-vision modelling techniques.

Recent architectures that combine vision encoders with pretrained Large Language Models (LLMs) to create multimodal Vision-Language Models (VLMs) have shown impressive performance in visual and language tasks. VLMs have been applied to healthcare tasks, including radiology report generation, typically by mapping image representations into the LLM’s token embedding space. The LLM is fine-tuned to respond to prompts like ‘Produce the findings section of a radiology report for this image’.

Despite improvements from scaling VLMs, hallucinations and disagreements with domain experts remain common. Hallucinations are unavoidable in LLMs, and whilst this represents a limitation of current VLM systems designed for radiology report generation, there are other important considerations of using such a system for this task. For current state-of-the-art systems, it is necessary to finetune a multi-billion parameter LLM (as well as projector weights) to perform visual instruction tuning, which is computationally intensive and can be prohibitively expensive. Additionally, the generated reports a VLM provides may not be faithful to the underlying computations of the image encoder – we should aim to design a framework that is verifiably faithful to the image model by reverse engineering the computations of the image encoder. This could yield more interpretable results and thus engender more trust in automated radiology reporting systems.

To this end, we introduce 'SAE-Rad', a framework which leverages SAEs, to directly decompose image class tokens from a pre-trained radiology image encoder into human-interpretable features.

Radiology Reporting Pipeline

Please see the pre-print on ArXiv for the detailed experimental setup. We trained an SAE with an expansion factor of 64. The SAE used had the architecture of a gated SAE but without normalising the decoder weights. At the end of training, the SAE had an  and an explained variance of . Figure 2 displays the radiology reporting pipeline.

Figure 2: SAE-Rad overview. Panel A: We learn a set of sparsely activating features by training a Sparse Autoencoder (SAE) on class tokens produced by a radiology-image encoder. Panel B: We retrieve the corresponding reference reports for highest activating images for a feature, from which we can produce text descriptions of each feature. Panel C: We pass a new image through the radiology-image encoder and SAE encoder to retrieve the highest activating features. Text descriptions of these features are subsequently used by a pretrained large language model (LLM) to generate a detailed radiology report.

Results

Quantitative Evaluation

For details of the metrics and datasets reported in this section, please refer to the preprint. 

We compared SAE-Rad to the current state-of-the-art radiology reporting systems. CheXagent is an instruction-tuned foundation model for CXRs trained on 1.1M scans for question-answering and text-generation tasks. MAIRA-1 &-2 are VLMS based on the LLaVA 1.5 architecture. MAIRA-2 is trained on 510,848 CXRs from four datasets and sets the current state-of-the art for report generation. The MAIRA systems are not publicly available for result replication, and thus we quote their evaluation values directly as our upper-bound. CheXagent is publicly available, and we therefore performed independent replications for this model for a direct comparison.

As Table 1 demonstrates, SAE-Rad underperforms on generic NLG metrics such as BLEU-4. This is expected as we do not try to optimize for any specific ‘writing style’ by fine-tuning an LLM on the reference reports. Conversely, SAE-Rad demonstrates strong performance on radiology-specific metrics which are clinically relevant, outperforming CheXagent by up to 52% in the CheXpert F1 score (macro-averaged F1-14), and achieving 92.1% and 89.9% of the performance of MAIRA-1 and MAIRA-2 on these scores, respectively.

Table 1: Report generation performance on the official MIMIC-CXR test split. BL4 = BLEU-4, RG-L = ROUGE-L, MTR = Meteor. Ma-5 (Macro-F1-5), Ma-14 (Macro-F1-14), Mi-5 (Micro-F1- 5), and Mi-14 (Micro-F1-14) represent the clinical CheXbert labeler scores. Bolding represents best performance in the current study or between the upper bound models.

SAE Features

Figure 1 displays some clinically relevant SAE features. In this section we showcase highest activating images for a number of other features, as well as the corresponding feature explanations. We highlight the variety of features captured by SAE-Rad, from instrumentation features to visual features such as radiograph inversion, pathology-related features, and small details such as piercings.

Maximally activating images for a feature corresponding to bowel obstruction.
Maximally activating images for a feature corresponding to deep brain stimulators which are typically used to treat Parkinson’s disease.
Maximally activating images for a feature corresponding to orthopaedic rods and screws.
Maximally activating images for a feature corresponding to right sided intra-jugular lines in the presence of sternotomy wires.
Maximally activating images for a feature corresponding to piercings.
Maximally activating images for a feature corresponding to inverted radiographs.
Maximally activating images for a feature corresponding to female radiographs with no pathology detected.

Counterfactual Image Generation

We evaluated the interpretability and validity of our SAE features by intervening on SAE features and then reconstructing the resulting x-rays through a diffusion model. SAE features are interpretable if they correspond to distinct concepts that respond predictably to activation space interventions. We trained a diffusion model conditioned on the class tokens of a radiology image encoder, to reconstruct the radiographs. During inference, we passed a class token through the SAE, intervened on the hidden feature activations, and reconstructed a “counterfactual” class token via the SAE decoder, which conditioned the diffusion model to project interventions into imaging space. We tested whether: 1) interventions alter the reconstructed class token accordingly, 2) changes affect only the targeted feature, and 3) features can be “added” or “removed” by manipulating the same activation. Figure 3 shows the results for two features (cardiomegaly and pacemaker), demonstrating that our interpretations accurately reflect their impact on model behaviour. Figure 3 also illustrates how these methods can be used to ground the radiology report in the chest x-rays through unsupervised segmentation.

Figure 3: SAE-Rad enables counterfactual image generation and unsupervised segmentation with disentangled class tokens. Row 1 examines a pacemaker, and Row 2 investigates cardiomegaly. Column 1 shows original images, Column 2 shows model reconstructions, and Columns 3 and 4 depict counterfactuals by adding and removing features. The final column demonstrates unsupervised segmentation by comparing counterfactual and original images.


Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

SAE-Rad 放射学报告生成 稀疏自动编码器 可解释性 医学图像理解
相关文章