MarkTechPost@AI 2024年11月30日
Enhancing Deep Learning-Based Neuroimaging Classification with 3D-to-2D Knowledge Distillation
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

本文介绍了一种名为3D-to-2D知识蒸馏(KD)的框架,用于增强2D卷积神经网络(CNN)在神经影像分析中的能力,尤其是在数据有限的情况下。该框架利用一个3D教师网络编码体积信息,并将知识蒸馏到一个专注于部分体积数据的2D学生网络。通过将该方法应用于帕金森病的分类任务,研究人员取得了优异的性能,F1分数达到98.30%。这种方法弥合了3D和2D影像之间的模态差距,提高了泛化能力,并解决了医学影像分析中的数据限制问题。此外,研究还探讨了不同的2D投影方法,如单切片、早期融合和联合融合,并证明了3D-to-2D KD在提高性能方面的有效性。

🤔**3D-to-2D知识蒸馏框架:**该框架包含一个3D教师网络(编码体积信息)和一个2D学生网络(学习部分体积数据),通过知识蒸馏将教师网络的知识转移到学生网络,从而提升学生网络对体积信息的理解能力。

📊**帕金森病分类实验:**该方法应用于帕金森病分类任务,使用123I-DaTscan SPECT和18F-AV133 PET数据集,取得了98.30%的F1分数,证明了其在实际应用中的有效性。

💡**多种2D投影方法探索:**研究人员探索了包括单切片、相邻切片(早期融合和联合融合)以及基于秩池化的动态图像等多种2D投影方法,并结合3D-to-2D KD进行性能提升。

🚀**特征级损失的重要性:**消融实验表明,基于特征的损失(Lfg)比基于logits的损失(Llg)对性能提升的影响更大,这说明了学习特征表示在3D-to-2D KD中的重要作用。

🔄**跨模态泛化能力:**该方法在SPECT和PET影像数据上都表现出色,证明了其跨模态的泛化能力,并且在有限数据集的情况下也能取得显著的性能提升。

Deep learning techniques are increasingly applied to neuroimaging analysis, with 3D CNNs offering superior performance for volumetric imaging. However, their reliance on large datasets is challenging due to the high cost and effort required for medical data collection and annotation. As an alternative, 2D CNNs utilize 2D projections of 3D images, which often limits volumetric context, affecting diagnostic accuracy. Techniques like transfer learning and knowledge distillation (KD) address these challenges by leveraging pre-trained models and transferring knowledge from complex teacher networks to simpler student models. These approaches enhance performance while maintaining generalizability in resource-constrained medical imaging tasks.

In neuroimaging analysis, 2D projection methods adapt 3D volumetric imaging for 2D CNNs, typically by selecting representative slices. Techniques like Shannon entropy have been used to identify diagnostically relevant slices, while methods like 2D+e enhance information by combining slices. KD, introduced by Hinton, transfers knowledge from complex models to simpler ones. Recent advances include cross-modal KD, where multimodal data enhances monomodal learning, and relation-based KD, which captures inter-sample relationships. However, applying KD to teach 2D CNNs, the volumetric relationships in 3D imaging still need to be explored despite its potential to improve neuroimaging classification with limited data.

Researchers from Dong-A University propose a 3D-to-2D KD framework to enhance 2D CNNs’ ability to learn volumetric information from limited datasets. The framework includes a 3D teacher network encoding volumetric knowledge, a 2D student network focusing on partial volumetric data, and a distillation loss to align feature embeddings between the two. Applied to Parkinson’s disease classification tasks using 123I-DaTscan SPECT and 18F-AV133 PET datasets, the method demonstrated superior performance, achieving a 98.30% F1 score. This projection-agnostic approach bridges the modality gap between 3D and 2D imaging, improving generalizability and addressing challenges in medical imaging analysis.

The method improves the representation of partial volumetric data by leveraging relational information, unlike prior approaches that rely on basic slice extraction or feature combinations without focusing on lesion analysis. We introduce a “partial input restriction” strategy to enhance 3D-to-2D KD. This involves projecting 3D volumetric data into 2D inputs via techniques like single slices, early fusion (channel-level concatenation), joint fusion (intermediate feature aggregation), and rank-pooling-based dynamic images. A 3D teacher network encodes volumetric knowledge using modified ResNet18, and a 2D student network, trained on partial projections, aligns with this knowledge through supervised learning and similarity-based feature alignment.

The study evaluated various 2D projection methods combined with 3D-to-2D KD for performance enhancement. Methods included single-slice inputs, adjacent slices (EF and JF setups), and rank-pooling techniques. Results showed consistent improvements with 3D-to-2D KD, with the JF-based FuseMe setup achieving the best performance, comparable to the 3D teacher model. External validation on the F18-AV133 PET dataset revealed the 2D student network, after KD, outperformed the 3D teacher model. Ablation studies highlighted the superior impact of feature-based loss (Lfg) over logits-based loss (Llg). The framework effectively improved volumetric feature understanding while addressing modality gaps.

In conclusion, the study contrasts the proposed 3D-to-2D KD approach with prior methods in neuroimaging classification, emphasizing its integration of 3D volumetric data. Unlike traditional 2D CNN-based systems, which transform volumetric data into 2D slices, the proposed method trains a 3D teacher network to distill knowledge into a 2D student network. This process reduces computational demands while leveraging volumetric insights for enhanced 2D modeling. The method proves robust across data modalities, as shown in SPECT and PET imaging. Experimental results highlight its ability to generalize from in-distribution to out-of-distribution tasks, significantly improving performance even with limited datasets.


Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. If you like our work, you will love our newsletter.. Don’t Forget to join our 55k+ ML SubReddit.

Evaluation of Large Language Model Vulnerabilities: A Comparative Analysis of Red Teaming Techniques’ Read the Full Report (Promoted)

The post Enhancing Deep Learning-Based Neuroimaging Classification with 3D-to-2D Knowledge Distillation appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

神经影像 知识蒸馏 3D CNN 2D CNN 医学影像
相关文章