MarkTechPost@AI 2024年07月16日
MIT Researchers Propose IF-COMP: A Scalable Solution for Uncertainty Estimation and Improved Calibration in Deep Learning Under Distribution Shifts
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

MIT、多伦多大学和向量研究所的研究人员提出了一种名为 IF-COMP 的方法,该方法是一种可扩展的解决方案,用于在深度学习模型中估计不确定性并提高校准精度,特别是在数据分布发生变化的情况下。IF-COMP 通过对模型进行线性化,并使用温度缩放的玻尔兹曼影响函数来近似 pNML 分布,从而解决了传统贝叶斯和 MDL 方法的计算挑战。实验结果表明,IF-COMP 在不确定性校准、错误标签检测和分布外检测等任务中,始终优于其他方法。

🤔 IF-COMP 是一种用于深度学习模型中不确定性估计和校准的新方法,它利用了温度缩放的玻尔兹曼影响函数来线性化模型,从而有效地近似 pNML 分布。该方法通过最小化模型和观察数据的总码长来避免对先验的明确定义,从而克服了传统贝叶斯方法的计算成本和先验定义的挑战。

💪 IF-COMP 在不确定性校准、错误标签检测和分布外检测等任务中表现出色,在 CIFAR-10 和 CIFAR-100 数据集上的不确定性校准中,IF-COMP 在各种腐蚀级别下都比贝叶斯和其他基于 NML 的方法实现了更低的预期校准误差 (ECE)。此外,IF-COMP 在错误标签检测方面也表现出色,在 CIFAR-10 上的人为噪声和 CIFAR-100 上的不对称噪声方面,其接收器工作特性曲线下面积 (AUROC) 分别为 96.86 和 95.21,优于 Trac-IN、EL2N 和 GraNd 等方法。

🚀 IF-COMP 在分布外检测任务中也取得了最先进的结果。在 MNIST 数据集上,IF-COMP 在远分布外数据集上获得了 99.97 的 AUROC,在 OpenOOD 基准测试中显著优于所有 20 种基线方法。在 CIFAR-10 上,IF-COMP 在远分布外数据集上获得了 95.63 的 AUROC,树立了新的标准。这些结果表明,IF-COMP 在提供校准后的不确定性估计和检测错误标签或分布外数据方面非常有效。

💡 IF-COMP 方法显著推动了深度神经网络的不确定性估计。通过使用温度缩放的玻尔兹曼影响函数有效地近似 pNML 分布,IF-COMP 解决了传统贝叶斯和 MDL 方法的计算挑战。该方法在各种任务中(包括不确定性校准、错误标签检测和分布外检测)的出色表现,突出了其在增强现实世界应用中机器学习模型的可靠性和安全性方面的潜力。这项研究表明,当有效地实施时,基于 MDL 的方法可以为深度学习中的不确定性估计提供稳健且可扩展的解决方案。

Machine learning, particularly deep neural networks, focuses on developing models that accurately predict outcomes and quantify the uncertainty associated with those predictions. This dual focus is especially important in high-stakes applications such as healthcare, medical imaging, and autonomous driving, where decisions based on model outputs can have profound implications. Accurate uncertainty estimation helps assess the risk associated with utilizing a model’s predictions, determining when to trust a model’s decision and when to override it, which is crucial for safe deployment in real-world scenarios.

This research addresses the primary issue of ensuring model reliability and proper calibration under distribution shifts. Traditional methods for uncertainty estimation in machine learning models often rely on Bayesian principles, which involve defining a prior distribution and sampling from a posterior distribution. However, these methods encounter significant challenges in modern deep learning due to the difficulty in specifying appropriate priors and the scalability issues inherent in Bayesian approaches. These limitations hinder the practical application of Bayesian methods in large-scale deep-learning models.

Current approaches to uncertainty estimation include various Bayesian methods and the Minimum Description Length (MDL) principle. Although theoretically sound, Bayesian methods require extensive computational resources and face challenges defining suitable priors for complex models. The MDL principle offers an alternative by minimizing the combined codelength of models and observed data, thereby avoiding the need for explicit priors. However, the practical implementation of MDL, particularly through the predictive normalized maximum likelihood (pNML) distribution, is computationally intensive. Calculating the pNML distribution involves optimizing a hindsight-optimal model for each possible label, which is infeasible for large-scale neural networks.

The Massachusetts Institute of Technology, University of Toronto, and Vector Institute for Artificial Intelligence research team introduced IF-COMP, a scalable and efficient approximation of the pNML distribution. This method leverages a temperature-scaled Boltzmann influence function to linearize the model, producing well-calibrated predictions and measuring complexity in labeled and unlabeled settings. The IF-COMP method regularizes the model’s response to additional data points by applying a proximal objective that penalizes movement in function and weight space. IF-COMP softens the local curvature by incorporating temperature scaling, allowing the model to accommodate low-probability labels better.

The IF-COMP method first defines a temperature-scaled proximal Bregman objective to reduce model overconfidence. This involves linearizing the model with a Boltzmann influence function, approximating the hindsight-optimal output distribution. The resulting complexity measure and associated pNML code enable the generation of calibrated output distributions and the estimation of stochastic complexity for both labeled and unlabeled data points. Experimental validation of IF-COMP was conducted on tasks such as uncertainty calibration, mislabel detection, and out-of-distribution (OOD) detection. In these tasks, IF-COMP consistently matched or outperformed strong baseline methods.

Performance evaluation of IF-COMP revealed significant improvements over existing methods. For example, in uncertainty calibration on CIFAR-10 and CIFAR-100 datasets, IF-COMP achieved lower expected calibration error (ECE) across various corruption levels than Bayesian and other NML-based methods. Specifically, IF-COMP provided a 7-15 times speedup in computational efficiency compared to ACNML. In mislabel detection, IF-COMP demonstrated strong performance with an area under the receiver operating characteristic curve (AUROC) of 96.86 for human noise on CIFAR-10 and 95.21 for asymmetric noise on CIFAR-100, outperforming methods like Trac-IN, EL2N, and GraNd.

IF-COMP achieved state-of-the-art results in OOD detection tasks. On the MNIST dataset, IF-COMP attained an AUROC of 99.97 for far-OOD datasets, significantly outperforming all 20 baseline methods in the OpenOOD benchmark. On CIFAR-10, IF-COMP set a new standard with an AUROC of 95.63 for far-OOD datasets. These results underscore IF-COMP’s effectiveness in providing calibrated uncertainty estimates and detecting mislabeled or OOD data.

In conclusion, the IF-COMP method significantly advances uncertainty estimation for deep neural networks. By efficiently approximating the pNML distribution using a temperature-scaled Boltzmann influence function, IF-COMP addresses the computational challenges of traditional Bayesian and MDL approaches. The method’s strong performance across various tasks, including uncertainty calibration, mislabel detection, and OOD detection, highlights its potential for enhancing the reliability and safety of machine learning models in real-world applications. The research demonstrates that MDL-based approaches, when implemented effectively, can provide robust and scalable solutions for uncertainty estimation in deep learning.


Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter

Join our Telegram Channel and LinkedIn Group.

If you like our work, you will love our newsletter..

Don’t Forget to join our 46k+ ML SubReddit

The post MIT Researchers Propose IF-COMP: A Scalable Solution for Uncertainty Estimation and Improved Calibration in Deep Learning Under Distribution Shifts appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

IF-COMP 深度学习 不确定性估计 校准 分布偏移
相关文章