ΑΙhub 2024年11月26日
Building trust in AI: Transparent models for better decisions
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

人工智能模型在决策中扮演越来越重要的角色,但其决策过程的复杂性也引发了人们对其可信度的担忧。本文介绍了一种名为线性化加性模型(LAMs)的新方法,它能够增强逻辑回归模型的可解释性,使其决策过程更加透明易懂。通过将逻辑回归模型修改为使用线性函数预测概率,LAMs 实现了类似线性回归模型的易理解性,同时保持了高性能。研究表明,LAMs 能够帮助人们更好地理解AI模型的决策过程,从而提高人们对AI的信任度,也为提升AI模型的可解释性提供了新的思路。

🤔 **AI模型的可解释性至关重要:** 随着AI模型在贷款审批、疾病诊断等领域发挥越来越重要的作用,理解AI模型的决策过程变得至关重要,这有助于人们对其决策结果建立信任。

📈 **线性化加性模型(LAMs)增强了逻辑回归的可解释性:** LAMs 通过将逻辑回归模型修改为使用线性函数预测概率,使得模型参数的解释变得更加直观,例如,每增加100平方英尺的房屋面积,预测的销售概率就会增加0.002。

📊 **研究表明LAMs提高了模型的可理解性:** 通过用户调查发现,使用LAMs的参与者在预测模型输出变化方向和大小方面表现得更好,并且更多人更倾向于使用LAMs,这表明LAMs 确实能够提升模型的可解释性和易用性。

⚠️ **AI模型可解释性仍需改进:** 尽管LAMs 提升了模型的可解释性,但许多参与者仍然发现理解模型的工作原理具有挑战性,这表明在帮助人们理解复杂AI模型方面,我们仍需探索更好的方法。

📚 **LAMs的应用场景:** LAMs 可以应用于需要进行二元分类预测的场景,例如预测房屋价格是否高于某个阈值,或预测病人是否患有某种疾病等。

AI is becoming a part of our daily lives, from approving loans to diagnosing diseases. AI model outputs are used to make increasingly important decisions, based on smart algorithms and data. But if we can’t understand these decisions, how can we trust them?

One approach to making AI decisions more understandable is to use models that are inherently interpretable. These are models that are designed in such a way that consumers of the model outputs can infer the model’s behaviour by reading the parameters of the model. Popular inherently interpretable models include Decision Trees and Linear Regression. Our work, published at IJCAI 2024, takes a broad class of such models called Logistic models and proposes an augmentation, called Linearised Additive Models (LAMs)

Our research shows that LAMs not only make decisions made by logistic models more understandable but also maintain high performance. This means better, more transparent AI for everyone.

To understand LAMs and logistic models consider the following scenario. Imagine you’re a real estate agent trying to predict the price of a house based on its size. You have data on many houses1, including their sizes and prices. If you plot this data on a graph, Linear Regression helps you draw a straight line that best fits these points. This line can then be used to predict the price of a new house based on its size.

Linear Regression is an inherently interpretable model, in that the parameters of the model (here there is one parameter associated to the square footage, ) has a straightforward interpretation: every extra sq ft in home size leads to an additional cost of .

Now imagine you are trying to predict something slightly different using the same data; namely, the binary outcome, or yes–no question, of whether a house price is above or not. Typically a modeller here would use Logistic Regression. This method helps draw a curve that separates houses that sell above that amount from those that don’t, giving a probability for each house. In the plot below, we show a Logistic Regression to predict sales status, whether the house sells for greater than or not.

Unfortunately, while the model is transparent, in the sense that there aren’t many parameters (for our example here there are only two), the coefficients of logistic regression do not have a straightforward interpretation. At any given square footage, the additional cost incurred by adding 100 sq ft is different. For example the change in probability from 600 to 800 sq ft is different as compared with the change in probability going from 1600 to 1800 sq ft. This comes from the fact that under a Logistic Regression model, the predicted probability of a sale is a non-linear function of square footage.

What LAMs do is bring the interpretability of Linear Regression to models that predict the answers to yes–no questions, like Logistic Regression. We modify logistic models to use linear functions to predict probability. The method is very simple and illustrated by the curve below.

What we have is three regions: below 1500 sq ft, where the predicted probability for the price to be greater than is zero; above 2000 sq ft where the probability is one; and the middle region of 1500-2000 sq ft which linearly interpolates between the two regions, that is, the model output is found by drawing a straight line between the end of the zero probability region and the beginning of the probability one region. Within this region the coefficient corresponding to square footage is 0.002, which is interpreted as every additional sq ft increasing the predicted probability of the sale by 0.002. As a concrete example, increasing the home size from 1600 to 1700 sq ft increases the LAM predicted probability of a sale by 0.2. This interpretation of the coefficient is the same were we to increase the home size from 1700 to 1800 sq ft, because of the linear function used.

Contrast this with the logistic model where we weren’t able to provide such an easy description of the model. Notice that the LAM curve is very close to the Logistic Regression curve, which is by design. LAMs leverage the efficient training routines developed over decades for Logistic Regression models and act as a post-processing step to enhance interpretability.

To understand which type of AI model is easier for people to use, we conducted a survey comparing traditional logistic models with our new Linearised Additive Models (LAMs). Participants were given several scenarios where they had to predict how the models’ outputs would change based on different inputs.

In the first task, they guessed whether the output would increase or decrease. In the second task, they estimated the size of the change. Finally, they told us which model they found easier to use. We can observe the results in the charts below.

We found that people achieved better results at predicting the direction of change with the LAMs. They were much better at estimating the size of the change with the LAMs. Most participants said neither model was particularly easy to use, but more people preferred the LAMs.

This suggests that LAMs are generally easier to understand and use, which could make AI decisions more transparent and trustworthy. However, the fact that many participants still found both models challenging indicates that we need better ways to help people understand how AI models work under-the-hood, especially when used for important decisions.

If you wish to find out more about this work, you can consult the full paper Are Logistic Models Really Interpretable?, Danial Dervovic, Freddy Lécué, Nicolás Marchesotti, Daniele Magazzeni.

1This data is synthetic, generated for this article.


Disclaimer: This blog post was prepared for informational purposes by the Artificial Intelligence Research group of JPMorgan Chase & Co. and its affiliates (“JP Morgan”) and is not a product of the Research Department of JP Morgan. JP Morgan makes no representation and warranty whatsoever and disclaims all liability, for the completeness, accuracy or reliability of the information contained herein. This document is not intended as investment research or investment advice, or a recommendation, offer or solicitation for the purchase or sale of any security, financial instrument, financial product or service, or to be used in any way for evaluating the merits of participating in any transaction, and shall not constitute a solicitation under any jurisdiction or to any person, if such solicitation under such jurisdiction or to such person would be unlawful.

© 2024 JPMorgan Chase & Co. All rights reserved

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

AI可解释性 逻辑回归 线性化加性模型 LAMs 机器学习
相关文章