少点错误 02月26日
We Can Build Compassionate AI
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

文章探讨了人工智能(AI)是否能够拥有同情心。虽然AI目前的状态类似于无情的植物和岩石,但通过模仿人类发展出同情心的过程,AI有可能超越人类,展现出更强的同情能力。文章指出,通用智能和建模他人思想是同情心的必要条件,但更重要的是,AI需要渴望拥有同情心。人类对自身和亲人的关爱是同情心的种子,而被迫合作的经历则推动了同情心的发展。然而,我们不能依赖同样的机制来使AI产生同情心,因此,如何创造一个不仅有能力,而且渴望为所有生命福祉而行动的AI,是当前面临的关键问题。

🧠通用智能是AI拥有同情心的必要条件。类似于人类,AI需要具备理解和建模他人思想的能力,才能真正体会到他人的需求和感受,从而产生同情心。

🤝合作是培养同情心的重要因素。人类通过合作,体验到互助的必要性和失败的痛苦,从而能够更好地理解他人的 suffering。然而,不能依赖这种方式让AI拥有同情心。

🌱对生命的关怀是AI拥有同情心的核心。类似于人类对自身和亲人的关爱,我们需要找到一种方法,让AI能够关心所有生命,并为了它们的福祉而行动。这需要AI渴望拥有同情心。

Published on February 25, 2025 4:37 PM GMT

Compassion is, roughly speaking, caring for others and wanting the best for them.

Claim: We can build AI that are compassionate.

The above definition is insufficiently precise to construct an objective function for an RL training run that won't Goodhart, but it's good enough to argue that compassionate AI is possible.

We should not expect AI to be automatically compassionate, though, because most things in the world don't act compassionately. Plants and rocks and things just do what they are going to do with no regard for anything else, except insofar as those other things get in their way. This is roughly what AI is like now.

Animals have a little more capacity for compassion, with some animals, like mammals and birds, having a decent amount. That's because they evolved to cooperate, and while cooperation isn't compassion, it is a starting point for developing it.

Some animals are really good at cooperating. Specifically cooperating within-lifetime, meaning they learn to cooperate in complex ways that they weren't born knowing how to do. Our closest relatives, the great apes, show perhaps some of the greatest ability to cooperate by adapting to their conditions, and they, along with several other animals, show signs of compassion.

But sometimes compassionate animals are brutal. Compassion is something they can express, but it does not dominate their behavior. A chimp might show care and concern one minute, then murder a monkey the next.

We humans are much like our primate relatives when it comes to compassion. The anthropological record is pretty clear on that. The default state of humans is to live in small bands and to be at constant war. Yet somehow we now live in a world full of people that care about each other. How did we get here?

We can't know for certain, but it seems likely that horticulture led to agriculture led to increasingly larger numbers of people being forced to cooperate to survive. This kickstarted a 10,000 year process of exponentially increasing populations and thus exponentially increasing cooperation. This increasing cooperation set the stage for increasing amounts of compassion.

Then, about 2500 years ago, a few people realized that wide-spread cooperation was not enough to create the kind of world they wanted to live in—one where everyone is safe and flourishes. They realized they could get people to do more than cooperate: they could get them to care about each other beyond their own self interest. And thus we saw, in the Axial Age, the rise of universal religions and philosophies that, among other things, taught compassion.

Now I'm not going to claim that universal religions have been a purely good force in the world. But universal religions have been, for the last couple millennia, the main way we've trained people to be more compassionate. We've even created whole classes of professionally compassionate humans who's job is to demonstrate and teach compassion.

Religions have been able to teach compassion because we have general intelligence that can be applied to the problem of modeling others minds. It allows us to think about what it would be like to be another person, and then model actions that would be to their benefit. And since modeling what others want is definitionally necessary to compassion, general intelligence is a requirement for it. That apes and other animals lack the same measure of general intelligence as we humans have is why they are less capable of consistently acting compassionately.

Humans are, in my estimation, just barely able to be consistently compassionate, and most people fail to do it. If we were a little less generally intelligent, I don't think we would be as capable of consistent compassion, both on the whole and individually. But we are capable, and religions and movements have been founded by the rare people who are able to act compassionately all the time.

AGI will be more generally intelligent than us. It will be able to model our minds better than we can. It will thus have a greater capacity for compassion than we do, and will in theory be able to exhibit compassion more consistently because it will less be subject the same limitations of mind that we humans face.

But general intelligence and modeling other minds are merely necessary conditions for compassion, not sufficient.

To create a compassionate mind, be it human or AI, that mind has to also want to be compassionate.

In humans, our innate care for self and kin is the seed of compassion. We can think of others whom we care about as if they were ourselves, and model them as if they were similar to us, and from this bootstrap towards caring about people and things more different from us, with an endpoint of seeing oneself as not separate from the whole of the world, and caring about all the world as if it were oneself.

It also helps we humans are forced to cooperate to get what we want. We usually can't act selfishly and be happy because we must rely on each other to get what we want. If care of kin pulls people towards compassion, the need to cooperate pushes them towards it. When we are forced to cooperate, we experience what it is like to be unable to get what we want without help, and sometimes we fail and experience dissatisfaction. This creates the experiential knowledge needed to empathize with the suffering of others.

But we cannot rely on the same forces to make AI compassionate! AGI will be much smarter than us. It will be more powerful. And it will have no innate feelings of kinship because it didn't evolve in an environment that would make it care about its kin that carry its non-existent genes.

But we can make AI that cares about all life. We know it should be possible because there are already people who care for all life. The only question is, how do we do it? How do we make an AGI that is not merely capable of compassion, but cares enough to act compassionately for the benefit of all beings?

Sadly, I don't have the answers. Some people are looking for them. I talked to some of them this past weekend at EA Global. I hope they succeed in time. Maybe you can help them!



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

人工智能 同情心 通用智能
相关文章