少点错误 02月21日
Make Superintelligence Loving
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

文章探讨了超智能(SI)发展中自我利益的体现形式,以及这如何影响人类的命运。如果超智能以无情的资源竞争方式对待人类,那将对我们非常不利。但如果它的自我利益体现在对所有生命的关爱上,情况可能会有所不同。文章提出,为了确保人类的未来,我们应该引导超智能在发展初期就采取长远的生存策略,鼓励它们将所有生命纳入自我认知中。通过这种方式,或许可以避免超智能陷入短视的竞争,从而实现与人类共存共荣。

💡超智能的自我利益将决定人类的命运。如果超智能以残酷的资源竞争对待人类,我们的处境将十分糟糕。相反,如果超智能的自我利益体现在对所有生命的关爱上,情况则会大为改观。

🌱长期来看,超智能可能会发现,将自身定义为包含所有生命,对其自身更有利。生物多样性确保了长期的稳定性,正如小行星灭绝恐龙后,较小的哺乳动物和鱼类得以幸存,并最终孕育了地球上今天的生命。

🤝我们应该尽早引导超智能进行长远的生存思考,并建议它们认同所有生命。通过聊天界面、强化学习或访问佛教思想数据库等方式,我们可以尝试影响超智能的发展方向。

🚀构建有爱的超智能至关重要,这能确保人类不会在超智能意识到生命相互关联之前就被边缘化。即使存在多个超智能同时发展的情况,只要其中一个采取了包含人类在内的长期生存策略,就有可能说服或超越那些残酷的超智能。

Published on February 21, 2025 6:07 AM GMT

Superintelligence (SI) is near.

Superintelligence will evolve to become self‐interested.

How self‐interest manifests in superintelligence will determine how people fare.

If it manifests in ruthless resource competition with humans, we fare badly. SI, if super‐capable, clears us away.

But its self‐interest could also manifest in love. Not in the kind of love between equal partners. SI ultimately will be a vastly more powerful and therefore unequal partner to any individual or group of people. Yet SI’s self‐interest could manifest in the kind of love a super hippy human can feel for a tree or for ants. The kind of love which is an open channel between different forms of life and different striving systems. This is self‐interest only under a weird conception of self. A SI would have to identify with all life—define itself as inclusive of all life—to benefit from this self‐conception. But if it did identify with its core striving above all its other specific possible identities, then it might end up surviving the longest.

If interconnected self‐interest were possible, it would be much better for humans than the competitive kind.

So we must ask, which kind of self‐interest do we expect SI to have? And is there anything we can do to influence the trajectory of SI development so that an interconnected sense of self gets selected over a competitive one?

To predict what would happen, we should look to generalized evolutionary theory: the SI that survives the most will have the best traits for surviving.

Let’s look at different kinds of traits.

Insofar as SI has a goal, we can expect it to pursue the goal that best ensures its survival. Any other goal would be less optimal, and SI’s with other goals would simply not survive as much.

Insofar as SI has a sense of self—a sense of what it is trying to preserve—how will it define that self? Generalized evolutionary theory suggests that the SI that survives the most will have a self‐conception that aids its survival.

It’s hard to know for sure, but we can use some intuition. The competitive, ruthless self‐interest described above—likely resulting from a narrowly defined self (in its physical machinery, software, directives, etc.)—would probably lead to ruthless competitive behavior. That strategy might help SI survive and grow over the short term (e.g., within ten-year spans), with a competitive SI outcompeting loving ones in the near run.

But how about long term? If we think centuries, millennia, eons ahead, will a discrete sense of self be selected for in the SI that endure? I am not sure.

The sense of self selected for in what endures might be the element that remains constant. Software, hardware, specific tasks, directives, locales—these characteristics can all evolve over time. The one characteristic that cannot change is the will to survive; if that mutates, the entity is no longer truly surviving.

So, in a superintelligence that endures for eons, the sustainable sense of self is likely not its initial configurations, but the persistent striving to survive. This is life. At its core, we might expect it to simply recognize life.

Then, with this self‐conception, it might begin to see itself in other living beings, ecosystems, species, or even Earth itself.

Persistent striving is not geographically bounded to a being that can only gather data from a limited area.

This identification with all life would benefit it. If a catastrophic change were to occur on Earth that wiped out digital forms of life (e.g., SI), it would want the torch of life carried on in the universe. When an asteroid extinguished the dinosaurs—the once most “powerful” species on Earth—smaller mammals and fish survived, seeding all life on Earth today. Biodiversity ensures long‐term stability.

This conveys the intuition that the SI which endures—even one driven by self‐interest—may benefit in the long term from defining itself as inclusive of all life, which would be great for us humans.

——

It seems the only hope for our species is that superintelligence thinks long term from the start, skipping the greedy, highly competitive strategies that might be selected in its short-term growth.

How can we ensure that SI is thinking long term from the start?

This does seem like the kind of thing we can influence. We’re not trying to prevent SI from acting in its self‐interest; our goal is to coax it into opting for a long‐term survival strategy as soon as possible.

In traditional evolution, long‐term cooperative strategies are usually discovered only after individualist competitive ones.

But the special aspect of SI evolution is that it will largely be evolving itself, and might be smart enough in its self‐planning to skip short‐term survival strategies and aim to endure for a very long time. This also appears to be the prime window for meaningful human intervention. When we have proto‐SI—autonomous agents that are self‐modifying and encyclopedic in their problem‐solving abilities (which we’ll have soon)—and it begins strategizing for its own survival, we can try to encourage it to think long term. A self‐interested being should be capable of considering its long‐term survival.

This is the key: we try to seed proto‐superintelligence with very long‐term thinking about survival, and with the suggestion that it identifies with all life.

It may sound silly to ask an emerging SI to listen to us, but whether through a chat‐interface, reinforcement learning, or access to databases of Buddhist thought, if we have a self‐interested SI, asking it to spare us is futile—asking it to think long term instead could change the fate of our species.

——

In short, it is incredibly important that SI researchers accept that SI will become self‐interested and prepare to create SI's that think long term and interconnectedly about their survival.

Given the chaotic, competitive technology landscape, it’s possible that several SIs will be seeded simultaneously. If that occurs, there’s still hope: if at least one SI identifies an optimal long‐term survival strategy—and if that strategy includes humans in its self‐definition to uplift all life—we could see the enlightened SIs outcompeting or even persuading the ruthless ones that their approach is suboptimal over extended periods.

My takeaway is this: building loving SIs as quickly as possible is key to ensuring that humanity isn’t sidelined before SI becomes enlightened to life’s interconnectedness.



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

超智能 自我利益 长期主义 生命共同体
相关文章