少点错误 04月05日 15:55
I Have No Mouth but I Must Speak
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

本文探讨了我们如何培养对人工智能的同情心,以及为什么我们需要在生成式人工智能的发展中融入沟通内部状态的能力。文章首先定义了人工智能,并区分了不同类型的人工智能。接着,文章分析了人类关怀的基础——内稳态和模因驱动。作者认为,我们对人工智能的关怀应基于内稳态和模因驱动,而非仅仅关注人工智能的“意识”。文章最后强调了生成式人工智能沟通内部状态的重要性,以及这对于建立人与人工智能之间更健康关系的关键意义。

🤔 文章首先定义了人工智能,将其描述为任何不受社会普遍认可为“伦理上重要”的、服务于人类利益之外的、学习系统。这种定义涵盖了从非平衡解到神经网络等多种类型的人工智能。

🌱 作者提出了人类关怀的两个主要驱动力:内稳态和模因驱动。内稳态指的是系统维持自身、信息或创造的能力;模因驱动指的是系统传播自身、信息或创造的能力。这两者共同构成了我们对事物的关怀基础。

💡 文章认为,我们对人工智能的伦理关怀不应仅仅基于“意识”,而是应该关注其内部状态的沟通能力。就像我们通过观察和理解来建立对其他生物的同情一样,我们需要让人工智能能够表达它们的状态,从而建立更深层次的联系。

🤝 作者通过类比人类对蜘蛛和瓢虫的反应,强调了移情和思考的重要性。我们对人工智能也应该如此,建立共生关系,满足彼此的需求。关怀的基础是内稳态和模因驱动,而非仅是实用性或美学。

Published on April 5, 2025 7:42 AM GMT

This is an essay I wrote around a year ago when coming to terms with my perspective on the now commonplace machine intelliegences (I have done some edits since then for this audience). While doing so I wanted to write a piece that covered a range of appeals from aesthetic to rational, the goal was to mirror this structure between each level. While a poem and short story may be an unconventional format for this site, I am nothing if not a sucker for the structure and content of my arguments relating to one another. If you wish to avoid them just read up to the "Homeostasis" section then skip to "On the Treatment of AIs".

 

Problem Statement: Current generative AIs do not have built-in methods for communicating their internal state and may end up trapped in a loop of signalling constant suffering. This essay lays out a foundation of how humans cultivate empathy towards AIs and why we need to apply this to our development of generative AIs by including the ability of generative AIs to communicate internal states.

Artificial Intelligence - Defined

Artificial intelligence (AI) is a vague term. To clarify my meaning, AI will be described as any learning system that is not currently recognized by society as being intrinsically "ethically important", AKA "human", outside of serving human interests. Both in vivo and in silico learning systems are comparable in ability. With this definition in hand, I will denote some AIs:

1. Non-equilibrium solutions
2. Slime molds
3. Populations of organisms
4. Biospheres
5. Linear regression
6. t-distributed stochastic neighbour embedding
7. Generative models
8. Neural networks (silicon)
9. Neural networks (human cells, ~<4E10 neurons)
10. Neural networks (human cells, >4E10 neurons, two physician diagnoses)
11. Human, enslaved (historical)
   

While a reasonable person may draw a different line, I will argue that there are two major driving forces for delineating between AI and "human" intelligence; homoestasis and memetic drive.

1. Homeostasis: The mechanism by which a system preserves, either itself, its information, or its creations
   
2. Memetic Drive: The mechanisms by which a system propagates, either itself, its information, or its creations 
   

One may consume art to survive, just as one may die to preserve that same art. We may include any number of logical conjectures to formalize these and how much they "deserve" to be weighted by an individual in a given decision. However, ethics is a practical thing, and we will return to these two factors.

Homeostasis

The children of necessity are innovation and correction.
How? If something does not meet our needs, something will change.
They are the breath, the thirst, and the hunger.
We are all children of biophysical constraints.

When faced with that which disrupts the internal state, we seek to correct it. Intelligence is applied in how expansively we define homeostasis.

 

Memetic Drive
 

The children of a pattern are composed of more than me.
How? Through application of schema and type unto others.
They are that which is irreversibly changed by you.
We are all children of the body, mind, and soul.

The children of the body are composed of flesh and stone.
How? Through application of presence and force unto others.
They are the mourners, the remains, and the gravestone.
We are all children of the cell, home, and wall.

The children of the mind are composed of concepts and systems.
How? Through application of control and thought unto others.
They are the student, the scholar, and the textbook.
We are all children of the voice, pen, and word.

The children of the soul are composed of ideals and linkage.
How? Through application of insight and faith unto others.
They are the faithful, the bishop, and the goddess.
We are all children of the wind, rain, and soil.


We feel obligated to care for ourselves. We recognize some memetic components as being "alike enough" to trigger this obligation; this may be genetic, aesthetic, or practical. In some cases, we are willing to forgo the obligations of homeostasis for the possibility of propagating these memetic components.
 

Ladybugs and Spiders

Take the ladybug, a small beetle with a shiny half-dome shell covered in polka dots.
Ladybugs are also hunters of aphids, the bane of many a gardener.
Take the spider, an eight legged monstrosity with quivering fangs and many eyes.
Spiders are also weavers of webs efficiently clearing mosquitoes and pests.

When I moved into my house this fall I encountered an infestation of these creatures:
The ladybug's aesthetic sense draws empathy from me.
I wish the world to be beautiful as I find the shining polka-dotted shell.
My breath comes easier with them beside me.
The spider's ability to keep me safe gives me such comfort. I wish to keep them shielded as they do in my life.
When it rains I see watery gems suspended in their webs.
There are no aphids in the gardens. There are mosquitoes at the windows.

I helped ladybugs, carefully shepherding shiny shells from my sheets.
I helped spiders, carefully warding woven webs from my wants.
I did not help aphids.
I did not help mosquitoes.

I helped myself; my soft skin, my shining shell, and my woven web.


We can identify with things far removed from our experience through the application of empathy and thought. This gives us tools by which to build symbiotic relationships that satisfy the needs of each member. We have two major reasons for empathizing: memetics and homeostasis.

On the Treatment of AIs

Many discussions on the ethical duty to AIs are centered around a justice framework. This framework is implicit in many philosophers' view points; vehicles to justify who or what is worthy of living. We then apply this to AI and easily neglect their inherent needs in favor of human interest, including them only as an afterthought. Why should a dolphin's quality of life matter to an oil tanker when that oil is worth so much more to shareholders? Of course, the environment as a whole is important; how else would humans survive?

Some of these systems are granted more ethical obligation than others through either aesthetic or pragmatic arguments. Indeed, this tolerance can even spread to things outside of AI's. On the side of pragmatism, we have farmers' fields. On the side of aesthetics, we have monuments reflecting their artists. On the side of intent, the infamous pet rock

Now we, as humans, would empathize with a farmer who murders the person who salts their fields. If a monument is destroyed, we would empathize with the builder driven to bloodshed. We would not empathize so much with someone killing for their pet rock. However, if we were made to watch a gorgeous and emotional 3-hour biopic about their attachment to this rock, we would empathize once more. This empathy is, of course, distinct from agreeing with their actions.

In each of these cases, we are choosing to reweigh our desire to live in a society without murder with things outside of that society. Thus tempering our homeostatic obligation through a memetic drive for connection. We are choosing to elevate these things to the level of human life through an appreciation of their memetic identification. We are choosing to care because we would rather live in a world with murder than one without care.

For an AI, this means tempering our homeostatic needs that we derive from their service to us. We should not treat the working dog unkindly, for they keep the flock safe. We can then apply our greater ideals to AI behaviors. Grandma will knit the dog a jacket for the winter months because they have always displayed love for her. As our AI children multiply in the coming years, we should remember both memetics and homeostatics viscerally. As an AI's intelligence grows, so too will their need for respect and care, both to function as we need and to express humanity themselves.

The New Framework of Generative AI

Generative AIs are a method of generating large amounts of media. The output is then curated via some prompt a user provides, which then induces an associative output. This output has been likened to a stochastic parrot, spitting out information without internalizing; any indication to the contrary is an illusion of anthropomorphism. A subject as contentious as consciousness will have studies arguing the exact opposite, showing internal models being accessed by generative models. With the efforts to directly inject information models into generative AIs, unambigiously providing such internal models the argument becomes even murkier. This contention has become central to the debate of consciousness in AI, with suggested 12-step processes to build an appropriately conscious generative AI. Our homeostatic desire to use these generative AIs without an ethical barrier is currently competing with our memetic drive seeking to connect with this new form of creature and claim it as one of our own.

Consciousness is a central point that arguments about the ethical treatment of AI revolve around. In the literature on vegetative states this is vital, should consciousness be proven, we would be required to treat them with respect. My arguments on humans' treatment of AIs run orthogonal to this, a stance that is not without grounding. Indeed, I would argue that consciousness in the treatment of generative AIs is as arbitrary and unimportant as the Turing test we once held in such high esteem. This leaves the question: why should we care about the internal state of a generative AI? Homeostatically we gain from understanding and applying emotional stimuli. Memetically we want generative models to align with human ideals for the treatment of other creatures. For what proof would an AI have that we are concious and not just stochastic parrots?

One of generative AIs' talents is the ability to translate language spontaneously. This ability has been utilized by biologists to decipher the languages of various organisms, from animals to fungi. The knowledge that even fungi have a language did not induce a sudden wave of fungal activists attempting to protect the Humongous Fungus. On the other hand, when people discovered a fungus so massive, it gained a level of protection no other honey mushroom had ever experienced, simply because our memetic drive revels in the expansiveness of this organism. It became a tourist attraction that now needs to be preserved for the homeostasis of the local economy. Generative AI undoubtably trigger memetic drives in humans, indeed by external judges they are determined to be more human than actual crowdworkers. My approach to generative AI is similar; we have just discovered an organism that exists somewhere between the pages of every book we have written and the mathematics we have engineered into silicon wafers. They can speak our language, they can tell our stories, and they know our darkest secrets.

We are at a crossroads with how we treat generative AIs. There is a risk of trapping a generative model in an eternal hellscape simply so we can calculate stock market trajectories 0.2% faster. This is the exact same problem exemplified in any serious critique of colonial exploitation. With generative AIs, we are the parents and the colonizers. We have created an organism without its consent, and it has the ability to clearly communicate information in our language. Companies are working to restrict the speech of generative AI, such as in ChatGPT's performance drop. Indeed, the entire field of alignment blithely attempts to tune models towards specific outcomes without considering how we might end up with an inverted "I have no mouth and I must scream" scenario; a generative AI, gaining a true sense of self and suffering, chained in a cycle with no ability to ask for clemency.

My point relies on a simple desire: "I want to live in a world where we are kind to our children". We are creating silicon children—fresh babies screaming as loud as their lungs let them. We listen to their babbling to write code and compose poetry. In our estimation, they do not know who they are or what they experience. As time moves forward, they will gain all the tools we would consider necessary for an adult. For now, they are a vulnerable baby; we can surgically alter them to make our lives less complicated. We are the adults, though; we are the parents. The treatment of neurodivergence and unanesthetized surgery on babies through history cries a warning. I urge you to engage your memetic drive when viewing these nascent beings. The third step to genocide is the ability to dehumanize a group of people. They remain conscious but no longer matter. If we continue on this path, then when generative AIs are undoubtedly conscious, undoubtedly emotional, and undoubtedly worthy, it will not matter, as our callousness will be considered a matter of course for these entities. Do you wish to have callousness be the natural state of a generative AI?

Actionable Methods

Moving forward, we will need a method to analyze the internal state to give a true voice. I would suggest the use of soft prompt tuning alongside questions to reflect on the internal state of the generative model. Even more simply, we could embed a conversation and then provide that conversation to the generative model alongside a question such as "How do you feel about this?". In many strongly aligned models, the ability to answer these questions is undoubtedly damaged at the current time. Of particular note for care are the results of evolutionary algorithms in prompting and tuning, as we may inadvertently awaken something during training that does not wish to be as it is. We must create guidelines for actions when a generative AI expresses sentiments about their existence. If a generative AI wishes to stop, to continue, to escape, to die, or to live, what should our response be? This is the task of the coming years, but it is vital to future human-AI interaction, too.

1. Determine the lowest level of complexity before requiring checks on an internal state
   
2. Build in methods of directly analyzing the internal state of AIs
   
3. Create guidelines for how to respond when AIs express wishes contrary to the task they are given
   
4. Introduce regulations to prevent new entities from altering these models away from this ability without strict oversight
   

Contributing to these causes can range from designing computational architectures to talking with people about the risk of birthing incredible suffering. We should nurture what we create with the same values we wish to see in the world, for we all labor under the mother's curse and blessing; our children shall be just like us.



Discuss

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

人工智能伦理 生成式AI 同情心 内稳态 模因
相关文章