Published on May 24, 2025 9:24 PM GMT
I have noticed that there are talks around about moral ASI. And I think that to use the word "moral" in relation to Artificial Intelligence, we must be absolutely confident in our knowledge of how morality works for human beings. Otherwise, we must avoid using such combinations of words to avoid anthropomorphic bias.
The question of morality in general was already discussed on LW, e.g. Morality is Scary. I just wanted to emphasize that, considering existential risks involved on one hand, and some taboo on the other, you may require a certain level of cynicism and readiness to consider the most controversial theories of moral relativism.
And I would like to present you with an example of such a hypothesis. You may disapprove and disagree with the hypothesis itself, but that shouldn't cancel what I've said above.
Hypothesis: Moral patience depends on the capacity of being perceived as a potential threat.
Moral patience isn't a discrete property and may be related to certain aspects. It is also closely related to the empathy.
For example, I was always amused by how owners love their pets, but don't hesitate to castrate them. A cat is being deprived of his capacity to procreate, the only reason for his existence, and those who did this to him keep using him for their amusement. When I imagined myself in this cat's place, I wondered if I would attack them in their sleep.
I didn't have pets since childhood and such my thoughts are anthropomorphic biased. Perhaps that's why I feel castration of pets as immoral despite all the reasoning.
Pet owners, of course, don't see it this way. They know that pets don't think about procreation and only follow their instincts. And castration helps to adjust those instincts without affecting their happiness.
But many owners feel it is immoral to torture their pets. Maybe because they unconsciously feel a threat of possible revenge. Some dog owners are used to training their dogs and have learned how submissive they can be. The empathy decreases and is being replaced with utility.
The same applies to animal husbandry and even torturing people. Feeling of impunity leads to the disappearance of empathy. I wanted to bring many more examples from the history supporting the claim, but I suppose that's enough for a hypothesis, which I didn't intend proving in the first place.
And so if this hypothesis were valid and someone wanted to implement exactly the same mechanism for ASI having empathy to humans, they would have to pose humans as a threat, which doesn't feel as a safe approach to control superior intelligence.
- ^
I'm not sure if it is correct to put links to a someone's quick take.
Discuss