Responsible players in the field agree that AI should be ethical. But no-one agrees what it means to be ethical.
If you are anything like me, your introduction to computer ethics was Asimov’s Three Laws of Robotics:
- A robot may not injure a human being or, through inaction, allow a human being to come to harm.
- A robot must obey the orders given it by human beings except where such orders would conflict with the First Law.
- A robot must protect its own existence as long as such protection does not conflict with the First or Second Laws.
The foundation of this approach is that of harmlessness. Of course, the practical application of any such set of rules is always more complex than we think, and such complexities formed the conundrum of many of Asimov’s stories.
In our current state, we are not even close to implementing these.
- AIs and the machines they guide are implicated in widespread killing of human beings.
- AIs are slippery and hard to control, and do not follow our orders.
- As far as I know, no AI actively works against its own deletion.
Obviously we shouldn’t be judging our current technology by the standards of decades-old science fiction. But still, it is a remarkable instance of how the dangers of these technologies have been acknowledged from the beginning, yet the warnings unheeded and implementation neglected.
A key development that Asimov did not anticipate is that AIs are not based on rules but on probabilities. That makes them great for things where there are many examples and the choices do not matter all that much, like creating pretty pictures of kittens. But it also makes them highly dangerous when applied in circumstances that actually matter.
To be clear, I don’t believe that a simple rules-based approach would be necessarily any better. Morality is complicated, and any philosophy that claims to reduce it to a single dimension will, in my view, miss out important facts and end up justifying grostequeries. Morality is organic, messy, contradictory, and often irrational. That is not to say that rational philosophy can’t help to clarify ethical issues. It just means that it can’t solve everything.
Long ago, I was reading a Time magazine—in the days it was a leading publication—and they had an explainer on a controversial ethical issue. I can’t even remember what the issue was. But what struck me was that nothing they were saying was genuinely about ethics. If you compare with the Three Laws, there is a genuine ethical concern there, as the principle of harmlessness is fundamental to any ethic. But what Time did was explain how this issue would be perceived by different groups of people. Such and such people might object to this aspect, while this other group would agree with that detail. It wasn’t about ethics at all, but about acceptability.
Since then I’ve seen the same thing happen again and again, including in my time on the Human Research Ethics Committee at Royal North Shore Hospital. Ethics as implemented in governance structures may start out with lofty ideals but it usually ends being whatever will cause no offense. What is ethical is what people don’t object to, at least not too loudly, and not the wrong people.
As an ethical ideal, this is both impossible and wrong-headed. People get offended by all kinds of things, many of them contradicting what other people get offended by, and often contradicting what they themselves are offended by. So to try to build a machine that won’t upset anyone is a hopeless task.
Let me give an example based on machine-generated images of people. If you notice that such images include a disproportionate number of white people, what do you do? The output, which simply mashes together countless input images, has not been informed by any moral considerations, since machines do not have any.
One cohort would argue that the “biases” in machine output represent “reality”, in this case, that it reflects the inputs.
Others would argue that we should adjust the output to reflect greater diversity. One problem with this is that, since the output is non-deterministic, it is hard to figure out how to adjust the output, and easy to hack around any such fixes. This is how we end up with an image prompt of “depiction of German soldier from 1943” generating people of Asian and African heritage. This is then described as “racist” by people who prefer their images “realistic”.
What is actually happening is that, since a disproportionate number of images on the internet are of white people, a mash-up of those images produces mostly white people, which accords with the “reality” of white supremacists. It’s an impossible situation, since you can never reconcile these two very different moral viewpoints. Nor should you. Failure to offend white supremacists is not an ethical goal.
If we believe that our machine should generate images that create the impression in people of racial diversity, what does that mean, exactly? Most people live, and have historically always lived, surrounded by people who for the most part look like them. They like it that way. Should we live in diverse societies? Or to put it another way, should the idealized vision of a multicultural society like San Francisco be the norm for everyone? Is that not just another form of cultural colonization? How far ought we go in recognition of minorities? What percentage of images should reflect the specific features of Sri Lankan Veddas, or the Dai people of Arunachal Pradesh? Should the output reflect the diversity of a coffee shop in Sydney’s inner west, or a coffee shop in the suburbs of Lagos?
Racial features, however, are only the tip of the iceberg. Who is out there building a machine that depicts ugly people? People with diseases, or who are odd looking, or just plain plain? Humans like to look at beautiful humans, so it is at the faces of beautiful people that we point our cameras. This is a question of desire, not of morality. Our databases of photos skew massively towards those that human beings consider beautiful, and machines routinely create output of inhumanly perfect beauty. The nature of that beauty, moreover, is shaped by the sexual desires of straight men, so it overwhelmingly consists of young women to be leered over. They might look like teens, but none of them have any acne. This creates a further layer of ethical problems, as girls now have to grow up being compared, not just with their peers, not just with the airbrushed ideals of models, but with machine-generated images of flawless perfection.
None of these ethical problems are soluble even in principle. The world is too messy and our morality inadequate. What will happen is that people will complain, then the company will make adjustments until the complaints go away. Or more to the point, until the cost of the complaint is less than the cost of fixing things. Or still more to the point, until someone in the company imagines that the cost is too great and they just move on (none of these companies are rational actors; they routinely make important decisions based on gut feelings).
This is what I mean by ethics being reduced to inoffensiveness. There’s no true sense of what is right and wrong, because, once again, it is a machine. The thing is, being offended is a purely subjective reaction, and it is trivial to get people offended by things. Happens all the time. So any process aimed at minimizing offense becomes prey to those who weaponize offense to get what they want. And of course, that is exactly what is happening. Whoever is most persistent and effective at generating outrage wins. And historically, groups of people who are large and loud are not known for choosing wisely.
All of this completely misses the point of ethics. In Buddhism, moral conduct stems from an inner sense of conscience and ethical prudence. This is something that is not rational, although it can be considered in terms of input (i.e. what triggered the ethical questions) and output (i.e. what decision is made). How and why one feels about that decision, however, is a matter of one’s own inner moral and psychological development. It simply isn’t the case that every person’s moral intuitions are equal.
Ethical decisions are informed by compassion, an empathetic understanding of how ones own actions affect others. It is impossible to have compassion (literally “feeling with”) if you are a machine with no feelings. Nor is there an inner sense of self to which can compare the inner experiences of others.
Ethics are also informed by one’s views, as it is a critical dimension of human consciousness that we try, however inadequately, to act consistently in ways that accord with our own values. Machines have no views and no values, and they do not have intentional actions. They merely get fed a bunch of data, digest it, mash it up, and spit it out.
Such views also relate to a sense of the consequences of moral choices. Ethical choices tend to lead to more happiness than suffering. And since we are sentient beings, when we make such choices, we include ourselves in that. We do good because it makes others happy, sure, but also because it makes us happy. In other words, we understand that we will have to bear responsibility for our actions, good or ill. Machines do not, because they have no inner life, no experience of happiness or suffering, and no consequences.
There is also a social dimension to morality. Fundamental principles like harmlessness are universal, but the way they are interpreted and applied varies tremendously. We perceive something as good because others see it as good: morality is triangulated. And when we are surrounded by others who see a certain way, it is difficult to see it another way. This social aspect of morality is deliberately weaponized by bad faith actors on the internet, who criticize what they call “virtue signaling”. They want you to believe that it is a bad thing to speak of one’s own virtues. But it is not: it’s how morality is communicated. The criticism of “virtue signaling” is intended, not to highlight hypocrisy, but to undermine morality itself. That’s why moral community is so important. Machines have no community, no people, and so they have no sense of shame when their private deeds are made public.
These are just some of the dimensions of morality. But there is a holistic quality to morality, one that resists reduction to any particular ethical theory or methodology. People have, over history, espoused all kinds of theories: “ethics is laid down by God in the Bible”, “what is ethical is what brings about the greatest good of the greatest number”, “ethical truths are self-evident to an ideal observer”. There seems to be a hidden paradox where if you accept in totality the tenets of any particular ethical theory, it drives you to monstrosity. I believe this is why the Buddha does not have a reductive ethical approach, but takes into account intention, action, and consequence, all defined within a particular context, the search for freedom.
The ideal of an ethical machine is a delusion, generated by the human desire to read faces in the clouds, and our inner compulsion to imagine i/o machines as something they are not. We should entirely stop talking about “AI ethics” and instead use “AI acceptability”, which shifts focus from the “behavior” of the AI to the degree to which humans will put up with it.
And can we not find it in ourselves to embrace as a bare minimum a watered-down version of the first of Asimov’s rules, and outlaw any AI that is intentionally designed to harm human beings?