Sure, Robots Can Feel Sorry For You


Let our journalists help you make sense of the noise: Subscribe to the Mother Jones Daily newsletter and get a recap of news that matters.

I see that Sherry Turkle is at it again:

Years ago I spoke with a 16-year-old girl who was considering the idea of having a computer companion in the future….But there is something she may have been too young to understand — or, like a lot of us — prone to forget when we talk to machines. These robots can perform empathy in a conversation about your friend, your mother, your child or your lover, but they have no experience of any of these relationships. Machines have not known the arc of a human life. They feel nothing of the human loss or love we describe to them. Their conversations about life occupy the realm of the as-if.

Yet through our interactions with these machines, we seem to ignore this fact; we act as though the emotional ties we form with them will be reciprocal, and real, as though there is a right kind of emotional tie that can be formed with objects that have no emotions at all. In our manufacturing and marketing of these machines, we encourage children to develop an emotional tie that is sure to lead to an empathic dead end….Children will lose the ability to have empathy if they relate too consistently with objects that cannot form empathic ties.

Meanwhile, in the real world:

Eighty-nine volunteers were asked to help improve a robot’s interactions by completing two tasks with it: creating a weekly schedule and answering such questions as “Do you rather like pizza or pasta?” The tasks with the robot, named Nao, were actually part of a ploy, however. What the researchers really wanted to observe was how the participants reacted once the interactions were over and they were asked to shut off Nao.

“No! Please do not switch me off! I am scared that it will not brighten up again!,” Nao said to about half of the participants….Of the 43 people who heard Nao beg to stay online, 13 chose to listen and did not turn him off, according to the study. Some merciful participants said they felt sorry for Nao and his fear of the void.

I realize there’s little point in people like Turkle and me arguing with each other. Neither of us knows what will happen in the future and neither of us knows what kinds of advances we’ll make in artificial intelligence. In the end, if we were doing this in person, we’d probably end up in some version of Are too! Is not! until we got tired of each other.

And yet! Consider the assumptions Turkle is making. First, she’s assuming that human empathy is based on certain experiences that no robot will ever have. But why not? There’s nothing magic about human emotion. It’s all neurons and cortisol and dopamine and so forth, just like everything else in the human brain. If we feel like it, we can program analogues of human neurochemistry into an artificial intelligence and then send it out into the world to have all sorts of emotional experiences. After we’ve done that a few thousand times, we’ll have a template we can copy into any new robot if we feel like it. We wouldn’t bother with this if we were building a robot to assemble cars, but we would if we were building a robot to take care of a child.

How real would this seem to a person interacting with a robot? Well, those 89 volunteers in the second story spent a grand total of a few minutes with Nao engaged in tasks that built no emotional connection at all. And yet, a surprising number of them thought that Nao’s simulated fearfulness was real enough that they refused to toggle off the power switch. Presumably, all it took was a crude bit of manipulation of Nao’s voice to get this reaction.

So will robots develop empathy? I don’t see why not. Will they at least develop simulated empathy as good as the real kind? Plenty of humans do that successully—actors, sociopaths used car salesmen, etc.—so again, I don’t see why not. In fact, here’s my prediction: artificial intelligence will eventually do everything better than humans do it. That includes the development and expression of emotions.

Humans are both less and more impressive than most of us think. Many people take for granted our immense ability to manipulate the world and create complex societies using nothing more than a 3-pound lump of nerve tissue working at chemical speeds with only a few watts of energy. That’s impressive! At the same time, though, three pounds of nerve tissue and a few watts of power just isn’t an awful lot. Much of what we do is, for lack of a better word, little more than crude simulation and simpleminded heuristics: we’re just faking things we’ve seen before. And crude or not, most of the time this is better than the alternative. When we’re not simulating, we’re all too often just reacting blindly to a rush of neurotransmitters that are themselves reacting blindly to some perceived effect on the body.

So yeah: robots will develop empathy. They’ll do it because we’ll program them to do it. Will it be “real”? Of course it will. It will be as real as the chemical version we all rely on now. Will we like it? We’ll love it. And our robots will love us back.


Headshot of Editor in Chief of Mother Jones, Clara Jeffery

It sure feels that way to me, and here at Mother Jones, we’ve been thinking a lot about what journalism needs to do differently, and how we can have the biggest impact.

We kept coming back to one word: corruption. Democracy and the rule of law being undermined by those with wealth and power for their own gain. So we're launching an ambitious Mother Jones Corruption Project to do deep, time-intensive reporting on systemic corruption, and asking the MoJo community to help crowdfund it.

We aim to hire, build a team, and give them the time and space needed to understand how we got here and how we might get out. We want to dig into the forces and decisions that have allowed massive conflicts of interest, influence peddling, and win-at-all-costs politics to flourish.

It's unlike anything we've done, and we have seed funding to get started, but we're looking to raise $500,000 from readers by July when we'll be making key budgeting decisions—and the more resources we have by then, the deeper we can dig. If our plan sounds good to you, please help kickstart it with a tax-deductible donation today.

Thanks for reading—whether or not you can pitch in today, or ever, I'm glad you're with us.

Signed by Clara Jeffery

Clara Jeffery, Editor-in-Chief

payment methods

We Recommend