In the new Netflix action / sci-fi movie ‘Atlas’, we get a
revisiting of a now-familiar tale: AI decides, in its infinite wisdom, that the
single biggest threat to humanity is, well, humans. So, the AI decides to kill
off vast swaths of humanity for humanity’s sake. Queue drama.
The set-up to the movie’s action amusingly includes a clip
from a real television interview with Geoffrey Hinton, sometimes called the ‘godfather
of AI’, a now-76-years-old retired researcher whose work at the University of
Toronto (and later Google) really broke open the current AI revolution, by
pioneering the idea that ‘neural networks’ could enact ‘deep learning’ algorithms,
leveraging graphical processing units (computer chips called GPUs) and the vast
stores of data on the internet (see my 2016 feature,
The Learning Machines). That philosophy is what has led us to
large language models like OpenAI’s ChatGPT. The clip in the movie shows Hinton
saying: “If we produce things that are more intelligent than us, how do we know
we can keep control? I mean, Isaac Asimov once said, if you make a smart robot,
the first rule should be 'do not harm people’.” Hinton is notably, famously,
concerned about the development of strong AI and how it might go rogue.
Atlas plays on this idea, but then, amusingly, takes a twist (no plot spoilers, this is in the preview and other movie reviews):
in order to save the world from a ‘bad AI’ named Harlan, Lopez must rely on a ‘good AI’ named Smith.
(I wrote to Hinton about all this, by the way, and asked whether anyone asked for his blessing to be in the film. He replied succinctly: "they did not ask me. Its a silly movie.")
Let’s take a look at a few of the themes here and see how
they’re interacting with reality. I’ll just note here that I’ve been writing a
lot about AI for Nature lately, as well as editing a lot of great features
written by other great writers; I’m drawing on a lot of what I have learned
from that experience here, with thanks to those amazing reporters.
We do not have completely humanoid robots (of the kind you’d
mistake for a person), but we do have some startlingly adept AI-driven machines
(see The AI Revolution is Coming to Robots). The company Figure has created a humanoid robot that
in a demo video recently showed its stuff: a person said “give me something to
eat” and the robot picked out an apple from the table and handed it to him.
This is pretty cool: it requires the robot to recognize an apple, know that it’s
food, and be able to pick it up and hand it over without crushing it or shoving
it in the person’s face, none of which is particularly obvious. Likewise,
Boston Dynamics is making some pretty impressive stunt-capable robots, the
latest of which is (by coincidence, I think) named Atlas.
There is a great debate as to how intelligent (in the human
sense) the large language models feeding these systems are; certainly they can
out-perform PhD level scientists on exams, speak fluently and comprehend
language as well the average person (see AI now beats humans at basic tasks). They are also fallible. AIs that
generate text and images can mistake parody for fact, present falsehoods with
great confidence, be racist and sexist, and fall victim to psychological
foibles like implicit bias (they can make up logical-sounding but irrational reasons
for why they come to certain conclusions) (see Psychology and neuroscience crack open AI).
Such ‘mistakes’ just make AI seem ever-more human. If a
chatbot gives the appearance of having sentience, does it have sentience? It is
distressingly hard to tell. One expert famously called his chatbot “slightly conscious”, which made many people laugh, and others take a pause.
In Star Trek (sorry for another fictional reference, but it’s
where we are), there is a famous court case in which they try to ascertain
whether the AI robot Data has rights or is ‘just a toaster’ that can be
considered property. As the defence rightly points out, while Data can’t prove
he is sentient, neither can anyone else prove that they themselves are. To be human and have human rights you
don’t need to have ‘normal feelings’ (some don’t), nor be smart (some aren’t).
To be alive you don’t have to have a brain (many organisms don’t). We are at a
point now in the real world where the fuzziness of these terms and definitions is coming to a
point. What is AI, exactly? That’s up for debate. If and when we develop a
fleet of AI-robots, can we justifiably treat them as property, or would that
make them slaves?
At any rate, since chatbots can go wrong, so can robots, and
that’s worrying. Wielding falsehoods is different from wielding, say, a knife. Isaac
Asimov did indeed lay out 3 useful rules of robots: in brief 1) robots shouldn’t
hurt people; 2) robots should follow orders, unless that hurts people; 3) a
robot should protect itself, unless that violates the first two rules. And yes,
roboticists are actually using these principles to program their robots: fiction
has come to life.
OpenAI is now engaged in a massive ‘superalignment’ project,
the intent of which is to bake an ethical concern for humanity into its AI –
basically to make sure that its AI is pro-humanity, and not evil. How exactly they
aim to do this remains unclear, but the training process for AI is vaguely
similar to parenting. An AI responds to a situation and the humans say: ‘yeah,
that’s good’, or ‘no no no, that’s not the way we do things’, until the
bot/child no longer hits people / mistakes garbage for food / thinks torturing
animals is okay.
It seems to me that speaking to any AI should be thought of
like speaking to a person (not like dealing with a calculator): in any given
interaction, we tend to make judgements about a person’s level of intelligence,
empathy, political leanings and more, and use that to help us assess what they
are saying, and how many grains of salt to add to what we take from it (which
we don’t do with a calculator). Of course, these assessments also can go wrong:
people tend to put greater trust, irrationally and sometimes to their detriment,
in ‘people like me’. And a good con artist can inspire trust when it isn’t
deserved.
This is the worry (also Hinton’s worry): a smart AI will be
a very, very good con artist, capable of manipulating peoples’ beliefs and
actions (as Harlan is, in the movie). I don’t think AI (at least not for a long
time yet) will have its own agenda, per se (whether that’s to save humanity or
take over the world). But that doesn’t stop someone from programming one in.
People who want to sway an election or convince people that smoking isn’t unhealthy
or that climate change isn’t a worry could harness convincing-AI very
powerfully. And that is worrying.
Even well-intentioned programming is slippery and hard to
define. Philosophers and ethicists know this. Would/should you push one person
off a bridge onto a train track in order to save hundreds of lives? There’s no
one right answer to that question. This came up with self-driving cars, when
programmers realized they had to get their cars to make hard choices: whether
to hit a pedestrian, say, or instead swerve into a concrete barrier, potentially
killing the driver. The moral machine project, in response to this problem, aimed to collate ethical thinking
from different cultures around the world.
It's worth noting that self-driving cars have been involved
in fatal incidents: the National Highway Traffic Safety Administration (NHTSA) has
apparently tracked more than 11 deaths across the US involving cars with AI
(though so far it seems AI cars are safer than human-driven ones; accidents
happen). So: in one sense at least, AI robots have already killed people.
I’m not so much worried about AIs developing their own
agency and agendas, but I am worried about people harnessing AI to fulfill their
own agendas. (On the whole I’m pretty optimistic about human nature, but bad
actors do exist.) Future AIs – chatbots or robots – should come with
disclaimers, saying who paid for their programming, what political beliefs they
hold and how often they get their facts wrong. Interestingly, this is what we
already have (and need investigators to track down) for things like newspapers
and prominent political figures. The skills we have – and teach our children –
for assessing the value of information will be ever, ever more important.
So: my takeaways from the movie? I bet we’ll be able to put
in safeguards to stop terrorist AIs from killing humanity. But I also bet we
can’t make AI that’s perfectly wholesome and trustworthy and right. The reality
is somewhere in between Harlan and Smith, and is likely to stay there.