Thursday, May 30, 2024

The movie Atlas begs the question: will AI robots kill us all?

In the new Netflix action / sci-fi movie ‘Atlas’, we get a revisiting of a now-familiar tale: AI decides, in its infinite wisdom, that the single biggest threat to humanity is, well, humans. So, the AI decides to kill off vast swaths of humanity for humanity’s sake. Queue drama.


The set-up to the movie’s action amusingly includes a clip from a real television interview with Geoffrey Hinton, sometimes called the ‘godfather of AI’, a now-76-years-old retired researcher whose work at the University of Toronto (and later Google) really broke open the current AI revolution, by pioneering the idea that ‘neural networks’ could enact ‘deep learning’ algorithms, leveraging graphical processing units (computer chips called GPUs) and the vast stores of data on the internet (see my 2016 feature, The Learning Machines). That philosophy is what has led us to large language models like OpenAI’s ChatGPT. The clip in the movie shows Hinton saying: “If we produce things that are more intelligent than us, how do we know we can keep control? I mean, Isaac Asimov once said, if you make a smart robot, the first rule should be 'do not harm people’.” Hinton is notably, famously, concerned about the development of strong AI and how it might go rogue.

Atlas plays on this idea, but then, amusingly, takes a twist (no plot spoilers, this is in the preview and other movie reviews): in order to save the world from a ‘bad AI’ named Harlan, Lopez must rely on a ‘good AI’ named Smith. 

(I wrote to Hinton about all this, by the way, and asked whether anyone asked for his blessing to be in the film. He replied succinctly: "they did not ask me. Its a silly movie.")

Let’s take a look at a few of the themes here and see how they’re interacting with reality. I’ll just note here that I’ve been writing a lot about AI for Nature lately, as well as editing a lot of great features written by other great writers; I’m drawing on a lot of what I have learned from that experience here, with thanks to those amazing reporters.

We do not have completely humanoid robots (of the kind you’d mistake for a person), but we do have some startlingly adept AI-driven machines (see The AI Revolution is Coming to Robots). The company Figure has created a humanoid robot that in a demo video recently showed its stuff: a person said “give me something to eat” and the robot picked out an apple from the table and handed it to him. This is pretty cool: it requires the robot to recognize an apple, know that it’s food, and be able to pick it up and hand it over without crushing it or shoving it in the person’s face, none of which is particularly obvious. Likewise, Boston Dynamics is making some pretty impressive stunt-capable robots, the latest of which is (by coincidence, I think) named Atlas.

There is a great debate as to how intelligent (in the human sense) the large language models feeding these systems are; certainly they can out-perform PhD level scientists on exams, speak fluently and comprehend language as well the average person (see AI now beats humans at basic tasks). They are also fallible. AIs that generate text and images can mistake parody for fact, present falsehoods with great confidence, be racist and sexist, and fall victim to psychological foibles like implicit bias (they can make up logical-sounding but irrational reasons for why they come to certain conclusions) (see Psychology and neuroscience crack open AI).

Such ‘mistakes’ just make AI seem ever-more human. If a chatbot gives the appearance of having sentience, does it have sentience? It is distressingly hard to tell. One expert famously called his chatbot “slightly conscious”, which made many people laugh, and others take a pause.

In Star Trek (sorry for another fictional reference, but it’s where we are), there is a famous court case in which they try to ascertain whether the AI robot Data has rights or is ‘just a toaster’ that can be considered property. As the defence rightly points out, while Data can’t prove he is sentient, neither can anyone else prove that they themselves are. To be human and have human rights you don’t need to have ‘normal feelings’ (some don’t), nor be smart (some aren’t). To be alive you don’t have to have a brain (many organisms don’t). We are at a point now in the real world where the fuzziness of these terms and definitions is coming to a point. What is AI, exactly? That’s up for debate. If and when we develop a fleet of AI-robots, can we justifiably treat them as property, or would that make them slaves?

At any rate, since chatbots can go wrong, so can robots, and that’s worrying. Wielding falsehoods is different from wielding, say, a knife. Isaac Asimov did indeed lay out 3 useful rules of robots: in brief 1) robots shouldn’t hurt people; 2) robots should follow orders, unless that hurts people; 3) a robot should protect itself, unless that violates the first two rules. And yes, roboticists are actually using these principles to program their robots: fiction has come to life.

OpenAI is now engaged in a massive ‘superalignment’ project, the intent of which is to bake an ethical concern for humanity into its AI – basically to make sure that its AI is pro-humanity, and not evil. How exactly they aim to do this remains unclear, but the training process for AI is vaguely similar to parenting. An AI responds to a situation and the humans say: ‘yeah, that’s good’, or ‘no no no, that’s not the way we do things’, until the bot/child no longer hits people / mistakes garbage for food / thinks torturing animals is okay.

It seems to me that speaking to any AI should be thought of like speaking to a person (not like dealing with a calculator): in any given interaction, we tend to make judgements about a person’s level of intelligence, empathy, political leanings and more, and use that to help us assess what they are saying, and how many grains of salt to add to what we take from it (which we don’t do with a calculator). Of course, these assessments also can go wrong: people tend to put greater trust, irrationally and sometimes to their detriment, in ‘people like me’. And a good con artist can inspire trust when it isn’t deserved.

This is the worry (also Hinton’s worry): a smart AI will be a very, very good con artist, capable of manipulating peoples’ beliefs and actions (as Harlan is, in the movie). I don’t think AI (at least not for a long time yet) will have its own agenda, per se (whether that’s to save humanity or take over the world). But that doesn’t stop someone from programming one in. People who want to sway an election or convince people that smoking isn’t unhealthy or that climate change isn’t a worry could harness convincing-AI very powerfully. And that is worrying.

Even well-intentioned programming is slippery and hard to define. Philosophers and ethicists know this. Would/should you push one person off a bridge onto a train track in order to save hundreds of lives? There’s no one right answer to that question. This came up with self-driving cars, when programmers realized they had to get their cars to make hard choices: whether to hit a pedestrian, say, or instead swerve into a concrete barrier, potentially killing the driver. The moral machine project, in response to this problem, aimed to collate ethical thinking from different cultures around the world.

It's worth noting that self-driving cars have been involved in fatal incidents: the National Highway Traffic Safety Administration (NHTSA) has apparently tracked more than 11 deaths across the US involving cars with AI (though so far it seems AI cars are safer than human-driven ones; accidents happen). So: in one sense at least, AI robots have already killed people.

I’m not so much worried about AIs developing their own agency and agendas, but I am worried about people harnessing AI to fulfill their own agendas. (On the whole I’m pretty optimistic about human nature, but bad actors do exist.) Future AIs – chatbots or robots – should come with disclaimers, saying who paid for their programming, what political beliefs they hold and how often they get their facts wrong. Interestingly, this is what we already have (and need investigators to track down) for things like newspapers and prominent political figures. The skills we have – and teach our children – for assessing the value of information will be ever, ever more important.

So: my takeaways from the movie? I bet we’ll be able to put in safeguards to stop terrorist AIs from killing humanity. But I also bet we can’t make AI that’s perfectly wholesome and trustworthy and right. The reality is somewhere in between Harlan and Smith, and is likely to stay there.

 

No comments:

Post a Comment