‘Creepy Facebook Al’ Story sweeps media

The newspapers have a scoop today – it seems that artificial intelligence (AI) could be out to get us.

‘Robot intelligence is dangerous’: Expert’s warning after Facebook AI ‘develop their own language'”, says the Mirror.

Similar stories have appeared in the Sun, the Independent, the Telegraph and in other online publications.

It sounds like something from a science fiction film – the Sun even included a few pictures of scary-looking androids.

So, is it time to panic and start preparing for apocalypse at the hands of machines?

Probably not. While some great minds – including Stephen Hawking – are concerned that one day AI could threaten humanity, the Facebook story is nothing to be worried about.

Where did the story come from?

Way back in June, Facebook published a blog post about interesting research on chatbot programs – which have short, text-based conversations with humans or other bots. The story was covered by New Scientist and others at the time.

Facebook had been experimenting with bots that negotiated with each other over the ownership of virtual items.

It was an effort to understand how linguistics played a role in the way such discussions played out for negotiating parties, and crucially the bots were programmed to experiment with language in order to see how that affected their dominance in the discussion.

Although some reports insinuate that the bots had at this point invented a new language in order to elude their human masters, a better explanation is that the neural networks had simply modified human language for the purposes of more efficient interaction.

As technology news site Gizmodo said: “In their attempts to learn from each other, the bots thus began chatting back and forth in a derived shorthand – but while it might look creepy, that’s all it was.”

AIs that rework English as we know it in order to better compute a task are not new.

Google reported that its translation software had done this during development. “The network must be encoding something about the semantics of the sentence” Google said in a blog.

And earlier this year, Wired reported on a researcher at OpenAI who is working on a system in which AIs invent their own language, improving their ability to process information quickly and therefore tackle difficult problems more effectively.

The story seems to have had a second wind in recent days, perhaps because of a verbal scrap over the potential dangers of AI between Facebook chief executive Mark Zuckerberg and technology entrepreneur Elon Musk.


But the way the story has been reported says more about cultural fears and representations of machines than it does about the facts of this particular case.

Plus, let’s face it, robots just make for great villains on the big screen.

In the real world, though, AI is a huge area of research at the moment and the systems currently being designed and tested are increasingly complicated.

One result of this is that it’s often unclear how neural networks come to produce the output that they do – especially when two are set up to interact with each other without much human intervention, as in the Facebook experiment.

That’s why some argue that putting AI in systems such as autonomous weapons is dangerous.

It’s also why ethics for AI is a rapidly developing field – the technology will surely be touching our lives ever more directly in the future.



Latest posts

KEWOPA, Google partner to improve gender equity

Claire Wanja

Tech pioneer Sir Clive Sinclair dies aged 81

Claire Wanja

Nokia, ATU to speed up digital transformation in Africa

Claire Wanja

This website uses cookies to improve your experience. We'll assume you're ok with this, but you can opt-out if you wish. Accept Read More