Fourth Industrial Revolution

Should we be concerned about the safety of AI and robotics?

Share:
Our Impact
What's the World Economic Forum doing to accelerate action on Fourth Industrial Revolution?
The Big Picture
Explore and monitor how Fourth Industrial Revolution is affecting economies, industries and global issues
A hand holding a looking glass by a lake
Crowdsource Innovation
Get involved with our crowdsourced digital platform to deliver impact at scale
Stay up to date:

Fourth Industrial Revolution

This article is published in collaboration with The Conversation.

Roboticist Sabine Hauert, from the Britain’s University of Bristol, wrote in Nature earlier this year:

My colleagues and I spend dinner parties explaining that we are not evil […]

People are worried, she said.

They hear a mostly one-sided discussion that leaves them worried that robots will take their jobs, fearful that AI poses an existential threat, and wondering whether laws should be passed to keep hypothetical technology ‘under control’.

These fears are not helped by a continuing epidemic of artificial intelligence (AI) and robophobic screenplays emanating from Hollywood.

151214-robots growth market AI technology Japan Robotics Association

It is hard to give examples of recent mainstream films with robots and AIs in them that are not infected with AI-phobic (Transcendence, Eagle Eye) or robophobic (Oblivion, any of the Robocop movies, Ex Machina) scenes.

The indie caper picture Robot and Frank and the relatively mild dumping of the human by the AI in Her, who then runs off to hang out with cooler Alan Watts-based superintelligences, are the only ones coming to mind that do not succumb to the prevailing moods of AI panic and robophobia.

Robots and AIs do make good cinematic villains. However, in reality no one has much clue how to make a robot “want” or “feel” anything in a phenomenologically credible way as yet; let alone how to make them sociopaths hell bent on world domination and the extermination of Homo sapiens.

They are more likely to become innocently dangerous idiot savants than malevolent overlords seeking to get psychotic kicks by making humans “bend the knee”.

Rule-driven robots play a mean game of chess but feel nothing about winning or losing. They just pick moves that optimise a mathematical value function. Humans associate intelligence with desire but “desire” as formally modelled in the rulebook of a Turing machine is a very different thing from the combustive forces of fury, jealousy and “star-crossed love” that drive humans.

Responsible robots

Two new AI and robotics nonprofits launched over the weekend. In different ways, both are responses to public concerns about the safety of AI and robotics.

The first is The Foundation for Responsible Robotics (FRR), which wants to “promote responsibility for the robots embedded in our society”.

The FRR is headed by Noel Sharkey and various robotics experts. It aims to engage policymakers, create interdisciplinary teams of robotic, legal, ethical and societal scholars, work to explore what it means to be “responsible” as robotics researchers and designers, run workshops and engage the public.

Sharkey, who is known for his activism with the Campaign to Stop Killer Robots, is concerned that “we are rushing headlong into the robotics revolution” without giving enough policy thought to the social problems that might arise.

He says governments are looking to robotics as a “powerful new economic driver” but “only lip service is being paid to a long list of potential societal hazards”.

New technologies could cause mass unemployment or there might be an acceleration of social inequality caused by robots and automation leading to a society divided between robot-owners (living in gated communities) and a robot-less underclass (struggling on a brown burnt Earth) such as was depicted in Elysium .

The FRR wants to ensure that the public have confidence in robotics research and that robots will be developed with due regard for their human rights and freedom of choice.

AI for everyone

Open AI is backed by Elon Musk, Peter Thiel and various technology entrepreneurs.

Their focus is more on research and on making advanced AI freely available. They seek to develop innovations in “deep learning” a technique where rather than “hand-code a new algorithm for each problem, you design architectures that can twist themselves into a wide range of algorithms based on the data you feed them”.

They aim to “advance digital intelligence in the way that is most likely to benefit humanity as a whole, unconstrained by a need to generate financial return”.

When “human-level AI” arrives, Open AI feels that it is important that there be “a leading research institution which can prioritize a good outcome for all over its own self-interest”.

They say “our researchers will be strongly encouraged to publish their work, whether as papers, blog posts, or code, and our patents (if any) will be shared with the world”.

Open AI’s backers have committed a billion dollars in funding though they expect to “only spend a tiny fraction of this in the next few years.”

Open AI is mainly about open access to advanced AI thus reducing the risk of a world of AI haves and have nots. This is a good idea. If the team can keep competitive and advanced AI open source this should reduce the risk of people being excluded from advanced AI for financial reasons.

The Foundation for Responsible Robotics has a broader agenda of policy engagement and raising professional and public awareness of robot ethics issues. Again, this is a worthwhile endeavour. AI and robotics researchers tend to be hard scientists unused to ethical debate.

Scientists need to step out of the empirical and into the normative. As trusted thought leaders of the citizenry, they should cross the line between “is” and “ought” and participate in policy debate.

Hopefully both these groups will help provide cures for the current epidemics of AI panic and robophobia.The Conversation

Publication does not imply endorsement of views by the World Economic Forum.

To keep up with the Agenda subscribe to our weekly newsletter.

Author: Sean Welsh is a Doctoral Candidate in Robot Ethics at the University of Canterbury.

Image: Humanoid robot “blushes” during a ceremony. REUTERS/Wolfgang Rattay.

Don't miss any update on this topic

Create a free account and access your personalized content collection with our latest publications and analyses.

Sign up for free

License and Republishing

World Economic Forum articles may be republished in accordance with the Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International Public License, and in accordance with our Terms of Use.

The views expressed in this article are those of the author alone and not the World Economic Forum.

Related topics:
Fourth Industrial RevolutionEconomic Growth
Share:
World Economic Forum logo
Global Agenda

The Agenda Weekly

A weekly update of the most important issues driving the global agenda

Subscribe today

You can unsubscribe at any time using the link in our emails. For more details, review our privacy policy.

More on Fourth Industrial Revolution
See all

Future of the internet: Why we need convergence and governance for sustained growth

Thomas Beckley and Ross Genovese

April 25, 2024

About Us

Events

Media

Partners & Members

  • Join Us

Language Editions

Privacy Policy & Terms of Service

© 2024 World Economic Forum