Fourth Industrial Revolution

Our language needs to evolve alongside AI. Here's how

Commuters wait for the subway train to stop in New York City October 7, 2005. Part of the section of New York's Pennsylvania Station where Amtrak trains operate was sealed off on Friday morning because of what authorities said was a police situation, a day after officials warned of a possible attack on the system. The precaution came one day after New York officials said they had intelligence that the city's subway system was under a credible and specific threat of possible attack. REUTERS/Seth Wenig - RP2DSFIMBZAA

Every age has its own technological advancements as well as new words to go alongside them. Image: REUTERS/Seth Wenig

Byron Reese
Contributor, LSE Business Review
Share:
Our Impact
What's the World Economic Forum doing to accelerate action on Fourth Industrial Revolution?
The Big Picture
Explore and monitor how The Digital Economy is affecting economies, industries and global issues
A hand holding a looking glass by a lake
Crowdsource Innovation
Get involved with our crowdsourced digital platform to deliver impact at scale
Stay up to date:

The Digital Economy

New technologies don’t just change the world, they change our language too. The automobile era gave us over a thousand new words and phrases such as limousine, drive through, hot rod, tailgate, and muscle car. Computers gave us chips, CPUs, beta testers, operating systems, programs and bugs. The Internet has already contributed hundreds of new words and new meanings to old words, including spam, meme, hashtag and trolling.

This all makes sense. Technology gives us new experiences, new abilities, and new problems. Thus we need new words to keep up with this changing world. The great technology of our time, artificial intelligence, will do the same – it will change the world and our language with it.

Image: Statista

Usually the new words flow from necessity and are created organically. But with AI, I am going to jump the gun a bit and suggest some new words that we are going to need to adjust to this new technology.

AiporiaUncertainty as to whether you are dealing with a human or an AI. From the ancient Greek aporia (ᾰ̓πορῐ́ᾱ), meaning a state of puzzlement.

Aiporia is a new phenomenon, but can be experienced on a regular basis. For instance, say you are on a website that has a real-time chat feature. You click on it, and up comes a message from “Sarah” asking if you need any help. You may wonder if Sarah is 1) a person named Sarah, 2) a person from another country not named Sarah presenting herself as “Sarah” to be more accessible, or 3) a chatbot named “Sarah.”

The confusion of aiporia is magnified further by the fact that all three may be true. Sarah may start out as a chatbot, but when you ask a more challenging question than “How much does shipping cost?”, the chat is continued by someone else, perhaps in another country. If your questions might be elevated to a more senior person, perhaps in the United States, and perhaps and coincidentally named Sarah. Expect aiporia to become more common.

AinigmaA decision made by an AI that a human cannot understand. Derived from enigma.

Pretend your company ranks third in Google for some search that is important to you, such as “Akron pool cleaning” and your competitor ranks second. Further suppose you found a Google engineer and asked them why this is. Likely they would shrug and say there is no way to know. There are simply too many factors at work for a human to understand those kinds of fine distinctions. It is an ainigma.”

Aithics – n – The moral aspects of an AI’s programming.

AI instantiates the morals and ethics of the programmers who create them. There is no escaping this. AI makes choices, choices involve relative values, and values are the manifestations of ethics. You can’t make an AI without any moral implications. The AI always has aithics of some kind. The question is simply whose it has.

BaisThe biases of an AI that stem from the data used to train it. From bias.

AIs are trained on data. Not all the data in the world, obviously, but a selection of data. Selecting which data to use is a choice made by a human, and such choices involve human bias. The AI isn’t biased. It is perfectly representing all the data in its world. But it is baised.

Brotherism – n – The use of AI to monitor and predict human behaviour. This can include “listening” to all phone conversations and “reading” all emails. Often done by governments. Derived from Big Brother, the Orwellian trope from 1984.

Data mining tools, the kinds we build for noble purposes such as finding new treatments for diseases, can just as easily be used to monitor people. They can read every email, convert every cell phone conversation to text, even use surveillance cameras to read lips as well as a human can. All of this together can be used to model every one of us, and make accurate predictions on whether we might commit a crime or who we might vote for. When wielded by the government, this is enormous power. This is brotherism.

Have you read?

Eception – n – an AI that is deliberately programmed to seem human and designed to trick humans. Often on social media.

Have you ever received an automated call that begins with, “Hi, this is Rachel. Can you hear me?” That is an eception. So are the millions or billions of bots that pretend to be human to try to get you to believe something or buy something.

Elize – The act of treating a machine as though it has emotions and feelings due to its life-like behaviour. From the 1960s computer program ELIZA.

In the 1960s, computer pioneer Joseph Weizenbaum was horrified to see people treat his simple psychologist chatbot, ELIZA, as though it were human. We instinctively do this. If something seems alive to us, we treat it differently than we treat, say, a power saw. As machines become more human-like, we are likely to elize them further, and be hesitant to interrupt them or put them in harm’s way.

Fauxnesis – n – Machine intelligence, as different from biological intelligence. From the Ancient Greek Phronesis, meaning intelligence.

Can a machine think? Alan Turing posed this question in 1950 and we still don’t have an answer. In reply, Noam Chomsky said, “That’s like asking if submarines swim.” Maybe the problem isn’t that we don’t know how to answer the question, rather that we don’t yet have a word to answer it. The word we need is fauxnesis.

Zomek – Machine life, as different from biological life. From the Ancient Greek zoe, meaning life, and mekhane, meaning machine.

There is no consensus definition of life. That’s a pretty interesting fact; that we can’t put into words what life is. Then it comes as no surprise that we are at a loss to say whether computers can be alive. Maybe they aren’t alive like us, but they have zomek, machine life.

Don't miss any update on this topic

Create a free account and access your personalized content collection with our latest publications and analyses.

Sign up for free

License and Republishing

World Economic Forum articles may be republished in accordance with the Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International Public License, and in accordance with our Terms of Use.

The views expressed in this article are those of the author alone and not the World Economic Forum.

Related topics:
Fourth Industrial RevolutionEmerging Technologies
Share:
World Economic Forum logo
Global Agenda

The Agenda Weekly

A weekly update of the most important issues driving the global agenda

Subscribe today

You can unsubscribe at any time using the link in our emails. For more details, review our privacy policy.

Generative AI is rapidly evolving: How governments can keep pace

Karla Yee Amezaga, Rafi Lazerson and Manal Siddiqui

October 11, 2024

About us

Engage with us

  • Sign in
  • Partner with us
  • Become a member
  • Sign up for our press releases
  • Subscribe to our newsletters
  • Contact us

Quick links

Language editions

Privacy Policy & Terms of Service

Sitemap

© 2024 World Economic Forum