It’s OK to call it Artificial Intelligence


Update 9th January 2024: This post was clumsily written and failed to make the point I wanted it to make. I’ve published a follow-up, What I should have said about the term Artificial Intelligence which you should read instead.

My original post follows.


We need to be having high quality conversations about AI: what it can and can’t do, its many risks and pitfalls and how to integrate it into society in the most beneficial ways possible.

Any time I write anything that mentions AI it’s inevitable that someone will object to the very usage of the term.

Strawman: “Don’t call it AI! It’s not actually intelligent—it’s just spicy autocomplete.”

That strawman is right: it’s not “intelligent” in the same way that humans are. And “spicy autocomplete” is actually a pretty good analogy for how a lot of these things work. But I still don’t think this argument is a helpful contribution to the discussion.

We need an agreed term for this class of technology, in order to have conversations about it. I think it’s time to accept that “AI” is good enough, and is already widely understood.

I’ve fallen for this trap myself. Every time I write a headline about AI I find myself reaching for terms like “LLMs” or “Generative AI”, because I worry that the term “Artificial Intelligence” over-promises and implies a false mental model of a sci-fi system like Data from Star Trek, not the predict-next-token technology we are building with today.

I’ve decided to cut that out. I’m going to embrace the term Artificial Intelligence and trust my readers to understand what I mean without assuming I’m talking about Skynet.

The term “Artificial Intelligence” has been in use by academia since 1956, with the Dartmouth Summer Research Project on Artificial Intelligence—this field of research is nearly 70 years old now!

John McCarthy’s 2nd September 1955 proposal for that workshop included this:

The study is to proceed on the basis of the conjecture that every aspect of learning or any other feature of intelligence can in principle be so precisely described that a machine can be made to simulate it. An attempt will be made to find how to make machines use language, form abstractions and concepts, solve kinds of problems now reserved for humans, and improve themselves.

I think this is a very strong definition, which fits well with the AI models and use-cases we are talking about today. Let’s use it.

Why not LLMs?

I’ve spent the past year mainly talking about LLMs—Large Language Models—often as an alternative to the wider term “AI”.

While this term is accurate, it comes with a very significant downside: most people still don’t know what it means.

I find myself starting every article with “Large Language Models (LLMs), the technology behind ChatGPT and Google Bard...”

I don’t think this is helpful. Why use a relatively obscure term I have to define every single time, just because the word “intelligence” in the common acronym AI might be rejected by some readers?

The term LLM is already starting to splinter as language models go multi-modal. I’ve seen the term “LMMs” for Large Multimodal Models start to circulate, which risks introducing yet another piece of jargon that people need to understand in order to comprehend my writing!

The argument against using the term AI

My link to this post on Mastodon has attracted thoughtful commentary that goes well beyond the straw man argument I posed above.

The thing that’s different with the current wave of AI tools, most notably LLMs, is that at first glance they really do look like the AI from science fiction. You can have conversations with them, they exhibit knowledge of a vast array of things, and it’s easy to form an initial impression of them that puts them at the same level of “intelligence” as Jarvis, or Data, or a T-800.

The more time you spend with them, and the more you understand about how they work, the more this illusion falls apart as their many flaws start to become apparent.

Where this gets actively harmful is when people start to deploy systems under the assumption that these tools really are trustworthy, intelligent systems—capable of making decisions that have a real impact on people’s lives.

This is the thing we have to fight back against: we need to help people overcome their science fiction priors, understand exactly what modern AI systems are capable of, how they can be used responsibly and what their limitations are.

I don’t think refusing to use the term AI is an effective tool to help us do that.

Let’s tell people it’s “not AGI” instead

If we’re going to use “Artificial Intelligence” to describe the entire field of machine learning, generative models, deep learning, computer vision and so on... what should we do about the science fiction definition of AI that’s already lodged in people’s heads?

Our goal here is clear: we want people to understand that the LLM-powered tools they are interacting with today aren’t actually anything like the omniscient AIs they’ve seen in science fiction for the past ~150 years.

Thankfully there’s a term that’s a good fit for this goal already: AGI, for “Artificial General Intelligence”. This is generally understood to mean AI that matches or exceeds human intelligence.

AGI itself is vague and infuriatingly hard to define, but in this case I think that’s a feature. “ChatGPT isn’t AGI” is an easy statement to make, and I don’t think its accuracy is even up for debate.

The term is right there for the taking. “You’re thinking about science fiction there: ChatGPT isn’t AGI, like in the movies. It’s just an AI language model that can predict next tokens to generate text.”

Miscellaneous additional thoughts

There’s so much good stuff in the conversation about this post. I already added the new sections Why not LLMs?, The argument against using the term AI and Let’s tell people it’s “not AGI” instead based on those comments.

I’ll collect a few more miscellaneous thoughts in this section, which I may continue to grow in the future.

Here’s what Glyph Lefkowitz had to say about this last point:

A lot of insiders—not practitioners as such, but marketers & executives—use “AI” as the label not in spite of its confusion with the layperson’s definition, but because of it. Investors who vaguely associate it with machine-god hegemony assume that it will be very profitable. Users assume it will solve their problems. It’s a term whose primary purpose has become deceptive.

At the same time, a lot of the deception is unintentional. When you exist in a sector of the industry that the public knows as “AI”, that the media calls “AI”, that industry publications refer to as “AI”, that other products identify as “AI”, going out on a limb and trying to build a brand identity around pedantic hairsplitting around “LLMs” and “machine learning” is a massive uphill battle which you are disincentivized at every possible turn to avoid.

Glyph’s closing thought here reflects my own experience: I tried to avoid leaning too hard into the term “AI” myself, but eventually it felt like an uphill battle that was resulting in little to no positive impact.