A place to cache linked articles (think custom and personal wayback machine)
Du kannst nicht mehr als 25 Themen auswählen Themen müssen mit entweder einem Buchstaben oder einer Ziffer beginnen. Sie können Bindestriche („-“) enthalten und bis zu 35 Zeichen lang sein.

index.md 23KB

title: Artificial General Intelligence and the bird brains of Silicon Valley url: https://softwarecrisis.dev/letters/ai-bird-brains-silicon-valley/ hash_url: f23d043d8e

The problem is, if one side of the communication does not have meaning, then the comprehension of the implicit meaning is an illusion arising from our singular human understanding of language (independent of the model). Contrary to how it may seem when we observe its output, an LM is a system for haphazardly stitching together sequences of linguistic forms it has observed in its vast training data, according to probabilistic information about how they combine, but without any reference to meaning: a stochastic parrot.

Emily M. Bender, Timnit Gebru, et al., On the Dangers of Stochastic Parrots: Can Language Models Be Too Big?.

Bird brains have a bad reputation. The diminutive size of your average bird and their brain has lead people to assume that they are, well, dumb.

But, bird brains are amazing. Birds commonly outperform mammals with larger brains at a variety of general reasoning and problem-solving tasks. Some by a large margin. Their small brains manage this by packing numerous neurons in a small space using structures that are unlike from those you find in mammals.

Even though birds have extremely capable minds, those minds are built in ways that are different from our own or other mammals. Similar capabilities; different structure.

The ambition of the Silicon Valley AI industry is to create something analogous to a bird brain: a new kind of mind that is functionally similar to the human mind, possibly outperforming it, while being built using very different mechanisms. Similar capabilities; different structure.

This effort goes back decades, to the dawn of computing, and has had limited success.

Until recently, it seems.

If you’re reading this, you’ve almost certainly interacted with a Generative AI, however indirectly. Maybe you’ve tried Bing Chat. Maybe you’ve subscribed to the paid tier for ChatGPT. Or, maybe you’ve used Midjourney to generate images. At the very least you’ve been forced to see the images or text posted by the overenthusiastic on social media.

These AI models are created by pushing an enormous amount of training data through various algorithms:

  • Language models like ChatGPT is trained on a good chunk of the textual material available in digital form in the world.
  • Image models like Midjourney and Stable Diffusion are trained on a huge collection of images found on the internet.

What comes out the other end is a mathematical model of the media domain in question: text or images.

You know what Generative AI is in terms of how it presents to you as software: clever chatbots that do or say things in response to what you say: your prompt. Some of those responses are useful, and they give you an impression of sophisticated comprehension. The models that generate text are fluent and often quite engaging.

This fluency is misleading. What Bender and Gebru meant when they coined the term stochastic parrot wasn’t to imply that these are, indeed, the new bird brains of Silicon Valley, but that they are unthinking text synthesis engines that just repeat phrases. They are the proverbial parrot who echoes without thinking, not the actual parrot who is capable of complex reasoning and problem-solving.

A zombie parrot, if you will, that screams for brains because it has none.

The fluency of the zombie parrot—the unerring confidence and a style of writing that some find endearing—creates a strong illusion of intelligence.

Every other time we read text, we are engaging with the product of another mind. We are so used to the idea of text as a representation of another person’s thoughts that we have come to mistake their writing for their thoughts. But they aren’t. Text and media are tools that authors and artists create to let people change their own state of mind—hopefully in specific ways to form the image or effect the author was after.

Reading is an indirect collaboration with the author, mediated through the writing. Text has no inherent reasoning or intelligence. Agatha Christie’s ghost does not inhabit the words of Murder on the Orient Express. Stephen King isn’t hovering over you when you read Carrie. The ghost you feel while reading is an illusion you’ve made out of your own experience, knowledge, and imagination. Every word you read causes your mind to reconstruct its meaning using your memories and creativity. The idea that there is intelligence somehow inherent in writing is an illusion. The intelligence is all yours, all the time: thoughts you make yourself in order to make sense of another person’s words. This can prompt us to greatness, broaden our minds, inspire new thoughts, and introduce us to new concepts. A book can contain worlds, but we’re the ones that bring them into being as we read. What we see is uniquely our own. The thoughts are not transported from the author’s mind and injected into ours.

The words themselves are just line forms on a background with no inherent meaning or intelligence. The word “horse” doesn’t come with the Platonic ideal of a horse attached to it. The word “anger” isn’t full of seething emotion or the restrained urge towards violence. Even words that are arguably onomatopoeic, like the word “brabra” we use in Icelandic for the sound a duck makes, are still incredibly specific to the cultures and context they come from. We are the ones doing the heavy lifting in terms of reconstructing a picture of an intelligence behind the text. When there is no actual intelligence, such as with ChatGPT, we are the ones who end up filling in the gaps with our memories, experience and imagination.

When ChatGPT demonstrates intelligence, that comes from us. Some of it we construct ourselves. Some of it comes from our inherent biases.

There is no ‘there’ there. We are alone in the room, reconstructing an abstract representation of a mind. The reasoning you see is only in your head. You are hallucinating intelligence where there is none. You are doing the textual equivalent of seeing a face in a power outlet.

This drive—anthropomorphism—seems to be innate. Our first instinct when faced with anything unfamiliar—whose drives, motivations, and mechanisms we don’t understand—is to assume that they think much like a human would. When that unfamiliar agent uses language like a human would, the urge to see them as near or fully human is impossible to resist—a recurring issue in the history of AI research that dates all the way back to 1966.

These tools solve problems and return fluent, if untruthful, answers, which is what creates such a convincing illusion of intelligence.

Text synthesis engines like ChatGPT and GPT-4 do not have any self-awareness. They are mathematical models of the various patterns to be found in the collected body of human text. How granular the model is depends on its design and the languages in question. Some of the tokens—the smallest unit of language the model works with—will be characters or punctuation marks, some of them will be words, syllables, or even phrases. Many language models are a mixture of both.

With enough detail—a big enough collection of text—these tools will model enough of the probabilistic distribution of various words or characters to be able to perform what looks like magic:

  • They generate fluent answers by calculating the most probable sequence of words, at that time, which would serve as the continuation of or response to the prompt.
  • They can perform limited reasoning tasks that correlate with textual descriptions of prior reasoning tasks in the training data.

With enough of these correlative shortcuts, the model can perform something that looks like common sense reasoning: its output is text that replicates prior representations of reasoning. This works for as long as you don’t accidentally use the wrong phrasing in your prompt and break the correlation.

The mechanism behind these systems is entirely correlative from the ground up.What looks like reasoning is incredibly fragile and breaks as soon as you rephrase or reword your prompt. It exists only as a probabilistic model of text. A Generative AI chatbot is a language engine incapable of genuine thought.

These language models are interactive but static snapshots of the probability distributions of a written language.

It’s obviously interactive, that’s the whole point of a chatbot. It’s static in that it does not change when it’s used or activated. In fact, changing it requires an enormous amount of computing power over a long period of time. What the system models are the distributions and correlations of the tokens it records for the texts in its training data set—how the various words, syllables, and punctuation relate to each other over as much of the written history of a language as the company can find.

That’s what distinguishes biological minds from these algorithmic hindsight factories: a biological mind does not reason using the probability distributions of all the prior cultural records of its ancestors. Biological minds learn primarily through trial and error. Try, fail, try again. They build their neural network, which is functionally very different from what you see in a software model, through constant feedback, experimentation, and repeated failure—driven by a chemical network that often manifests as instinct, emotion, motivation, and drive. The neural network—bounded, defined, and driven by the chemical network—is constantly changing and responding to outside stimuli. Every time an animal’s nervous system is “used”, it changes. It is always changing, until it dies.

Biological minds experience. Synthesis engines parse imperfect records of experiences. The former are forward-looking and operate primarily in the present, sometimes to their own detriment. The latter exist exclusively as probabilistic manifestations of imperfect representations of thoughts past. They are snapshots. Generative AI are themselves cultural records.

These models aren’t new bird brains—new alien minds that are peers to our own. They aren’t even insect brains. Insects have autonomy. They are capable of general problem-solving—some of them dealing with tasks of surprising complexity—and their abilities tolerate the kind of minor alterations in the problem environment that would break the correlative pseudo-reasoning of a language model. Large Language Models are something lesser. They are water running down pathways etched into the ground over centuries by the rivers of human culture. Their originality comes entirely from random combinations of historical thought. They do not know the ‘meaning’ of anything—they only know the records humans find meaningful enough to store. Their unreliability comes from their unpredictable behaviour in novel circumstances. When there is no riverbed to follow, they drown the surrounding landscape.

The entirety of their documented features, capabilities, and recorded behaviour—emergent or not—is explained by this conceptual model of generative AI. There are no unexplained corner cases that don’t fit or actively disprove this theory.

Yet people keep assuming that what ChatGPT does can only be explained as the first glimmer of genuine Artificial General Intelligence. The bird brain of Silicon Valley is born at last!

Because text and language are the primary ways we experience other people’s reasoning, it’ll be next to impossible to dislodge the notion that these are genuine intelligences. No amount of examples, scientific research, or analysis will convince those who want to maintain a pseudo-religious belief in alien peer intelligences. After all, if you want to believe in aliens, an artificial one made out of supercomputers and wishful thinking feels much more plausible than little grey men from outer space. But that’s what it is: a belief in aliens.

It doesn’t help that so many working in AI seem to want this to be true. They seem to be true believers who are convinced that the spark of Artificial General Intelligence has been struck.

They are inspired by the science fictional notion that if you make something complex enough, it will spontaneously become intelligent. This isn’t an uncommon belief. You see it in movies and novels—the notion that any network of sufficient complexity will spontaneously become sentient has embedded itself in our popular psyche. James Cameron’s skull-crushing metal skeletons have a lot to answer for.

That notion doesn’t seem to have any basis in science. The idea that general intelligence is an emergent property of neural networks that appears once the network reaches sufficient complexity, is a view based on archaic notions of animal intelligence—that animals are soulless automata incapable of feeling or reasoning. That view that was formed during a period where we didn’t realise just how common self-awareness (i.e. the mirror test) and general reasoning is in the animal kingdom. Animals are smarter than we assumed and the difference between our reasoning and theirs seems to be a matter of degree, not of presence or absence.

General reasoning seems to be an inherent, not emergent, property of pretty much any biological lifeform with a notable nervous system.

The bumblebee, despite having only a tiny fraction of the neurons of a human brain, is capable of not only solving puzzles but also of teaching other bees to solve those puzzles. They reason and have a culture. They have more genuine and robust general reasoning skills—that don’t collapse into incoherence at minor adjustments to the problem space—than GPT-4 or any large language model on the market. That’s with only around half a million neurons to work with.

Conversely, GPT-3 is made up of 175 billion parameters—what passes for a “neuron” in a digital neural network. GPT-4 is even larger, with some estimates coming in at a trillion parameters. Then you have fine-tuned systems such as ChatGPT, that are built from multiple interacting models layered on top of GPT-3.5 or GPT-4, which make for an even more complex interactive system.

ChatGPT, running on GPT-4 is, easily a million times more complex than the “neural network” of a bumblebee and yet, out of the two, it’s the striped invertebrate that demonstrates robust and adaptive general-purpose reasoning skills. Very simple minds, those belonging to small organisms that barely have a brain, are capable of reasoning about themselves, the world around them, and the behaviour of other animals.

Unlike the evidence for ‘sparks’ of AGI in language models, the evidence for animal reasoning—even consciousness—is broad, compelling, and encompasses decades of work by numerous scientists.

AI models are flawed attempts at digitally synthesising neurologies. They are built on the assumption that all the rest—metabolisms, hormones, chemicals, and senses—aren’t necessary for developing intelligence.

Reasoning in biological minds does not seem to be a property that emerges from complexity. The capacity to reason looks more likely to be a built-in property of most animal minds. A reasoning mind appears to be a direct consequence of how animals are structured as a whole—chemicals, hormones, and physical body included. The animal capacity for problem-solving, social reasoning, and self-awareness seem to increase, unevenly, and fitfully with the number of neurons until it reaches the level we see in humans. Reasoning does not ‘emerge’ or appear. Some creatures are better at it than others, but it’s there in some form even in very small, very simple beings like the bumblebee. It doesn’t happen magically when you hook up a bunch of disparate objects together in a complex enough network. A reasoning mind is the starting point of biological thinking, not the endpoint that only “emerges” with sufficient complexity.

The internet—a random interconnected collection of marketing offal, holiday snaps, insufferable meetings, and porn—isn’t going to become self-aware and suddenly acquire the capacity for general reasoning once it reaches a certain size, and neither will Large-Language-Models. The notion that we are making autonomous beings capable of Artificial General Intelligence just by loading a neural network up with an increasingly bigger collection of garbage from the internet is not one that has any basis in anything we understand about biology or animal reasoning.

But, AI companies insist that they are on the verge of AGI. Their rhetoric around it verges on the religious as the idea of an AGI is idealised and almost worshipped. They claim to be close to making a new form of thinking life, but they refuse to release the data required to prove it. They’ve built software that performs well on the arbitrary benchmarks they’ve chosen and claim are evidence of general intelligence, but those tests prove no such thing and have no such validity. The benchmarks are theatrics that have no applicability towards demonstrating genuine general intelligence.

AI researchers love to resurrect outdated pseudoscience such as phrenology—shipping AI software that promises to be able to tell you if somebody is likely to be a criminal based on the shape of their skull. It’s a field where researchers and vendors routinely claim that their AIs can detect whether you’re a potential criminal, gay, a good employee, liberal or conservative, or even a psychopath, based on “your face, body, gait, and tone of voice.”

It’s pseudoscience.

This is the field and the industry that claims to have accomplished the first ‘spark’ of Artificial General Intelligence?

Last time we saw a claim this grand, with this little scientific evidence, the men in the white coats were promising us room-temperature fusion, giving us free energy for life, and ending the world’s dependence on fossil fuels.

Why give the tech industry the benefit of the doubt when they are all but claiming godhood—that they’ve created a new form of life never seen before?

As Carl Sagan said: “extraordinary claims require extraordinary evidence.”

He didn’t say “extraordinary claims require only vague insinuations and pinky-swear promises.”

To claim you’ve created a completely new kind of mind that’s on par with any animal mind—or, even superior—and provides general intelligence using mechanisms that don’t resemble anything anybody has ever seen in nature, is by definition the most extraordinary of claims.

The AI industry is backing their claims of Artificial General Intelligence with hot air, hand-waving, and cryptic references to data and software nobody outside their organisations is allowed to review or analyse.

They are pouring an every-increasing amount of energy and work into ever-larger models all in the hope of triggering the ‘singularity’ and creating a digital superbeing. Like a cult of monks boiling the oceans in order to hear whispers of the name of God.

It’s a farce. All theatre; no evidence. Whether they realise it or not, they are taking us for a ride. The sooner we see that they aren’t backing their claims with science, the sooner we can focus on finding safe and productive uses—limiting its harm, at least—for the technology as it exists today.

After everything the tech industry has done over the past decade, the financial bubbles, the gig economy, legless virtual reality avatars, crypto, the endless software failures—just think about it—do you think we should believe them when they make grand, unsubstantiated claims about miraculous discoveries? Have they earned our trust? Have they shown that their word is worth more than that of independent scientists?

Do you think that they, with this little evidence, have really done what they claim, and discovered a literal new form of life? But are conveniently unable to prove it because of ‘safety’?

Me neither.

The notion that large language models are on the path towards Artificial General Intelligence is a dangerous one. It’s a myth that directly undermines any effort to think clearly or strategise about generative AI because it strongly reinforces anthropomorphism.

That’s when you reason about an object or animal as if it were a person. It prevents you from forming an accurate mental model of the non-human thing’s behaviour. AI is especially prone to creating this reaction. Software such as chatbots trigger all three major factors that promote anthropomorphism in people:

  1. Understanding. If we lack an understanding of how an object works, our minds will resort to thinking of it in terms of something that’s familiar to us: people. We understand the world as people because that’s what we are. This becomes stronger the more similar we perceive the object to be to ourselves.
  2. Motivation. We are motivated to both seek out human interaction and to interact effectively with our environment. This reinforces the first factor. The more uncertain we are of how that thing works, the stronger the anthropomorphism. The less control we have over it, the stronger the anthropomorphism.
  3. Sociality. We have a need for human contact and our tendency towards anthropomorphising objects in our environment increase with our isolation.

Because we lack cohesive cognitive models for what makes these language models so fluent, feel a strong motivation to understand and use them as they are integrated into our work, and, increasingly, our socialisation in the office takes on the very same text conversation form as a chatbot does, we inevitably feel a strong drive to see these software systems as people. The myth of AGI reinforces this—supercharges the anthropomorphism—because it implies that “people” is indeed an appropriate cognitive model for how these systems behave.

It isn’t. AI are not people. Treating them as such is a major strategic error as it will prevent you from thinking clearly about their capabilities and limitations.

Believing the myth of Artificial General Intelligence makes you incapable of understanding what language models today are and how they work.