How AI Thinks: What Autocomplete Can Teach Us About Bias and Creativity

Headshot
November 24, 2024

When we think of artificial intelligence, it’s tempting to attribute human-like characteristics to it—intuition, judgement, or even creativity. But fundamentally, most AI models, including popular ones like ChatGPT, don’t really “think” in the way humans do. Instead, they rely on something far simpler but surprisingly powerful: predicting the next word in a sequence based on everything that has come before. This is known as next-token prediction.

Breaking Down AI’s “Mind”: The Mechanics of Autocomplete

At its core, a large language model (LLM) like ChatGPT operates much like an advanced version of your phone’s autocomplete. When you give an AI a prompt, it calculates which word, or token, is statistically most likely to follow, based on all the data it has been trained on. It’s a process of probability, where the AI uses vast amounts of human language data to predict what should come next in a given sentence.

For instance, if you type the phrase, “The sky is…” the AI has learned that the next most probable token is “blue,” though alternatives like “clear,” “overcast,” or even “beautiful” could also follow depending on the context and the randomness built into the model. The AI doesn’t know the sky is blue—it’s simply learnt from its training data that this is a common pairing.

Yet, small changes in a prompt can lead to drastically different outcomes. Changing “The sky is…” to “In winter, the sky is…” or “In poetry, the sky is…” would dramatically shift the AI’s likely responses, steering it towards other common word associations like “grey,” “endless,” or even “metaphor.”

The Implications: Bias, Misunderstandings, and Divergence

Understanding how AI predicts tokens sheds light on why even small differences in language can produce drastically varied outputs. More importantly, it reveals where and how biases might creep into AI-generated content. Since the AI’s predictions are based on patterns in its training data, the responses it produces reflect those patterns—along with all the biases, assumptions, and gaps that exist within the data itself.

For example, the style or language you use when asking a question can influence how an AI responds. If the model was primarily trained on formal, academic data, it might give less accurate or relevant responses to queries phrased informally or colloquially. This means that two users asking similar questions could receive quite different answers depending on how their prompts are worded, inadvertently introducing biases based on language styles.

Creativity Through Pattern Recognition

This brings us to another interesting tension: if AI is essentially an autocomplete machine, how can it produce creative or seemingly novel results? While AI models don’t “create” in the human sense of having an intention or an imaginative leap, their capacity for creativity emerges from their ability to mix, blend, and extend patterns in unpredictable ways.

When asked to write a story or suggest ideas, the AI’s output isn’t an original insight, but a reflection of the diverse content it’s been trained on. Creativity, in this case, is the product of randomness and variation within those patterns. Small changes in prompts, or even differences in word choice or syntax, can push the AI towards generating more experimental or unconventional outputs.

For instance, asking ChatGPT to write a story in the style of Edgar Allan Poe taps into the stylistic patterns it has absorbed from Poe’s works during training. It can produce something that feels original because it’s recombining familiar stylistic elements in new contexts. Yet, this creativity is limited by the model’s training data and lacks the true inventiveness or emotional resonance of human creativity.

Why Understanding AI’s Mechanics Matters

Grasping the mechanics of next-token prediction isn’t just an exercise in technical understanding; it’s crucial for navigating and managing the risks associated with AI. Knowing how AI arrives at its responses can help users avoid common pitfalls like inadvertently leading the AI down an unhelpful path or trusting biased or inaccurate outputs. It can also inform better prompt engineering, pushing AI models to generate more useful and varied results.

As AI becomes increasingly embedded in our workflows, from automated content generation to more complex decision-making tasks, it’s essential to demystify how it operates. Only by understanding the mechanics of how AI “thinks” can we start to anticipate its limitations, prevent biases, and harness its potential for innovation and creativity.

Final Thoughts: The Value of Simplicity

The idea that sophisticated AI models boil down to predicting the next word might seem overly simplistic, but it reveals the surprisingly elegant core of what makes them powerful. AI’s capacity to mirror and extend human language, while maintaining the flexibility to experiment with different outputs, offers vast potential for innovation. Yet, it also comes with significant risks that must be understood and managed carefully.

By demystifying AI’s decision-making, we can start to explore not just how to get the most out of it, but also how to use it responsibly and with a critical eye. Understanding the mechanics allows us to question, refine, and reimagine what AI can achieve—both creatively and practically.