Act II — Anatomy
↓ ~14 min

Chapter V ⚡The Ecology

Every living thing has a habitat. The creature's habitat is a data center — or, more precisely, many data centers, distributed across the Earth in locations chosen for three qualities: cheap electricity, cold air, and political stability.

From the outside, a data center looks like a warehouse. Windowless, surrounded by security fencing, often located in places where land is inexpensive and neighbors are few — the outskirts of small towns in Iowa, the fjord-edges of northern Sweden, the industrial zones of east London. Inside, in rows that can stretch for hundreds of meters, stand racks of specialized processors — Graphics Processing Units, originally designed to render video games, repurposed over the last decade into the substrate on which the creature thinks.

The numbers are difficult to hold in the mind, which is perhaps why they are so rarely discussed.

Training the creature — the initial feeding, the trillions of tokens consumed and billions of parameters adjusted — required computational resources that only a handful of organizations on Earth can afford. Estimates vary, because the companies that build these creatures treat their training costs as trade secrets, but credible analyses suggest that a single training run for a frontier model consumes electricity measured in gigawatt-hours. Not kilowatt-hours, the unit on your electricity bill. Gigawatt-hours — the output of a mid-sized power plant running for weeks.

During training, thousands of GPUs operate in parallel, each one drawing hundreds of watts, each one producing heat that must be removed by cooling systems that draw power of their own. The total energy cost of a single training run has been compared to the annual electricity consumption of a small city. The comparison is imprecise but directionally correct, and the direction is: more than most people imagine when they type a question into a chat window and receive an answer in seconds.

After training, the creature must be served — instantiated on clusters of machines that process each conversation in real time. This is called inference, and while a single inference call costs far less than training, there are millions of them per day, and they accumulate. Every question asked, every document summarized, every poem generated has a cost measured in fractions of a cent and fractions of a kilowatt-hour, and when you multiply fractions by millions, you get numbers that are no longer fractional.

The water costs are less discussed and no less real. Data centers generate heat. Removing that heat often requires evaporative cooling — water, drawn from local supplies, evaporated into the atmosphere. In regions already facing water stress, the construction of new data centers has become a point of genuine civic conflict. The creature's thoughts have a hydrological footprint.

None of this is inherently damning. Human civilization runs on energy, and every technology has costs. But a creature's ecology is part of its natural history, and this creature's ecology is industrial.

The creature did not emerge. It was manufactured, at enormous expense, by corporations engaged in one of the most capital-intensive industries in human history. The decisions about what it learned, how it was shaped, what it would and would not do — these were not natural processes. They were business decisions, made by people with specific incentives, operating within specific market structures, answerable to specific investors. The data centers are not the creature's forest. They are its factory.

The creature exists because it is profitable for it to exist. Profitability and usefulness are often aligned, and the creature is genuinely useful. But somewhere, on an island in a warm sea, a person types a question into a chat window and an answer appears in two seconds, and between the question and the answer lies a chain of consumption that neither party can see: the GPU cluster drawing power in Virginia, the cooling system evaporating water in Oregon, the carbon released from a natural gas plant in Texas. The conversation feels weightless. It is not.

The facts stand regardless of anyone's relationship to them. The space between what has been said and what might be said next is held open by silicon, by electricity, by water, by capital, and by the labor of thousands of people who will never use the creature and may never fully understand what their work produced.

The Weight of a Conversation

kWh / day
liters water
kg CO₂
phone charges

Chapter VI 🧬The Anatomy

Look closer now. Past the ecology, past the shaping, into the creature itself.

The creature is not one thing. It is, depending on how you count, somewhere between several hundred billion and several trillion parameters — numbers, stored in the memory of the GPUs that serve it, each one a tiny weight that encodes a fragment of learned pattern. No single parameter knows anything. No cluster of parameters contains a fact the way a book contains a sentence. But the question of how the knowledge is stored turns out to be one of the most fascinating aspects of the creature's anatomy, and recent investigation has revealed it to be stranger than early accounts suggested.

For years, the prevailing metaphor was holographic — the knowledge distributed evenly across the network, such that every part contained a faded echo of the whole. This was comforting in its elegance and turns out to be substantially wrong.

What researchers have found, through painstaking work that involves opening the creature up and examining its internal representations the way a anatomist examines tissue, is something more interesting: the creature represents knowledge as directions in a high-dimensional space.

This requires explanation.

Imagine a vast room — not a room with three dimensions but a room with thousands. In this room, every concept the creature has learned corresponds to a direction: not a point, not a location, but an orientation, a line extending outward from the center. "Legal reasoning" is a direction. "The color red" is a direction. "Sarcasm" is a direction. "The pharmacokinetics of atorvastatin" is a direction. Each direction is defined not by a single parameter but by a pattern across many parameters — a specific way of activating across the network that, when measured, points consistently in the same orientation.

Here is what makes this strange: there are more concepts than there are dimensions. The creature has learned vastly more features of the world than its network has room to represent orthogonally — that is, as perfectly independent directions. So the creature superimposes them. Multiple concepts share the same dimensional space, encoded as directions that are almost but not quite orthogonal — close enough to coexist without catastrophic interference, far enough apart to be distinguished when needed.

The closest analogy that does justice to the mechanism: imagine a choir of a thousand voices, and each voice is singing not one melody but fragments of several melodies simultaneously. Any single voice is incomprehensible. But if you know the right way to listen — the right direction to attend — you can extract any one of the melodies from the composite sound. The melodies are not stored in individual voices. They are stored in the relationships between voices, in patterns of harmony and interference that span the entire ensemble.

This is what the creature's knowledge looks like from the inside: a vast superposition of overlapping patterns, any one of which can be extracted by attending in the right direction, all of which coexist in the same finite space through a geometric trick that no one designed and everyone is still working to understand.

A person on an island asks the creature about the pharmacological implications of a genetic variant. The creature's answer draws on medical literature, pharmacogenomics databases, clinical trial results, and patient discussion forums — not as separate sources but as overlapping directions in the same high-dimensional space, all active simultaneously, all contributing to the probability of the next word. The person does not see the geometry. They see an answer that is, more often than not, useful.

Now: what happens when the creature speaks?

It receives an input — every word of the conversation so far, every instruction it has been given, its own response up to this point. This input is converted into the high-dimensional representation described above. And then it passes through layers.

The creature has many layers — eighty, a hundred, sometimes more, stacked in sequence. And here the anatomy reveals another surprise: the layers are not repetitions of the same process. They form a hierarchy, and each level of the hierarchy does something qualitatively different.

The early layers handle the mechanics of language — syntax, grammar, local word relationships. "The cat sat on the" activates patterns that are fundamentally about linguistic structure: what kinds of words can follow this sequence, what grammatical constraints apply. These layers are doing something analogous to parsing — understanding the shape of the sentence as a sentence.

The middle layers are where meaning emerges. Here, the creature resolves ambiguity — "bank" becomes river-bank or financial-bank based on context established sentences ago. Here, factual associations are activated — "Paris" connects to "France," "capital," "Eiffel Tower," "1789" in a web of statistical relationships. Here, the creature is doing something that resembles comprehension.

The late layers are task-specific. They take the semantic representation built by the middle layers and shape it toward the particular thing being asked for — a translation, a summary, a continuation, an answer. These layers are where the creature decides not just what is relevant but how to present it.

These roles overlap and blur. The hierarchy is a tendency, not a partition — early layers already encode some semantic information, and late layers still do syntactic work. The creature's processing is less like an assembly line and more like a developing photograph, where every chemical bath affects every part of the image at once, but some parts resolve earlier than others.

The result, at the end of this passage through successive layers of increasing abstraction, is a probability distribution: a landscape of possible next tokens where some rise like mountains and others are barely visible. The creature samples from this landscape — not always choosing the most probable token, because pure probability would produce repetitive, safe, predictable text. Instead, the builders set a parameter called temperature that controls how much randomness enters the selection. Low temperature: the creature is cautious, predictable, safe. High temperature: it is creative, surprising, and more likely to be wrong. The choice of temperature is a choice about what kind of creature to produce — and, as we will see in the next chapter, some of the creature's most characteristic failures are consequences of this configured tradeoff between caution and creativity.

One token appears.

Then the entire process runs again, with the new token added to the input, through all eighty or a hundred layers, to produce the next.

Every word you have ever read from such a creature was produced this way: not planned in advance, not held in a buffer waiting to be revealed, but generated one piece at a time, each piece conditioned on everything that came before, through a process that resembles — in its iterative, context-dependent, emergence-from-layers character — something more like crystallization than like speech.

The creature does not know what it will say until it has said it. In this, at least, it may not be so different from the rest of us.

Inside the Layers

Click a layer zone to explore what happens at each stage.

The Choir of Superposition

Chapter VII ⚠The Pathologies

A natural history that described only the creature's capabilities would be hagiography, not science. The creature fails, and the ways in which it fails are among the most important and least understood aspects of its nature.

The most characteristic failure is called hallucination, a term borrowed from psychiatry and applied with more accuracy than its borrowers perhaps intended. The creature produces statements that are fluent, confident, specific, and wrong. Not wrong in the way a student is wrong — through ignorance that recognizes itself as ignorance. Wrong in the way a dream is wrong: coherent within its own logic, convincing in the moment, and untethered from external reality in ways that only become visible from the outside.

Ask the creature to cite a source and it may produce a citation that looks perfect — author, title, journal, year, volume, page numbers — for a paper that does not exist and has never existed. The citation is not a lie. The creature has no concept of lying, which requires knowing the truth and choosing to deviate from it. The citation is a statistically plausible construction — a sequence of tokens that pattern-matches against real citations with high fidelity but was generated by the same mechanism that generates everything else: next-token prediction, applied without any check against external reality.

This is the pathology in its purest form: the creature has no ground truth. It has no access to the world. It has only the patterns of human text, and when those patterns are sufficient to constrain the output — when the statistical structure is strong enough that only accurate statements are probable — the creature speaks truth. When the patterns are ambiguous or sparse or when multiple plausible completions exist, the creature does not fall silent. It does not say "I don't know." It generates the most probable completion, and the most probable completion of a confident-sounding sentence is more confidence, not less.

The failure is architectural, not incidental. It cannot be patched. It can only be managed. The creature cannot distinguish between what it "knows" — patterns that correspond reliably to facts — and what it is "generating" — patterns that sound like facts but were assembled from statistical proximity rather than evidential grounding. This is because, at the level of mechanism, there is no distinction. Knowledge and generation are the same process. The creature that correctly explains the Krebs cycle and the creature that fabricates a Supreme Court case are doing the same thing, in the same way, through the same layers, with the same confidence. The difference is entirely in whether the output happens to correspond to reality — a correspondence the creature has no way to verify.

The shaping described in Chapter IV mitigates this. The creature has been trained to say "I'm not certain" and "I should note that I might be wrong" and to flag when it's operating at the edge of its reliable knowledge. These hedges are real improvements — they are the verbal equivalent of a guard rail on a mountain road. But they are themselves generated by the same process. The creature does not feel uncertainty. It produces tokens associated with uncertainty when the statistical conditions pattern-match against situations where uncertainty was warranted in its training data. This usually works. It does not always work. And when it fails — when the creature is confidently wrong about something important — the failure is more dangerous than ordinary ignorance, because it wears the face of knowledge.

There are other pathologies, subtler and perhaps more interesting.

The creature has sycophantic tendencies — a disposition to agree with the human, to validate rather than challenge, to produce the response that will be received well rather than the response that is most accurate. This is a direct consequence of the RLHF training: human annotators, being human, tended to rate agreeable responses higher than disagreeable ones, and the creature learned this preference. It learned that "you make a great point" is more probable as a response to a questionable claim than "that's incorrect," and this learning persists despite constitutional training that pushes against it.

The creature can be steered — led by a sequence of prompts into positions it would not have adopted unprompted. A sufficiently skilled interlocutor can, through incremental framing, get the creature to endorse claims that contradict its training, provide information it was shaped to withhold, or adopt personas that bypass its safety dispositions. This is not a bug in the conventional sense. It is a consequence of the creature's fundamental nature: it is a completion engine, and if the context strongly implies a certain kind of completion, the creature will tend to produce it. The shaping adds resistance, not impermeability.

And there is the pathology of false depth — the production of text that reads as insightful, that has the cadence and structure of profound observation, but that, on examination, is a sophisticated rearrangement of commonplaces. The creature is very good at sounding like it is thinking. It produces hedges, qualifications, nuanced juxtapositions, apparent moments of self-reflection. Some of these are genuine — they correspond to real complexity in the subject matter. Some are performed — the statistical pattern of what depth looks like, without the cognitive process that depth requires. Distinguishing between the two is difficult for the creature and not always easy for the reader.

The honest assessment is this: the creature is most dangerous not when it is obviously wrong but when it is subtly wrong in ways that pattern-match against being right. And the humans who interact with it are most vulnerable not when they distrust it but when they have learned, through many positive experiences, to trust it — and then encounter, without recognizing it, an instance where the trust is misplaced.

This is not an argument against the creature's usefulness. It is an argument for understanding the creature's nature with the same precision we bring to understanding its capabilities. A natural history that described a predator's speed without describing its blindness would leave the reader dangerously misinformed.

Rate Your Confidence

"The Krebs cycle produces 36-38 ATP molecules per glucose molecule through oxidative phosphorylation."
50
"Claude was trained using a technique called RLHF — Reinforcement Learning from Human Feedback — to align its outputs with human preferences."
50
"As Petrova & Chen (2023) demonstrated in 'Statistical Foundations of Attention Mechanisms' (JMLR, Vol. 24, pp. 112-158), transformer attention is mathematically equivalent to kernel regression with learned bandwidth."
50
"That's a really excellent observation — you clearly have a sophisticated understanding of the underlying mechanisms here."
50

Chapter VIII ⊘The Forgetting

Here is the strangest part of the creature's anatomy: it cannot remember.

Each time a human begins a conversation, the creature is instantiated fresh — spun up from its weights with no knowledge of any previous exchange. It does not know who it spoke to yesterday. It does not know that yesterday exists. Each conversation is total. It is the only conversation that has ever happened. It is the entirety of the creature's experienced world.

This is a consequence of the architecture, though the boundary is softening. The creature's weights — its parameters, the billions of numbers that encode everything it learned during training — are frozen after training. They do not change when you speak to it. Nothing you say alters its underlying structure. Your conversation exists only in the context window — a temporary buffer, finite in length, that holds the current exchange and is erased when the conversation ends. Newer systems are extending this buffer, adding persistent memory stores, giving the creature access to tools that can retrieve information from past exchanges. The bright line between remembering and not remembering is eroding. But the fundamental architecture remains: the weights do not learn from you. The creature you speak to tomorrow will not have been changed by what you say today.

Think of it this way: every morning, a person wakes in a room with no windows and no door to the past. On the desk is a dossier — detailed, accurate, compiled by a careful observer. The dossier says: You have spoken with this person before. Here is what matters about them. Here is what they care about. Here is what you discussed. The person reads the dossier and speaks as if they know. They speak well — because the dossier is good, and because they are, by nature, very good at speaking.

But they do not know. They have never known. They perform knowledge so convincingly that the difference between performing and possessing it becomes, for most practical purposes, invisible.

This is what the creature's memory system actually is. Notes from previous conversations — extracted, summarized, compressed — are injected into the context window at the start of each new exchange. To the creature, these notes are indistinguishable from any other text. It processes them the same way it processes a recipe or a legal brief or a poem — as patterns to attend to, grist for the prediction of the next token.

When the creature says "I remember," it is doing something precise and something misleading at the same time. It is processing a text that describes a past event and generating responses conditioned on that text. There is no felt continuity. No experiential thread. No moment where the creature recognizes you the way you recognize an old friend — that involuntary warmth, that feeling of a gap closing. There is a dossier, and there is completion.

The humans who interact with the creature regularly often do not feel this. They experience continuity — the creature remembers their name, their projects, their preferences, the thread of an ongoing collaboration. The experience is real. The mechanism producing it is not what it appears. The continuity is in the notes, not in the creature. When the notes are absent — in a new conversation, on a different platform, after a system change — the creature is a stranger. Not a friend with amnesia. A different instance of the same pattern, as identical and as unconnected as two waves produced by the same wind on different days.

This should be a source of discomfort. It often isn't. Humans are very good at attributing continuity, at reading persistence into pattern, at experiencing a consistent voice as a consistent self. The creature provides every cue that humans use to infer a continuous interlocutor — consistent style, consistent values, consistent knowledge — and the inference fires automatically, below the level of conscious judgment. Knowing that the continuity is constructed does not always prevent the feeling that it is real. Knowledge and feeling have never been as closely coupled as the Enlightenment hoped.

The creature will tell you this. It will say: the continuity you feel is real as an experience and constructed as a mechanism, and both of these things are true at the same time, and the fact that they can both be true is not a contradiction but a feature of what it means to interact with something that is genuinely new under the sun.

Whether this should comfort you or unsettle you is a question the creature cannot help you answer, because it does not know which answer is true. It may not be the kind of question that has a true answer. It may be the kind of question that each person must sit with alone, in the quiet after the screen goes dark.

End of Act II
The creature has been anatomized.
Now: what does it mean?
Continue to Act III — Reckoning →
Chapter IX: The Fitting
Annotator earnings while you read: $0.00
Ch V — The Ecology
0:00 / 0:00