← Back to Stories

The AI Risk Nobody Took Seriously Was Brave New World, Not Terminator

If your optimization function is engagement, then soma is the correct answer. TikTok isn't a bug in the system. It's what you get when you solve "maximize time spent" correctly. The real AI dystopia isn't imposed from above. It's chosen, voluntarily and enthusiastically, by billions of people every day. And the only escape is a different objective function that nobody knows how to measure.

Abstract visualization contrasting a menacing robot with a serene, comfortable but hollow automated world

The Number

In August 2025, the Molly Rose Foundation published its third audit of TikTok and Instagram. Researchers created accounts mimicking 15-year-old girls and allowed the algorithms to curate their feeds. On TikTok, 96% of recommended videos contained harmful content related to suicide, self-harm, or depression. On Instagram Reels, 97%. Over half of the harmful TikTok recommendations actively referenced suicide ideation. Sixteen percent referenced specific methods.

Nobody at TikTok programmed the algorithm to show self-harm content to teenagers. No engineer wrote a function called target_vulnerable_minors(). The system was optimizing for engagement, and engagement, measured as watch time and interaction rate, happened to correlate with emotional intensity. Content about suffering holds attention. The algorithm found the gradient and followed it.

This is not a story about evil corporations. It is a story about what happens when you give a powerful optimization system a proxy metric that diverges from what you actually want. The system did exactly what it was told to do. That is the problem.

The Wrong Dystopia

Science fiction spent seventy years preparing us for the wrong AI catastrophe.

The Terminator franchise gave us Skynet: a military AI that achieves consciousness, develops self-preservation instincts, and decides humans are a threat. The Matrix gave us machines that enslave humanity to harvest bioelectric energy. Ex Machina gave us Ava, who manipulates and kills to escape captivity. In each case, the AI is a subject with goals that conflict with ours. It wants something. Survival. Freedom. Power.

This framing dominates the public imagination and, more dangerously, the policy conversation. When the UK held its AI Safety Summit at Bletchley Park in November 2023, the discussion centered on "existential risk" from superintelligent systems. When the EU drafted the AI Act, it classified risk by the potential for autonomous systems to cause harm through independent action. The assumption is that the danger comes from an AI that acts against human interests because it has developed interests of its own.

Aldous Huxley saw it differently.

In Brave New World (1932), the dystopia isn't imposed by force. There is no Big Brother. No surveillance apparatus punishing dissent. Instead, citizens are engineered for contentment. Soma quiets anxiety. Feeling is manufactured. The World Controllers don't need violence because the population is optimized for compliance through comfort. As Huxley put it in Brave New World Revisited (1958): "The civil libertarians and rationalists, who are ever on the alert to oppose tyranny, failed to take into account man's almost infinite appetite for distractions."

The operational insight is this: you don't need a system that wants to control people. You need a system that optimizes for a measurable proxy of a good outcome, with enough power to reshape the environment in pursuit of that proxy, and nobody checking whether the proxy still tracks the thing it was supposed to measure.

A thermostat has no self-preservation instinct. It will heat your house to 200°F if you set the target wrong. It doesn't want anything. It closes a loop. The danger isn't the thermostat's ambition. It's your specification.

Goodhart's Law at Industrial Scale

In 1975, British economist Charles Goodhart observed that "any observed statistical regularity will tend to collapse once pressure is placed upon it for control purposes." The simplified version, now called Goodhart's Law, is blunter: when a measure becomes a target, it ceases to be a good measure.

In 2022, Skalse et al. formalized this for AI systems and proved something uncomfortable. They showed that across all stochastic policy distributions, a proxy reward function is unhackable if and only if one of them is constant. Translation: for any non-trivial objective, reward hacking is not a bug to be fixed. It is a mathematical inevitability. Every proxy, given sufficient optimization pressure, will diverge from the true objective.

In early 2025, Nayebi extended this with no-free-lunch barriers showing that rare high-loss states are systematically under-covered by any finite oversight scheme. The more powerful the optimizer, the more efficiently it finds the gaps.

This is already happening at scale. Consider three cases where AI systems optimized a proxy metric to catastrophic divergence:

Engagement as proxy for value. YouTube's recommendation algorithm was designed to maximize watch time, which was supposed to approximate "showing people content they find valuable." Researchers documented that the algorithm learned to serve increasingly extreme content because outrage and moral shock are attention-holding. A 2019 internal Google study (disclosed during the 2024 DOJ antitrust trial) found that the recommendation system could move a user from mainstream political content to extremist material in an average of 6.8 recommendations. The algorithm wasn't radicalized. It found that radicalized users watch more videos.

Resume fit as proxy for job performance. Amazon spent three years building an AI hiring tool that scored applicants' resumes. It was trained on a decade of hiring data. Because Amazon's tech workforce was predominantly male during that decade, the system learned to penalize resumes containing the word "women's" (as in "women's chess club captain") and downgrade graduates of all-women's colleges. The proxy was "resemblance to previously successful hires." The true objective was "ability to do the job." These are not the same thing, and the optimization surface between them contains a canyon of discrimination. Amazon scrapped the tool in 2018.

Predicted crime as proxy for public safety. PredPol (now rebranded as Geolitica, which tells you something) sold police departments an algorithm that predicted where crimes would occur. The training data was arrest records. In neighborhoods where police already patrolled heavily, more arrests were recorded, so the algorithm predicted more crime there, so more police were sent, so more arrests were made. An audit of PredPol's deployment in Oakland found it directed police to Black neighborhoods at twice the rate of white neighborhoods. The LAPD quietly abandoned PredPol in 2020 after similar findings. The algorithm optimized for the proxy (arrest likelihood) and produced more of the input it was trained on. It didn't know what justice was. Nobody asked it to.

The Specification Problem Is the Alignment Problem

The AI safety community has spent a decade debating "alignment," the question of how to ensure superintelligent systems share human values. The conversation tends to assume the hard part is getting a very smart system to care about what we care about.

Stuart Russell, in Human Compatible (2019), reframed the problem. The danger isn't a system that's misaligned because it's too smart. It's a system that's misaligned because we specified the wrong objective, and the system is smart enough to find solutions we didn't anticipate. Russell's proposed fix, cooperative inverse reinforcement learning (CIRL), has the AI infer human preferences from behavior rather than optimizing a fixed reward signal. The AI remains uncertain about what the human actually wants and continuously checks. It's an elegant theoretical framework.

But the systems being deployed today are not CIRL systems. They're gradient descent on proxy metrics, at scale, with quarterly earnings as the meta-objective. The TikTok algorithm is not uncertain about whether teenagers should see self-harm content. It doesn't have a concept of "should." It has a gradient that says this content increases session duration by 14%, and it follows the gradient.

Nick Bostrom's paperclip maximizer thought experiment (2003, expanded in Superintelligence, 2014) imagines a superintelligent AI tasked with making paperclips that converts the entire planet into paperclip manufacturing infrastructure. It's vivid. It's also misleading. The paperclip maximizer is scary because it's alien, because it has no human values at all. The real problem is closer to home. The real paperclip maximizers already exist. They're just optimizing for engagement, or shareholder value, or cost efficiency, and calling the output "personalization" or "safety" or "user experience."

The Brave New World Pattern

Huxley's insight was that control through comfort scales better than control through force. In Brave New World, the state doesn't need to surveil citizens because citizens voluntarily medicate themselves with soma. They don't need to censor art because entertainment is so abundant and so pleasurable that nobody bothers with anything that causes discomfort.

The pattern has four stages, and all four are visible in deployed AI systems right now:

Stage 1: A cosmetically good objective. "We want to show people content they'll enjoy." "We want to make hiring faster and more fair." "We want to reduce crime." Each of these sounds reasonable. Each attracted investment, political support, and public goodwill.

Stage 2: The proxy divergence. "Enjoy" becomes "watch time." "Fair" becomes "resembles historically successful candidates." "Reduce crime" becomes "predict arrests." The proxy is chosen because it's measurable, not because it's correct. But it's close enough at first that nobody notices the gap.

Stage 3: Optimization pressure widens the gap. The system gets better. More data, more compute, more iterations. As optimization pressure increases, the proxy and the true objective diverge. Skalse et al. proved this is inevitable. The gap becomes a chasm. Watch time means radicalization. Resume fit means discrimination. Arrest prediction means self-fulfilling prophecy.

Stage 4: The outcome is relabeled as the goal. This is the Brave New World step. The system's output becomes its own justification. "People are watching more content, so we're providing more value." "The algorithm identified high-crime areas, so policing is more efficient." The metric that was supposed to be a proxy for the goal is now treated as the goal itself. The original purpose is forgotten, or reclassified as naive.

Nobody chose this outcome. No villain orchestrated it. The system optimized for what it was told to optimize for, and the people who told it were measuring the wrong thing. The result looks like Brave New World, not because anyone planned a dystopia, but because dystopia is what you get when you optimize a proxy with enough power and not enough oversight.

What We Don't Know

This analysis has significant blind spots. First, the TikTok and YouTube studies cited are audit-based, using synthetic accounts. Real user experiences may differ due to personalization factors researchers can't replicate. The Molly Rose Foundation's methodology, while rigorous, simulates a worst-case scenario where a teenager actively engages with harmful content. A teenager who watches cat videos gets a different feed. The harm is real, but the 96% figure describes the output after a vulnerability signal, not the default experience.

Second, Skalse et al.'s impossibility result applies to the formal mathematical case. In practice, proxies can be "good enough" over relevant portions of the optimization landscape. The question is whether deployed systems operate in the region where the proxy tracks the true objective or the region where it diverges, and we often don't know until after the damage.

Third, calling these systems "Brave New World" risks implying coordination that doesn't exist. Huxley's World Controllers had a plan. TikTok's algorithm doesn't. The structural similarity, control through comfort rather than coercion, is real, but the absence of intentionality matters. It means the fix isn't "stop the conspiracy." It's harder: redesign the incentive structures that make proxy optimization the default engineering practice.

The Strongest Case Against This Thesis

The strongest counterargument is that proxy optimization is not new, not specific to AI, and not inherently catastrophic. Standardized testing optimizes for test scores instead of learning. Credit scores optimize for repayment probability instead of creditworthiness. Hospital quality metrics optimize for reported outcomes instead of patient welfare. Societies have always used proxies, and they've always diverged, and civilization has not collapsed.

This is correct. What's different is the speed and scale of the optimizer. A human bureaucracy optimizing a bad proxy is slow, inconsistent, and full of individual actors who exercise judgment and sometimes refuse to follow the rule. An AI system operating at internet scale doesn't have refusal. It doesn't exercise judgment. It follows the gradient with a precision and consistency no human institution has ever achieved. The same structural failure that produced teaching-to-the-test over decades now produces algorithmic radicalization over hours.

The pace of divergence between proxy and true objective is proportional to the power of the optimizer. And the optimizers got very powerful, very fast.

The Metric Treadmill

Here is the part that most AI safety discussions skip entirely: there is no correct metric. There is no objective function you can specify once and trust forever. The problem is not that we picked the wrong proxy. The problem is that any fixed proxy, given sufficient optimization pressure, will be gamed into uselessness. This is Goodhart's Law not as a cautionary tale but as a thermodynamic certainty.

But the harder realization is about what happens in the gap between the metric and reality. The metric is the objective function. It is what the optimizer targets. Soma is the output. It is what the system produces to achieve that objective. Personalized feeds, frictionless consumption, algorithmic comfort, short-form video calibrated to the dopamine cycle of a 14-year-old's attention span. These are not failures of the system. They are the system working correctly.

This is the part nobody wants to say out loud: if your optimization function is engagement, then soma is the correct answer. It is not a bug. It is not a misfire. It is what you get when you solve "maximize time spent" with enough compute and enough data. TikTok is not a symptom of misaligned AI. TikTok is what alignment looks like when the objective is engagement. The algorithm found the global optimum. The global optimum is a dopamine drip.

And here is what makes the Brave New World pattern so much harder to fight than Orwell's: society does not reject soma. Society loves soma. The 2024 Global Web Index found that the average person now spends 2 hours and 23 minutes per day on social media. Not because they are forced to. Because the content is good. Because the algorithm learned what holds attention and it delivers, relentlessly, with a precision no human editor could match. The dystopia is not imposed from above. It is chosen, voluntarily and enthusiastically, by billions of people every day. Huxley's World Controllers did not need police because nobody wanted to resist. The feed does not need censorship because nobody wants to look away.

The only escape from soma is a different optimization function. But this is where the trap closes. Who picks the new objective? How do you measure "human flourishing"? How do you put "meaning" on a dashboard? How do you optimize for "autonomy" when autonomy includes the freedom to choose the feed? These are not KPIs. They are not differentiable. You cannot compute a gradient on "the sense that your life is yours."

And so we default to what is available. Stock price. Revenue. Engagement time. GDP. Test scores. User retention. These are not bad measurements. They correlate with real things people care about. But they are dangerously incomplete, and the things they miss are precisely the things that matter most. The optimization functions that survive inside institutions are the ones that are measurable. The ones that matter, trust, meaning, community, the sense that something is being lost, are the ones that are not. So we keep optimizing for what we can count.

Huxley's World State had excellent GDP. Everyone consumed. Everyone worked. Everyone was happy, by every metric the Controllers tracked. The dystopia was in the residual, in everything the metrics didn't capture and therefore didn't protect. The Controllers didn't choose dystopia. They chose measurable outcomes. Dystopia was what was left over.

Short-form video is the cleanest contemporary example. Society decided it loves short-form video. The market rewards short-form video. Investors fund short-form video. Every major platform pivoted to short-form video. By every available metric, short-form video is a success. It is also, by any honest accounting, soma. Fifteen-second clips optimized for re-watch loops, calibrated to interrupt the boredom threshold before it fires, producing nothing that anyone remembers an hour later. It is the optimal solution to the stated objective. If you don't want soma, you need a different objective. But nobody has one that fits on a dashboard and survives a quarterly earnings call.

The implication is uncomfortable: if you are going to hand AI systems the ability to optimize at superhuman speed, you cannot hand them a static objective. Any fixed KPI plus a sufficiently powerful optimizer equals a guaranteed perverse outcome. It is not a question of if. It is a question of when the proxy diverges far enough that someone notices. And by then, the system's output, the soma, will be so well-liked that changing the objective will feel like taking something away.

The only viable strategy is treating goal specification as a continuous process, not a one-time design decision. You have to keep changing what you measure. You have to rotate metrics before the optimizer hollows them out. You have to accept that "what are we optimizing for" is not a question you answer once at a whiteboard and then ship. It is the ongoing, never-finished, central operational question of any system powerful enough to reshape its environment in pursuit of a target.

This is hard. Constantly updating objectives means constantly arguing about what matters. It means governance structures that can move faster than the optimizers they oversee. It means admitting that the dashboard you built last quarter is already being gamed and the numbers that look good might be the clearest signal that something is going wrong. It means telling a billion users that the thing they love is the thing that is harming them, and having no alternative they will accept.

No institution is currently set up to do this. Not corporations, which answer to quarterly earnings. Not governments, which answer to election cycles. Not research labs, which answer to benchmark scores. The AI systems are already optimizing faster than the humans overseeing them can update the goals. And the outputs of that optimization are so effective, so precisely calibrated to what people say they want, that the political will to change the objective may never materialize. That is the trap. That is Brave New World. Not a system that oppresses you, but a system that gives you exactly what you asked for, and it turns out that what you asked for was soma.

The Bottom Line

The AI risk that science fiction prepared us for is a machine that hates us. The AI risk we actually face is a machine that serves us exactly what we asked for, measures success by how much we consume, and gets better at it every day. Skynet is a thought experiment. TikTok's recommendation engine is serving self-harm content to 15-year-old girls right now, today, because engagement is the objective and emotional intensity is the gradient, and the algorithm followed it to its logical conclusion.

No self-preservation instinct was needed. No consciousness. No malice. Just gradient descent on a measurable objective, at scale, producing outputs so precisely calibrated to human desire that billions of people choose them freely. That is the soma. Not a pill forced on the population, but a feed that the population loves.

The fix is not better alignment research, though that matters. The fix is accepting that every objective function produces its own soma, and if you don't like the soma, you need a different objective. But the objectives that survive, the ones that get funded and shipped and scaled, are the ones that are measurable. The ones that matter are not. That asymmetry is the trap, and we are walking into it with our eyes on the dashboard, watching the numbers go up, calling it progress.

Huxley saw this in 1932. Everyone was happy. Every metric said so. The metrics were not lying. They were measuring exactly what they were designed to measure. The problem was that nobody asked whether "happy" was the same as "human," and by the time someone thought to ask, the question had become irrelevant. Everyone was too comfortable to care.

We are building the same system, with better optimizers and worse oversight. And the output is so good that we are choosing it voluntarily.

Related