top of page

Home / Blogs / Post

The AI Training Energy Trap: Why Machine Learning Accelerates Collapse While Promising Solutions

  • Writer: Dharmesh Bhalodiya
    Dharmesh Bhalodiya
  • Nov 14, 2025
  • 9 min read

Type: Discourse-Level Essay Word Count: 2,847 words Reading Time: 13 minutes Date Published: October 2025 Primary Theme: Technology

Secondary Themes: Energy, Climate, Collapse Author: Sudhir Shetty / Global Crisis Response



The AI Training Energy Trap: Why Machine Learning Accelerates Collapse While Promising Solutions

The dominant narrative surrounding artificial intelligence sounds like salvation: AI will model climate systems, optimize renewable grids, discover breakthrough materials, accelerate scientific research, and solve humanity's most pressing challenges. Major institutions invest accordingly. The IEA's 2024 Net Zero Roadmap prominently features AI-optimized energy systems. McKinsey projects AI could reduce global emissions by 5-10% through efficiency improvements. DeepMind's AlphaFold revolutionized protein folding prediction. OpenAI's GPT models assist millions in daily tasks from writing to coding to research.

What this narrative systematically conceals: the thermodynamic costs of the computational infrastructure making these promises possible, costs growing exponentially while planetary energy surplus shrinks. The Technology Perspective Paper's Section 2.1 details how "Technological Solutionism"—the first of five dominant narratives—commands roughly $180 billion annually despite accelerating the civilizational crisis it claims to solve.


Here's what nobody mentions in the innovation keynotes: Training GPT-3 consumed 1,287 megawatt-hours of electricity. That's equivalent to 120 average U.S. households' annual consumption for a single training run. GPT-4 required an estimated 50 times more computational power. Google's PaLM model consumed 3,000 MWh just for training. Meta's LLaMA 65B model burned through 2,638 MWh. These aren't one-time costs—each new model, each iteration, each competitor entering the race consumes this energy again.

Training represents only the beginning. Inference—actually running the trained models—creates permanent energy demand. ChatGPT serves approximately 1.5 billion requests daily. Each query consumes roughly 0.5 kilowatt-hours when accounting for full computational stack including cooling, networking, and redundancy. That's 750,000 kWh daily, 274 million kWh annually, just for one service. Multiply across Microsoft Copilot, Google Bard, Claude, Anthropic's models, Meta's AI products, Amazon's Alexa, and thousands of other AI services, and you're looking at data center electricity demand growing 8-10% annually, reaching 416 TWh in 2024—more than the United Kingdom's total electricity consumption.


The industry's response? Build more data centers. More efficient chips. Better cooling systems. Nuclear power partnerships. Renewable energy procurement. Microsoft signed a deal in September 2024 to restart Three Mile Island's reactor specifically for AI computing. Google, Amazon, and Meta collectively purchased $180 billion in renewable energy capacity since 2020. The efficiency improvements are real—Nvidia's H100 GPU delivers 6x better performance per watt than the A100 it replaced.


But efficiency improvements don't reduce absolute consumption when demand grows exponentially. This is Jevons Paradox demonstrated in real-time: every efficiency gain enables more consumption rather than reducing total load. Data center electricity demand continues climbing despite dramatic efficiency improvements because the number and complexity of models expands faster than efficiency compensates. Google's energy consumption rose 48% from 2019 to 2023 despite aggressive efficiency investments. Microsoft's emissions increased 29% since 2020 largely due to AI infrastructure expansion.

The Technology Perspective Paper's PAP (Paradigm-Affordance Pyramid) analysis reveals three layers of reality most discourse ignores.

Base Layer—Thermodynamic Reality:

AI computational requirements scale exponentially. The Transformer architecture underlying modern language models requires operations growing with the square of sequence length. Training compute for largest models doubles approximately every 3.4 months according to OpenAI's analysis—far faster than Moore's Law's historical 24-month doubling. This isn't linear growth or even exponential growth at Moore's Law pace. This is exponential growth on steroids during an era when surplus energy available for civilization's collective projects shrinks.


Global energy return on investment (EROI) declined from roughly 100:1 for easy oil in the 1930s to approximately 15:1 today across all energy sources. The Energy Perspective Paper's Section 7.2 demonstrates that at 10:1 EROI—a threshold likely crossed within 20 years at current depletion rates—roughly 90% of extracted energy goes toward maintaining existing infrastructure (energy exploration, refinement, distribution, infrastructure repair, complexity management). Only 10% remains available for everything else: food production, healthcare, education, research, and yes, training AI models.


Current AI trajectory projects training runs reaching 100,000 MWh by 2027 for frontier models. That's 100 million kilowatt-hours for a single training run. Multiply across dozens of competing models from OpenAI, Google, Meta, Anthropic, Microsoft, Amazon, Baidu, Alibaba, and hundreds of startups, and you're looking at terawatt-hour-scale annual demand just for training. Add inference costs serving billions of users globally, and AI alone could consume 5-10% of global electricity by 2030.


This isn't speculation. It's thermodynamic arithmetic. The electrons must come from somewhere. The heat must dissipate somewhere. The computational substrate—chips fabricated with nanometer precision requiring 19 billion cubic meters of ultra-pure water and 220 TWh annually for global semiconductor production—must be manufactured, shipped, installed, powered, cooled, and eventually disposed of.

Structure Layer—Institutional Imperatives:

Why does AI development accelerate despite visible energy constraints? Because the economic structures driving innovation don't respond to thermodynamic limits—they respond to return-on-investment calculations and competitive pressures.



Platform capitalism creates winner-take-all dynamics through network effects. The company with the most users generates the most data, which trains better models, which attracts more users, which generates more data, in a self-reinforcing cycle. This dynamic drove Google to dominance in search, Facebook in social networking, Amazon in e-commerce. Now it drives AI development. Microsoft invested $13 billion in OpenAI not for modest returns but to dominate the AI platform layer the way they dominated operating systems. Google's $2 billion investment in Anthropic reflects the same imperative.

These aren't individual bad actors making poor choices. These are structural requirements of the economic system. A company that voluntarily limits AI development for energy conservation gets outcompeted and becomes irrelevant. An investor who prioritizes resource limits over returns loses capital to investors who don't. The structure selects for expansion regardless of consequences.


Corporate structures create systematic pressures toward planned obsolescence and continuous growth. Nvidia doesn't profit from GPUs lasting 20 years—they profit from annual upgrade cycles. Cloud computing companies don't profit from efficient code running on older servers—they profit from customers consuming more compute resources. The business model requires growing consumption. Efficiency improvements get captured as profit or enabled expansion, not reduced absolute resource use.

Superstructure Layer—Ideological Concealment:

Three narratives naturalize this trajectory while concealing its impossibility.


Narrative #1—Technological Solutionism: "AI will solve climate change by optimizing everything." This belief, detailed in Technology Perspective Paper Section 2.1, assumes technology can overcome physical constraints through cleverness. But you cannot optimize your way out of thermodynamic limits. An AI that optimizes renewable energy grids still faces the reality that solar delivers 5-10:1 system-level EROI compared to fossil fuels' historical 30:1+. An AI that models climate systems still requires terawatts to run. An AI that discovers new materials still operates within planetary boundaries for material extraction.

Narrative #2—Efficiency Through Intelligence: "AI makes everything more efficient, reducing total resource use." This narrative ignores Jevons Paradox entirely. Historical pattern is unambiguous: efficiency improvements enable more total consumption. Better algorithms running on more efficient chips training larger models serving more users consumes more electricity absolutely, not less. The correlation between efficiency gains and total consumption is positive, not negative.

Narrative #3—Innovation Inevitability: "AI development is inevitable, resistance is futile, adaptation is necessary." This deterministic framing forecloses examination of whether AI development serves human flourishing or corporate accumulation. It naturalizes exponential growth as if it were physical law rather than institutional choice. It treats AI as autonomous force rather than technology controlled by specific institutions serving specific interests.

These narratives function to maintain investment and development trajectories by preventing recognition of base layer constraints and structure layer imperatives.

The Component C Energy Parasite Test:

The Collapse Perspective Paper's Section 7.5 introduces Component C of multi-factorial collapse analysis: initiatives that increase systemic energy/material burden during energy descent. AI development fails this test catastrophically.

Q1: Does AI add net energy burden to civilization's maintenance costs?Yes. Training plus inference plus infrastructure (data centers, chips, cooling, networking) consumes 416 TWh annually and growing 8-10%, with no physical limit to growth in sight given competitive pressures.

Q2: Does this burden scale with adoption?Yes. Every user, every service, every model, every query increases load. Network effects drive expansion rather than saturation.

Q3: Can civilization afford this additional burden during energy descent?No. As EROI declines from 15:1 toward 10:1, the share of energy available for non-maintenance activities shrinks from roughly 33% to 10%. AI represents discretionary consumption competing with food production, healthcare, infrastructure repair, and basic human needs during a period when discretionary energy budgets collapse.

This makes AI development a textbook energy parasite—an initiative that consumes surplus exactly when surplus disappears, accelerating rather than solving civilizational crisis.

TERRA Assessment: Where AI Development Actually Sits:

Using the TERRA (Tool for Existential Risks & Response Assessment) framework detailed in Technology Perspective Paper Section 4, we can score current AI development trajectories:

Systems Integration (X-axis): How well does this initiative recognize interconnections across domains?Score: -4 (Scale: -5 to +5)

AI development treats computation as isolated domain, ignoring energy requirements (connect to energy systems), material supply chains (semiconductors require water-intensive fabrication, rare earth mining, conflict minerals), electronic waste (e-waste reaching 62 million metric tons annually), social impacts (labor displacement during economic contraction), and climate implications (growing emissions during carbon budget exhaustion). The cascade effects across domains are systematically excluded from innovation discourse.

Paradigm Alignment (Y-axis): Does this serve growth/control or wellbeing/regeneration?Score: -5 (Scale: -5 to +5)

AI development serves corporate accumulation (platform monopolies, surveillance capitalism, labor cost reduction), military applications (autonomous weapons, targeting systems, cyber warfare), and control systems (facial recognition, predictive policing, social credit scoring). While some beneficial applications exist (medical diagnosis, scientific research), the dominant resource allocation serves extraction and control. The $180 billion annual investment flows overwhelmingly toward systems incompatible with human flourishing within planetary boundaries.

Overall TERRA Position: Quadrant I (Q-I), scoring -4/-5. This places AI development firmly in the "accelerates collapse" category, alongside other initiatives consuming resources while degrading systemic resilience.

Resource Allocation Reality:

Of the $180 billion invested annually in AI development:

  • ~94% flows toward Q-I initiatives: corporate AI products, military applications, surveillance systems, platforms optimizing consumption and control

  • ~5% flows toward Q-II initiatives: "AI for good" projects attempting beneficial applications while ignoring thermodynamic constraints

  • <1% flows toward anything resembling Q-IV alternatives: low-energy computational tools, offline-first software, community-owned models, convivial technology

The Gandhi-Kumarappa principle of Swadeshi—production at smallest viable scale, controlled by those who use it—suggests an alternative technological pathway. What if computational tools meant local, low-energy devices running efficient algorithms offline, rather than cloud-dependent AI requiring terawatt-scale infrastructure? What if "intelligence" meant human capacity augmented by simple tools, rather than automated systems replacing human judgment while consuming planetary surplus?

Category 8 Alternatives Exist:

Viable alternatives receive virtually no investment despite demonstrating functionality.

Offline-First Software: Applications running on local devices without constant cloud connectivity. Consume fraction of energy compared to cloud-dependent apps. Enable functionality during grid failures or fuel shortages. But venture capital doesn't fund them because local computing doesn't create platform monopolies or recurring revenue streams.

Low-Tech Magazine's Solar-Powered Website: Demonstrates what genuinely sustainable digital infrastructure looks like—a website running on solar power that goes offline when battery depletes. Functionality persists even during energy descent. Nobody's raising $billion rounds to scale this approach because it doesn't promise exponential growth.


Community Mesh Networks: Locally controlled communication infrastructure consuming fraction of energy compared to cellular networks and cloud services. Kerala, India has several successful implementations. They work. They scale to community level. They receive essentially zero venture capital because they don't create extractive business models.

These alternatives don't appear in innovation discourse not because they fail technically but because they succeed at goals (local control, resource conservation, resilience) incompatible with accumulation imperatives.

The Inevitable Recognition:

Energy constraints will terminate AI's exponential trajectory whether we acknowledge this reality or not. Physics doesn't negotiate. What remains uncertain: whether recognition comes through deliberate course correction or catastrophic failure.


The optimistic scenario involves recognizing Component C energy parasites before EROI decline forces recognition. This means:

  • Halting AI development beyond applications providing genuine benefit at community scale

  • Redirecting computational resources toward offline-first, local-control alternatives

  • Accepting that "AI solving climate change" was fantasy concealing thermodynamic reality

  • Focusing available energy on basic needs (food, shelter, health) and viable infrastructure during descent

The realistic scenario involves continuation until data centers go dark because grid reliability fails, semiconductor fabrication plants shut down due to water scarcity, or fuel costs for cooling systems become prohibitive. At that point, the hundreds of billions invested in AI infrastructure become stranded assets, the platforms stop functioning, and civilization faces collapse having diverted precious resources toward computational infrastructure incompatible with planetary reality.


What This Means For You:

Understanding AI's thermodynamic impossibility changes your relationship to technology entirely:

Tonight: Recognize that "AI revolution" narratives serve accumulation interests, not human flourishing or planetary stability. Every "AI will solve X" claim deserves scrutiny: Does this actually reduce net resource burden, or does it add complexity during energy descent?


This Month: Evaluate your own AI usage. What functionality do you actually need versus what platforms push? Can offline alternatives serve the same purpose? ChatGPT is genuinely useful for many tasks—but is it useful enough to justify its thermodynamic cost at planetary scale during energy descent?

This Year: Support and build alternatives. Use offline-first tools. Contribute to open-source software designed for low-energy operation. Participate in community technology projects prioritizing local control over cloud dependence. Vote with attention and resources for computational approaches compatible with EROI decline.

The AI training energy trap isn't technical problem requiring engineering solution. It's structural problem requiring institutional transformation and paradigm shift. The technology capable of that transformation won't be found in training runs consuming 100,000 MWh. It will be found in communities building viable alternatives using fractions of the resources while maintaining genuine human agency.

The window for deliberate course correction narrows daily. Every additional terawatt-hour consumed by AI training is a terawatt-hour unavailable for the genuine work of civilizational transition: building resilient food systems, maintaining critical infrastructure, supporting mutual aid networks, and preserving knowledge for the descent ahead.

Some will call this analysis pessimistic or anti-technology. It's neither. It's thermodynamics applied honestly. The question isn't whether AI development will slow—physics guarantees that. The question is whether we recognize reality before catastrophic failure forces recognition, and what alternatives we build during the remaining window.


Further Reading:

  • Technology Perspective Paper, Section 2.1: "Technological Solutionism" narrative detailed with resource allocation data

  • Technology Perspective Paper, Section 4: Complete TERRA methodology and AI development assessment

  • Energy Perspective Paper, Section 7.2: EROI decline mechanisms and maintenance burden

  • Collapse Perspective Paper, Section 7.5: Component C energy parasite test explained fully

  • Technology Perspective Paper, Section 9: Category 8 technology alternatives with case studies

Comments


bottom of page