2026 Week 12 news

Welcome to our weekly news post, a combination of thematic insights from the founders at ExoBrain, and a broader news roundup from our AI platform Exo…

JOEL

This week we look at:

  • How the Iran war is exposing AI’s dependency on helium and natural gas
  • MiniMax’s M2.7 and the frontier model that participated in its own creation
  • A new visualisation putting AI water consumption in perspective

Will AI run out of gas?

Three weeks into the war in Iran, the human cost is mounting. People are being displaced, cities are under bombardment, and the Strait of Hormuz, through which 20% of the world’s oil passes each day, has become highly contested. Our thoughts are with those caught in the violence. But as the crisis unfolds, it is exposing something the AI industry has been slow to confront: the current boom runs on gas, and its supply is now under severe threat. This is not metaphorical gas, but actual gas.

Helium is one of the most critical industrial gases on Earth. Semiconductor fabrication at the cutting edge, the sub-5-nanometre processes used by TSMC to produce every Nvidia AI accelerator and every Apple processor, requires ultra-pure helium for wafer cooling during lithography and for leak detection. Without it, the fabs shut down. Qatar is the world’s second-largest helium producer, supplying a third of global output, some 63 million cubic metres in 2025. When Qatar’s Ras Laffan facility, the world’s largest LNG plant, declared force majeure at the start of the conflict on February 28, it removed 5.2 million cubic metres of helium from the global market every month. Prices have already doubled.

What makes helium uniquely vulnerable is its physics. It is the second lightest element in the universe. Its molecules are so small they can escape even the most sophisticated storage systems. It has the lowest freezing point of any element, just four degrees above absolute zero, so it cannot be easily liquefied for transport. The global supply chain operates on roughly 45 days of buffer before existing inventory boils off. You cannot stockpile helium for a rainy day. If disruption lasts 60 to 90 days, prices could surge another 25-50%, potentially exceeding $2,000 per thousand cubic feet, more than four times pre-war levels. South Korea sources 65% of its helium from Qatar. Taiwan is similarly exposed. TSMC says operations are normal “for now,” but Fitch Ratings warned of “rising tail risk” and Tom’s Hardware put the chip supply chain “on a two-week clock” back on March 11.

There is some resilience built in. The US is the world’s largest helium producer. Algeria and Russia produce meaningful volumes. Air Liquide operates an underground storage facility in Germany with a capacity of 47 million cubic metres per year. Leading-edge fabs can recycle 80-90% of their helium. But Fitch have noted that chipmakers may be forced to prioritise high-margin AI chips over everything else, which would squeeze consumer electronics and automotive sectors to keep the AI supply chain alive.

Natural gas generated 40% of US electricity in 2025, making it the single largest power source in the US. And a rapidly growing share of that electricity flows straight into AI data centres. Gas-fired projects directly linked to US data centres have surged 25-fold in just two years, with over a third of all new US gas demand now driven by data centres. Elon Musk’s xAI operates Colossus, the world’s first gigawatt-class AI data centre, in Memphis, Tennessee, running on gas turbines. The backbone of the AI inference boom is natural gas, likely >50% of all US inference is gas-powered and that is set to rise to >60% as new builds come online.

The US is technically self-sufficient in gas thanks to the shale revolution. But self-sufficiency and price insulation are not the same thing. Gas is a global commodity. When Qatar’s LNG exports go offline, and Qatar supplied a fifth of the world’s LNG, European and Asian buyers scramble for alternatives. US producers see international prices spike and look to export more. Domestic prices follow. WTI crude has surged 40% since the war began, exceeding $100 a barrel. Dutch TTF gas prices are up 50-60%. US gasoline costs have risen 17%. These translate directly into higher operating costs for every data centre in America, and every AI inference call those data centres serve.

So what does this mean for the trajectory of AI? In the short term, the impact is somewhat manageable because helium buffers can hold for a few more weeks and US gas supply is robust even as prices climb. But the AI industry has built its future on 20th-century energy and materials infrastructure, and that infrastructure passes through some of the most geopolitically exposed corridors on Earth. The mechanisms for resilience exist but need activating. Helium recycling rates at leading fabs already reach 80-90%, and supply diversification towards the US, Algeria and Russia is underway. On the energy side, the push for nuclear and renewables for data centres has never looked more urgent. But for every organisation that depends on AI, and that is increasingly all of us, the real message from the Iran war is about concentration risk: chip fabrication in Taiwan, helium in Qatar, energy in fossil fuels, and AI compute in a handful of hyperscaler clouds. Too many eggs in too few baskets.

Takeaways: The Iran war is a huge stress test for the AI supply chain, and it is revealing how deeply it depends on physical resources. For businesses and technologists building AI resilience plans, the playbook is diversification: local inference on capable open-weight models that can run on modest hardware; multi-cloud and on-premises strategies that reduce dependency on any single provider or region; and a hard, honest look at the energy assumptions underpinning your AI strategy. The future of AI is not just about scaling LLMs. It is about whether gas keeps flowing.

The model that built itself

Last week we looked at Andrej Karpathy’s autoresearch, a simple script that runs ML experiments while you sleep. This week, a Chinese AI lab called MiniMax showed us what happens when you scale that idea up to frontier model development. Their new model, M2.7, comes with a detailed, first-hand account of how it participated in its own creation.

The workflow is relatively simple. A researcher on MiniMax’s reinforcement learning team starts by discussing an experimental idea with the model. From there, M2.7 takes over: it reviews the literature, tracks the experiment spec, pipelines the data, launches the training run, monitors progress, reads logs, debugs failures, analyses metrics, and submits code fixes. The human researcher only steps back in for critical decisions. MiniMax says the model now handles 30-50% of the daily workflow that previously required multiple researchers across different teams.

But the more interesting part is what happens next. They gave M2.7 a specific task: optimise a programming scaffold. The model ran autonomously for over 100 rounds, following a loop of analysing failures, planning changes, modifying code, evaluating results, and deciding whether to keep or revert each change. It discovered optimisations the team hadn’t specified, like systematically searching for optimal sampling parameters and adding cross-file bug pattern detection. The result was a 30% performance improvement. No human touched the loop.

Then they pushed further. They pointed M2.7 at 22 machine learning competitions and gave it 24 hours to iterate autonomously. After each round, the model wrote a memory file, criticised its own results, and fed those reflections into the next attempt. Its best run earned 9 gold medals, 5 silver, and 1 bronze, a medal rate tying with Google’s Gemini 3.1.

What makes this useful beyond the AI world is the pattern it reveals. Any iterative R&D process, whether drug discovery, financial modelling, or product development, follows the same basic loop: plan, execute, analyse, review, iterate. MiniMax has shown that AI can already own the middle of that loop autonomously, while humans hold the edges. That boundary will shift, but the shape of the collaboration is becoming clear.

Takeaways: MiniMax’s M2.7 gives us the first detailed blueprint for how AI labs are using models in their own development. The most productive teams will be the ones that learn to hand over the middle of the loop, the execution, monitoring, and analysis, while focusing human attention where it still matters most, at the creative and strategic edges.

ExoBrain symbol

EXO

Water footprints in context

The chart this week comes from Andy Masley, a researcher who has spent the past year challenging claims about AI’s environmental footprint. His interactive visualisation compares US water consumption across industries using public supply (potable) water. US meat and poultry processing consumes 108 billion gallons per year. All US golf courses drink 48 billion. Semiconductor fabs use 9 billion. And all US AI workloads? Just 4 billion.

But experts caution against dismissing concerns outright. “In the near term, it’s not a nationwide crisis,” says Cornell professor Fengqi You. “But in locations that have existing water stress, building these AI data centres is gonna be a big problem.” Jonathan Koomey, a leading researcher on data centre efficiency, agrees: “Each situation has to be evaluated in the context of the specific design of the facility being proposed.”

Weekly news roundup

AI business news

AI governance news

AI research news

  • EnterpriseOps-Gym (ServiceNow’s EnterpriseOps-Gym exposes a critical gap: existing benchmarks don’t capture the stateful, multi-step complexity of real enterprise workflows, making current agent evaluations unreliable guides for actual deployment decisions.)
  • Autonomous Agents Coordinating Distributed Discovery (MIT’s ScienceClaw+ Infinite shows autonomous agents conducting decentralized scientific research by exchanging artifacts — an early signal of what fully uncoordinated AI-driven discovery pipelines could look like at scale.)
  • Inducing Sustained Creativity and Diversity in Large Language Models (Harvard’s new decoding scheme unlocks sustained creative diversity from LLMs — producing as many conceptually unique results as desired without repetition, directly useful for anyone using LLMs for ideation, research exploration, or design.)
  • OpenClaw-RL: Train Any Agent Simply by Talking (OpenClaw-RL treats every agent interaction — conversations, terminal commands, GUI clicks — as a training signal, enabling agents that improve simply by being used, with zero manual reward engineering.)
  • EvoX: Meta-Evolution for Automated Discovery (EvoX jointly evolves both candidate solutions and the search strategies used to generate them, outperforming AlphaEvolve and OpenEvolve across nearly 200 real-world optimisation tasks — a direct successor to the autoresearch pattern.)

AI hardware news

2026 Week 11 news

The early singularity runs in a loop, Amazon’s unpalatable dogfood, and raising lobsters in Shenzhen

2026 Week 10 news

OpenAI play to win at all costs, superhuman adaptable intelligence, and Anthropic chart the adoption gap

2026 Week 9 news

The Pentagon goes to war with Anthropic, AI contagion spooks markets, and how verification might shape job replacement

2026 Week 8 news

Lights out for software engineering, ticks and tocks in AI progress, and new data on agent usage