xAI confirmed that Colossus 2 has officially gone online, reaching a power capacity of 1 Gigawatt (GW). This follows the purchase of a third massive building in Memphis—dubbed “MACROHARDRR”—which allowed xAI to scale its infrastructure to roughly 555,000 NVIDIA GPUs at a single site.

Hardware Specs: The 555,000 GPU Powerhouse
Colossus 2 is not just larger than its predecessor; it is fundamentally more powerful, transitioning from “Hopper” (H100) architecture to the latest Blackwell chips.
- GPU Mix: The cluster primarily consists of 520,000 NVIDIA GB200 units and 30,000 GB300 units (the latest Blackwell variant).
- Total Investment: The hardware alone is estimated at $18 billion, with each GPU rack utilizing liquid-cooling technology to manage the extreme thermal output of a gigawatt load.
- Compute Power: The system is estimated to deliver up to 50 ExaFLOPS of compute power, making it roughly twice as powerful as the original Colossus 1 cluster.
Solving the Power Problem: On-Site Generation
To achieve the 1GW milestone without crashing the local Memphis grid, Musk utilized a “self-contained” power strategy:
- Gas-Fired Power Plant: xAI constructed a private gas-fired power plant adjacent to the data center to bypass traditional utility interconnection queues.
- Tesla Megapacks: A massive array of Tesla Megapacks acts as a giant battery buffer, managing the intense “power swings” that occur during large-scale AI training runs.
- Water Access: The facility utilizes the Mississippi River watershed for a mandatory liquid-cooling system capable of dissipating 1.8 GW of heat.
| Metric | Colossus 1 (2024) | Colossus 2 (2026) |
| Power Capacity | ~250 MW | 1,000 MW (1 GW) |
| GPU Count | 200,000 (H100/H200) | 555,000 (GB200/GB300) |
| Primary Goal | Grok 3 & 4 Training | Grok 5 & AGI Research |
| Build Time | 122 Days | 19-Day Expansion Cycles |
The Road to Grok 5 and AGI
The primary purpose of Colossus 2 is to begin the pre-training of Grok 5. With over half a million Blackwell GPUs, xAI aims to slash training times from months to weeks.
- AGI Ambitions: Musk has hinted that the scale of Colossus 2 provides the “raw silicon” necessary to achieve Artificial General Intelligence (AGI) by the end of 2026.
- The “Muskonomy” Synergy: Beyond Grok, the supercomputer provides backend support for Tesla’s FSD (Full Self-Driving) v13+ and SpaceX’s flight trajectory simulations.
Conclusion: A New Blueprint for AI
The launch of Colossus 2 proves that the bottleneck for AI is no longer just software, but power and speed. By compressing a four-year data center construction timeline into mere months and generating its own electricity, xAI has created a new template for “frontier” AI training. As the 1GW site goes full-throttle, the rest of the industry—including Microsoft’s “Stargate” project—is now racing to match this single-site concentration of power.