>
Another One Bites the Dust: Former Iranian Foreign Minister Kamal Kharazi Dies from Injuries...
Department of Justice Launches Investigation Into the NFL: Report
Only Iran "Friendly" Ships Allowed Transit Through Strait, As Tankers Pile Up Near Hormuz.
FCC Set To "Supercharge" Starlink Space Internet With "Seven-Fold More Capacity"
'World's First' Humanoid Robot For Real Household Chores Launched With 16-Hour Battery
XAI Training 10 Trillion Parameter Model – Likely Out in Mid 2026
The $7 Powder That Beats Your $5,000 AC Unit!
Private credit is now a $3 trillion asset class and investors are receiving 45 cents on the dollar
Converting Diesel Vehicles to Run on Waste Vegetable Oil, by Polar Bear
Anthropic says its latest AI model is too powerful for public release and that it broke...
The CIA used a futuristic new tool called "Ghost Murmur" to find and rescue...
This Plant Replaces All Fertilizer FOREVER. Why Did the FDA Ban It?
China Introduces Pistol-Like Coil-Gun Based On Electromagnetic-Launch Systems

Grok Imagine V2
2 variants of 1-trillion-parameter models
2 variants of 1.5-trillion-parameter models
1 variant of a 6-trillion-parameter model
1 variant of a 10-trillion-parameter model
xAI would be leading in raw announced scale of parameters. No other lab has publicly confirmed training 10T or even 6T models right now. The 6T model alone is roughly double the rumored size of Grok 4 and far larger than most current estimates for GPT-5 or Claude 4.6.
Parameter count is only part of the story.
AI models are judged more on:
Active parameters per token (MoE efficiency).
Training data quality and "intelligence density" (xAI claims higher density per gigabyte).
Inference-time compute (reasoning modes, multi-agent orchestration).
Real-world benchmarks (coding, agentic tasks, multimodality).
Chips Needed & Costs for Pre-Training Runs
Exact per-model costs are not public (models are still training), but here are the best analyses and estimates.
Colossus 2 hardware: ~550,000 NVIDIA GPUs (mostly GB200/GB300 Blackwell variants) at ~$18 billion hardware cost alone (average ~$32k–$40k per GPU). This supports the full parallel training lineup.
Total CapEx is tens of billions of dollars for Colossus 2 (land, power infrastructure, cooling, networking). Includes on-site gas turbines/Megapacks for 400+ MW dedicated power and rapid buildout.
Per-model rough estimates (community/analyst extrapolations).
10T model needs ~$1.5 billion+ in compute (one early analyst call. scales with FLOPs and duration). Initial pre-training phase ~2 months on Colossus 2.
6T model needs Similar order of magnitude but lower. benefits from shared cluster efficiency.
Smaller 1T/1.5T runs: Significantly cheaper/faster due to parallelization.
