The Chinchilla orthodoxy is cracking
For three years, the field operated under a simple rule: compute and data should scale together. Train on roughly 20 tokens per parameter, keep the ratio consistent, and you get predictable capability gains. Chinchilla made this intuitive and empirically defensible.
GPT-4o breaks that story in ways that are still being absorbed.
The model achieves significantly better performance than similarly-sized predecessors not by training longer on more data, but by rethinking the relationship between pretraining compute, post-training refinement, and inference-time reasoning.
Multimodal scaling is different
The key insight from GPT-4o's design is that text, image, and audio don't scale the same way. A model trained jointly across modalities doesn't simply learn a weighted average of capabilities — it develops representations that are qualitatively different from the sum of its parts.
This means the scaling laws derived from text-only pretraining are insufficient models for multimodal systems. New compute allocation frameworks are needed, and nobody has published them yet.
What inference-time compute changes
The emergence of chain-of-thought reasoning as a first-class training objective — rather than an emergent property — fundamentally changes how we should think about scaling.
If a model can "think longer" at inference time and improve its answers, the relevant question isn't just "how many parameters?" but "how much compute are you willing to spend at inference?" These are different optimization axes.
The implications for deployment
Production teams are discovering that smaller, more capable models with inference-time reasoning often outperform brute-force scale on real tasks. This shifts the economics of AI deployment dramatically — fewer GPUs for training, but more complex inference infrastructure.
The era of "just scale it" is giving way to something more nuanced: architectural choices, data curation, and post-training refinement matter as much as raw compute. The teams that understand this will have a significant advantage in 2026.