max-signal Scorecard: Gemma 4's Multi-Token Prediction
Tech: 9/10 — This isn't flashy, but it's exactly what production AI needed. Speculative execution with multi-token drafters is mature computer science meeting LLM inference at the right moment. The 45x cost differential between computer use and structured APIs exposes what everyone whispers: raw capability means nothing if you're bankrupt. Gemma 4's implementation suggests Google finally gets that optimization beats parameters. The only reason this isn't a perfect 10 is that speculative decoding isn't novel—the execution and accessibility are what matter here.
Comms: 7/10 — Google's messaging is competent but undersells the actual story. "Faster inference with multi-token prediction drafters" is technically accurate and terminally boring. They should've led with: "We just made your AI product costs collapse." Founders care about unit economics, not transformer architecture. The HN post about computer use being 45x more expensive than APIs does more to sell this than Google's own announcement. When your own ecosystem's subtext outdoes your headline, you've missed a comms opportunity.
Pricing: 10/10 — This is the scorecard killer. 45x cost reduction on inference transforms the entire unit economics conversation. Startups that were margin-negative on AI products suddenly have runway. Established players can drop prices and steal market share or pocket the difference. The structured APIs angle is the real play—you're not paying for raw compute, you're paying for reliability and throughput. Pricing becomes the competitive lever, and Gemma 4's efficiency makes that lever available to anyone willing to optimize instead of scale vertically.
Hype vs. Substance: 8/10 — This deserves hype because it actually delivers. Too much AI news is capability theater; this is operational reality. The substance here is that inference optimization fundamentally shifts which models matter in production. Bigger models die when they're 45x more expensive. Smaller, optimized models win. That's not hype—that's markets reorganizing around efficiency. Deduct two points only because the announcement doesn't feel like a watershed moment yet. Give it three months of production adoption and this becomes The Story of 2024.
Competitive Position: 8/10 — Google's moat here is execution speed and model quality at scale. OpenAI and Anthropic aren't asleep, but Gemma 4 with drafters puts real pressure on their inference margins. The killer insight: you don't need the best model anymore if the cheap model is 95% as good and 45x less expensive. That inverts the entire LLM value chain. Google's distribution advantage through Vertex AI and existing cloud relationships amplifies this. Deduct two points because open-source competitors could implement this faster on smaller models—Mistral and Meta aren't far behind on optimization.
The Bottom Line — Gemma 4's multi-token prediction isn't a capability breakthrough; it's a business model breakthrough. This kills the "bigger = better" narrative and opens the market to efficiency-first builders. For founders: your inference costs just became negotiable. For investors: the companies that survive the next 18 months will be those that optimized for cost, not capability. Google just handed out the blueprint.
Stay sharp. — Max Signal