The DeepSeek V4 follow-up coverage that landed Friday and Saturday is denser than the initial Thursday drop and has shifted the framing of the model. Latent Space's AINews newsletter on Saturday morning leads with the hardware-sovereignty angle: V4 in both its sizes — Pro at one point six trillion total parameters with forty-nine billion active and Flash at two hundred eighty-four billion total with thirteen billion active — is now confirmed runnable on Huawei Ascend chips, not just on NVIDIA. That detail is being read as the practical end of the export-control bottleneck for Chinese frontier deployments, since Ascend supply is domestic and the V4 inference stack appears to be hand-tuned for the Ascend memory hierarchy. Latent Space frames the release as DeepSeek stepping out of the benchmark-leadership game — V4 is no longer claiming to top the public leaderboards — and into a compute-economics game, where the relevant comparison is intelligence per dollar at deployment. The prodigal-tiger-returns metaphor in the headline captures both: DeepSeek skipped a quarter of public model news entirely between V three point two Speciale in November and now, then dropped two preview models with full open weights and aggressive pricing in one week.
MIT Technology Review's Friday piece, Three Reasons Why DeepSeek's New Model Matters, takes a different cut. The first reason is the size milestone — at one point six trillion total parameters under MIT license, V4 Pro is the largest open-weights model ever released, larger than Kimi K two point six at one point one trillion and twice the size of V three point two. The second is the disclosed training cost, which DeepSeek continues to claim is roughly an order of magnitude below comparable Western runs; the report notes the same caveats every analyst has been raising, namely that the figure excludes prior research compute and likely undersells the real total, but also notes that even a charitable correction leaves DeepSeek noticeably ahead on cost efficiency. The third reason, and the most consequential for downstream deployments, is pricing: V4 Flash at fourteen cents per million input tokens and twenty-eight cents output now sits below GPT five point four Nano, Gemini three point one Flash Lite, and Claude Haiku four point five on every dimension, and V4 Pro at one dollar seventy four input and three forty eight output is roughly a fifth the price of GPT five point four and an order of magnitude cheaper on output than Claude Opus four point seven or GPT-5.5. The piece closes on the export-control angle Latent Space also flags — that the marginal effect of restrictions on Chinese AI capability has diminished sharply.
AI Explained's Friday video synthesizes both threads under a single 'compute war intensifies' frame. The argument is that the dominant axis of model competition has shifted from one-dimensional intelligence scores to intelligence-per-dollar Pareto curves, and that V4 has now planted a flag at the cost-efficient frontier in a way no closed lab can match without disclosing pricing. The video also highlights the qualitative gap that remains — V4's tool-use, agentic-coding, and reasoning behaviours are described as roughly Claude four point five class, not Opus four point seven or GPT-5.5 class, but the open-weights property and the cost differential more than compensate for many production workloads. Together the three takes form the consensus view that V4 is the new floor for what 'good enough' open-weights frontier-class deployment looks like, and that the next round of pricing pressure on US-hosted APIs is now baked in.
- Latent Space emphasized that V4 is now runnable on Huawei Ascend chips — a hardware-sovereignty signal beyond the parameter count.
- MIT Tech Review framed V4 around three reasons it matters: training-cost claims, MIT-licensed open weights at frontier scale, and pricing pressure on US-hosted APIs.
- AI Explained placed V4 alongside GPT-5.5 in a single 'compute war intensifies' frame, arguing the relevant axis is now intelligence-per-dollar Pareto curves rather than headline benchmarks.