DeepSeek v4

api-docs.deepseek.com

414 points by impact_sy 3 hours ago


simonw - an hour ago

I like the pelican I got out of deepseek-v4-flash more than the one I got from deepseek-v4-pro.

Flash: https://gist.github.com/simonw/4a7a9e75b666a58a0cf81495acddf...

Pro: https://gist.github.com/simonw/9e8dfed68933ab752c9cf27a03250...

Both generated using OpenRouter.

For comparison, here's what I got from DeepSeek 3.2 back in December: https://simonwillison.net/2025/Dec/1/deepseek-v32/

And DeepSeek 3.1 in August: https://simonwillison.net/2025/Aug/22/deepseek-31/

And DeepSeek v3-0324 in March last year: https://simonwillison.net/2025/Mar/24/deepseek/

nthypes - 2 hours ago

https://huggingface.co/deepseek-ai/DeepSeek-V4-Pro/blob/main...

Model was released and it's amazing. Frontier level (better than Opus 4.6) at a fraction of the cost.

rohanm93 - 3 minutes ago

This is shockingly cheap for a near frontier model. This is insane.

For context, for an agent we're working on, we're using 5-mini, which is $2/1m tokens. This is $0.30/1m tokens. And it's Opus 4.6 level - this can't be real.

I am uncomfortable about sending user data which may contain PII to their servers in China so I won't be using this as appealing as it sounds. I need this to come to a US-hosted environment at an equivalent price.

Hosting this on my own + renting GPUs is much more expensive than DeepSeek's quoted price, so not an option.

gardnr - a minute ago

865 GB: I am going to need a bigger GPU.

yanis_t - an hour ago

Already on Openrouter. Pro version is $1.74/m/input, $3.48/m/output, while flash $0.14/m/input, 0.28/m/output.

fblp - 2 hours ago

There's something heartwarming about the developer docs being released before the flashy press release.

seanobannon - 2 hours ago

Weights available here: https://huggingface.co/deepseek-ai/DeepSeek-V4-Pro

sidcool - an hour ago

Truly open source coming from China. This is heartwarming. I know if the potential ulterior motives.

mchusma - an hour ago

For comparison on openrouter DeepSeek v4 Flash is slightly cheaper than Gemma 4 31b, more expensive than Gemma 4 26b, but it does support prompt caching, which means for some applications it will be the cheapest. Excited to see how it compares with Gemma 4.

luew - 5 minutes ago

We will be hosting it soon at getlilac.com!

storus - 31 minutes ago

Oh well, I should have bought 2x 512GB RAM MacStudios, not just one :(

zargon - 2 hours ago

The Flash version is 284B A13B in mixed FP8 / FP4 and the full native precision weights total approximately 154 GB. KV cache is said to take 10% as much space as V3. This looks very accessible for people running "large" local models. It's a nice follow up to the Gemma 4 and Qwen3.5 small local models.

sibellavia - 8 minutes ago

A few hours after GPT5.5 is wild. Can’t wait to try it.

tariky - 13 minutes ago

Anyone tried with make web UI with it? How good is it? For me opus is only worth because of it.

gbnwl - 2 hours ago

I’m deeply interested and invested in the field but I could really use a support group for people burnt out from trying to keep up with everything. I feel like we’ve already long since passed the point where we need AI to help us keep up with advancements in AI.

jessepcc - 2 hours ago

At this point 'frontier model release' is a monthly cadence, Kimi 2.6 Claude 4.6 GPT 5.5, the interesting question is which evals will still be meaningful in 6 months.

CJefferson - 28 minutes ago

What's the current best framework to have a 'claude code' like experience with Deepseek (or in general, an open-source model), if I wanted to play?

Aliabid94 - 2 hours ago

MMLU-Pro:

Gemini-3.1-Pro at 91.0

Opus-4.6 at 89.1

GPT-5.4, Kimi2.6, and DS-V4-Pro tied at 87.5

Pretty impressive

clark1013 - 26 minutes ago

Looking forward to DeepSeek Coding Plan

jdeng - 2 hours ago

Excited that the long awaited v4 is finally out. But feel sad that it's not multimodal native.

taosx - 2 hours ago

MErge? https://news.ycombinator.com/item?id=47885014

luyu_wu - 2 hours ago

For those who didn't check the page yet, it just links to the API docs being updated with the upcoming models, not the actual model release.

aliljet - an hour ago

How can you reasonably try to get near frontier (even at all tps) on hardware you own? Maybe under 5k in cost?

KaoruAoiShiho - 2 hours ago

SOTA MRCR (or would've been a few hours earlier... beaten by 5.5), I've long thought of this as the most important non-agentic benchmark, so this is especially impressive. Beats Opus 4.7 here

namegulf - an hour ago

Is there a Quantized version of this?

swrrt - 2 hours ago

Any visualised benchmark/scoreboard for comparison between latest models? DeepSeek v4 and GPT-5.5 seems to be ground breaking.

mariopt - an hour ago

Does deepseek has any coding plan?

reenorap - an hour ago

Which version fits in a Mac Studio M3 Ultra 512 GB?

ls612 - 2 hours ago

How long does it usually take for folks to make smaller distills of these models? I really want to see how this will do when brought down to a size that will run on a Macbook.

rvz - 2 hours ago

The paper is here: [0]

Was expecting that the release would be this month [1], since everyone forgot about it and not reading the papers they were releasing and 7 days later here we have it.

One of the key points of this model to look at is the optimization that DeepSeek made with the residual design of the neural network architecture of the LLM, which is manifold-constrained hyper-connections (mHC) which is from this paper [2], which makes this possible to efficiently train it, especially with its hybrid attention mechanism designed for this.

There was not that much discussion around it some months ago here [3] about it but again this is a recommended read of the paper.

I wouldn't trust the benchmarks directly, but would wait for others to try it for themselves to see if it matches the performance of frontier models.

Either way, this is why Anthropic wants to ban open weight models and I cannot wait for the quantized versions to release momentarily.

[0] https://huggingface.co/deepseek-ai/DeepSeek-V4-Pro/blob/main...

[1] https://news.ycombinator.com/item?id=47793880

[2] https://arxiv.org/abs/2512.24880

[3] https://news.ycombinator.com/item?id=46452172

frozenseven - an hour ago

Better link:

https://news.ycombinator.com/item?id=47885014

https://huggingface.co/deepseek-ai/DeepSeek-V4-Pro

- 2 hours ago
[deleted]
hongbo_zhang - an hour ago

congrats

dhruv3006 - 42 minutes ago

Ah now !

creamyhorror - 2 hours ago

[dead]

hubertzhang - an hour ago

[dead]

maryjeiel - an hour ago

[dead]

minhajulmahib - an hour ago

[flagged]

slopinthebag - 30 minutes ago

OMG

OMG ITS HAPPENING

shafiemoji - 2 hours ago

I hope the update is an improvement. Losing 3.2 would be a real loss, it's excellent.

raincole - 2 hours ago

History doesn't always repeat itself.

But if it does, then in the following week we'll see DeepSeek4 floods every AI-related online space. Thousands of posts swearing how it's better than the latest models OpenAI/Anthropic/Google have but only costs pennies.

Then a few weeks later it'll be forgotten by most.