Dr.Sid
Philosopher
It's not only about resources needed to run it. It's also about resources needed for training.We know they are telling the truth about the computational resources it requires. It is open source and people have already got it running.
You can run distilled and downscaled models, but you have to train with full precision. And to train 671B parameters with 32 bit precision you simply need 2.7T of RAM, or better, GFX card with 2.7T of VRAM.
On the other hand Open AI nor google share how they did the training, what hardware they used, or how long did it take. So it's hard to judge.
IMHO the main contribution of Deepseek is that somebody other than OpenAI replicated the reasoning approach. It was assumed it's especially hard to train .. but mostly because of the lack of training data. Deepseek solved it. But again, we don't know how OpenAI did that, and if it was problem at all. And we don't know how Deepseek trained it. We can only check the results.
It's entirely possible Deepseek did it exactly the same way OpenAI did, by use of spionage, on similar hardware, by circumventing the embargos.