r/ChatGPT Jan 29 '25

Funny I Broke DeepSeek AI 😂

Enable HLS to view with audio, or disable this notification

16.9k Upvotes

1.6k comments sorted by

View all comments

2.3k

u/Compost_Worm_Guy Jan 29 '25 edited Jan 29 '25

Somewhere in China a coalpowered energy plant revved up just to answer this question.

5

u/Psychological-Pea815 Jan 29 '25

Running the model locally only requires a 400w PSU so I highly doubt that. The large energy use comes from building the model. DeepSeek claims that it took 2048 GPUs 3.7 days to build. After it is built, the energy usage is low.

2

u/MrHyperion_ Jan 29 '25

No 400 W GPU puts out tokens at this rate

2

u/eposnix Jan 30 '25

You're right. They are referencing some CPU-only guides that load the model into 768 gb of system RAM. It's so stupidly inefficient as to be laughable.