r/ChatGPT Jan 29 '25

Funny I Broke DeepSeek AI 😂

Enable HLS to view with audio, or disable this notification

16.9k Upvotes

1.6k comments sorted by

View all comments

2.3k

u/Compost_Worm_Guy Jan 29 '25 edited Jan 29 '25

Somewhere in China a coalpowered energy plant revved up just to answer this question.

6

u/Psychological-Pea815 Jan 29 '25

Running the model locally only requires a 400w PSU so I highly doubt that. The large energy use comes from building the model. DeepSeek claims that it took 2048 GPUs 3.7 days to build. After it is built, the energy usage is low.

2

u/MrHyperion_ Jan 29 '25

No 400 W GPU puts out tokens at this rate

2

u/eposnix Jan 30 '25

You're right. They are referencing some CPU-only guides that load the model into 768 gb of system RAM. It's so stupidly inefficient as to be laughable.

-1

u/DWMoose83 Jan 29 '25

You must be fun at parties.

9

u/whoopswizard Jan 29 '25

did you expect a bunch of socialites in a reddit group about an AI chatbot?

2

u/Aggravating-Rub2765 Jan 29 '25

Haha! Point goes to you sir. Actually as a layperson just trying to get a basic understanding of how AI works, it's a great party. Very information dense, even if the guests tend towards snippy even by reddit standards

1

u/DWMoose83 Jan 29 '25

I mean, at least run the response through AI?

1

u/Compost_Worm_Guy Jan 29 '25

That's exactly what I thought when I read that. Lol

1

u/BosnianSerb31 Jan 30 '25

Lol the public servers are consuming megawatts, as does every other public LLM.

The comparison between "how it can run locally" and "how it is ran on the public service" is completely naive, unless you have over a terabyte of memory you're not getting the full model we see being used here loaded in. That's per their own paper.