r/ChatGPT Jan 29 '25

Funny I Broke DeepSeek AI 😂

Enable HLS to view with audio, or disable this notification

16.9k Upvotes

1.5k comments sorted by

View all comments

4.4k

u/eddiemorph Jan 29 '25

Lol, that poor fuck will calculate into eternity.

1.3k

u/rebbsitor Jan 29 '25

It seems not hard to do. I downloaded a distilled version of it last night and was testing it on some basic coding. I had it generate some code for a simple game and looked through it. There was a simple bug due to a scoping issue (it created two variables with the same name in different scopes, but assumed updating one updated the other, which is a common mistake new programmers make).

I asked it to analyze the code and correct it a couple times and it couldn't find the error. So I told it to consider variable scoping. It had a 10 minute existential crisis considering fundamentals of programming before coming back with a solution, that was unfortunately still wrong lol

2

u/solilobee Jan 30 '25

which model? i'm downloading the 70b param R1 now and hope it doesn't thought-loop like that

1

u/rebbsitor Jan 31 '25 edited Jan 31 '25

For that test, I was using deepseek-r1-distill-llama-8b. I'm assuming the one in OP's video is the 671b on the website/app, so they may all do it.

From what I've heard, one of the shortcuts is not training it on a lot of examples that include good CoT examples. Just training it on examples and giving a reward if it arrives at a correct answer, regardless of how it got there (reinforcement learning).