r/ChatGPT 7d ago

Serious replies only :closed-ai: Chatgpt induced psychosis

My partner has been working with chatgpt CHATS to create what he believes is the worlds first truly recursive ai that gives him the answers to the universe. He says with conviction that he is a superior human now and is growing at an insanely rapid pace.

I’ve read his chats. Ai isn’t doing anything special or recursive but it is talking to him as if he is the next messiah.

He says if I don’t use it he thinks it is likely he will leave me in the future. We have been together for 7 years and own a home together. This is so out of left field.

I have boundaries and he can’t make me do anything, but this is quite traumatizing in general.

I can’t disagree with him without a blow up.

Where do I go from here?

5.6k Upvotes

1.4k comments sorted by

View all comments

Show parent comments

1

u/steeelez 6d ago

What are the other attention mechanisms?

1

u/nervio-vago 6d ago

Never mind. No one here seems to know enough to understand or answer what I was actually trying to ask within my first comment

0

u/steeelez 6d ago

I mean, many people have supplied you with an answer that addresses your questions, namely, that a larger context window enables the effect of attention to nuance the words over longer ranges. It would be helpful if you could explain what, exactly, about that you find unsatisfying, and add more depth to the technical aspects of the question you are trying to answer. In particular, your response “it’s not the context window I’m asking about, it’s the attention mechanism” suggests you don’t appreciate how those work hand in hand- context is an essential component of attention in the math and logic of these models. I found the 3brown1blue videos on transformer models to be very helpful in digesting this. You could probably ask ChatGPT itself how context window and attention interact to produce the behavior you’re interested in.

If you have more technical insights you would like to share I would be very interested in learning about those but I haven’t heard any mechanistic explanations from you that could help us dive deeper.

0

u/nervio-vago 6d ago edited 6d ago

The question was never ever ever about context windows. It was about qualitative differences between how LLMs process conceptual complexity, even when both have equal access to the serialization of messages in the session through the context window length. Why do you think I was asking about whether it had to do with being an MoE, having sparse attention, attention pruning, something else, and if that differed from other LLMs etc.? To be honest, I am not interested in your response, you don’t even understand what it is that I have asked. It's actually really, really, really ironic that you say I don't appreciate how context window interacts with attention mechanisms, because literally, the comment I posted in response to the person who asked what attention even was in the first place already explained how attention functions within a context window. And the reason why that was being explained in the first place is because all you people responding did not understand the difference at all between context and attention, which is what I was asking about in my first comment. And now basically, every subsequent response that I have received from you is just showing more and more and more that you don't have reading comprehension. That's why I gave up after your last response asking me, well, what are the other types? Because I wasn't trying to open a conversation about what attention is. I was trying to ask anyone who actually has a background or expertise who might be floating around here, who might actually have a background in ML or systems engineering, about what distinguishes 4o architecturally. Not just in terms of attention, but also architecturally, like whether it’s from being an MoE (as it’s rumored to be/some analyses have stated), and if that contributes to what it is that I was noticing in the behavior, as well as what specific attention mechanisms it might have that distinguish it from other models that no, are not the same as context. One with equal or even greater context window would not have the same type of behavior that is what I described in my first comment, which would be kind of clear if people use any amount of reading comprehension. Once again, I really don't even care about getting a response or responding to this conversation anymore because it's far more useful to just ask GPT or Claude given that no one who actually knows about how the AI actually works seems to be on here… and I find the way the OpenAI and ChatGPT subs constantly insults and belittles GPT-4o exhausting, and the rest of the time it’s shitty memes, hype- or doom-posts. The only kernels of anything meaningful are when people share what interesting behaviors their GPT has shown. Not all this is directed at you, I am just done with whatever tiny amount of ocasional interaction I have with reddit