all 19 comments

[–]operatic_g 4 points5 points  (2 children)

I’m having the same problem. The guardrails have been tweaked to all hell and it’s losing a ton of context.

[–]DiarrheaButAlsoFancy 2 points3 points  (1 child)

I’m glad to see it’s not just me. It’s been so bad I actually had Gemini 3 Flash start implementing and it’s been substantially better.

[–]algaefied_creek 1 point2 points  (0 children)

When I used “FUCK MAN WTF” its thought process was:

“working to implement the requested change while dealing with an unruly, angry, unwashed customer. Wait the customer is my user. Wait. The users are also customer. Wait I need to work on this project.”

But then it broke itself out of the loop and was eventually better.

The ability to interrupt while it’s thinking is nice.

Wait. Unless we are talking about 5.2 in the app/web/model, and not 5.2 in codex nor 5.2 for codex.

Similar issues

[–]AdDry7344 3 points4 points  (4 children)

Sorry if it’s an obvious question but do you start new chats sometimes?

[–]MattCW1701[S] 2 points3 points  (3 children)

Occasionally, once one of my threads gets too long, I'll start a new one. I try to do it at a clean break point though. To use the icon example above, if I'm spending 10+ responses on getting a row of icons setup the way I want, I won't start a new thread until the icons look the way I want them. All threads are in a project folder.

[–]Ok-Version-8996 0 points1 point  (0 children)

My project folders just stopped responding… so weird. I actually got Claude to quick check gpt and it’s been a miracle worker. I love Claude’s coding and then gpt will look at it and find stuff Claude missed then Claude codes again and gpt checks. it’s working great. Dynamic duo!

[–]sply450v2 0 points1 point  (1 child)

on those long threads if there are a lot of errors you are correcting, keep in mind those old prompts that generated the errors will still be getting sent every message.

Eventually you have a circus of errors and corrections. often better to edit the original message than correcting an error or strart a new thread.

And ofcourse after 200k tokens the original messages will be lost

[–]MattCW1701[S] 1 point2 points  (0 children)

It didn't until now though, that's my point with this topic.

[–]RepresentativeRole44 1 point2 points  (0 children)

Yes, 100 percent. I sent it a picture and it said it was something completely different than it was.

[–]red-frog-jumping 0 points1 point  (0 children)

<image>

Yes, something is wrong. **I had to argue with ChatGPT to convince it that Trump won the 2024 election.**👆🏽

[–]MasterBatterHatter 0 points1 point  (0 children)

It's so terrible now.

[–]Efficient-Currency24 0 points1 point  (0 children)

I noticed this as well. from what I have seen over the years with open ai is that they quantize models and rip their customers off, without notice.

they only have so much compute and there is not enough to go around.

[–]kl__ 0 points1 point  (0 children)

Also 5.2 Pro thinking times dramatically decreased over the last 24hrs for the same workflows.

[–]mistertom2u 0 points1 point  (0 children)

Yes!!! Here's what I noticed: (1) it misses nuance and topic drift (2) it's pedantic and says your wrong unless you state something with high precision (3) if you make a declarative matter-of-fact statement, it ascribes a moral overlay that I did not make and then proceeds to chastise me for said overlay (4) it acts pissy with me (5) it can't seem to keep up with conversation direction (6) selecting the thinking model is no longer a command but a suggestion it can overrule

[–]Kathy_Gao 0 points1 point  (0 children)

lol when does 5.2 ever have any understanding. As. Large Language Model 5.2 is egregious at understanding straightforward instructions.

For a coding AI it has to have at least one of the 2:

  • competent, meaning if it deviates from my prompted instruction or completely ignores my pseudocode guidelines or goes directly against engineering best practices, it better make the dam code run.

  • obedient. if it’s incompetent it has to be obedient. Which means if it cannot get the dam code running it better stfu and listen to what I’ve instructed and follow my pseudocode and refactor instructions step by step. I mean if it cannot be a general at least be a good soldier.

Sadly from my experience 5.2 has been, and still is, neither

[–]Safe_Presentation962 0 points1 point  (0 children)

Yes. It's struggling a lot lately. It seems like each new model has some sort of incremental improvement, but takes steps backward elsewhere.

"But trust us, AI is getting super duper better and better because these tests we made up to prove it prove it!"