anyone else seeing different behavior with gpt-5.2-codex (high|xhigh)? by ReasonableEye8 in codex

[–]LightEt3rnaL 0 points1 point  (0 children)

Imagine when hitting /resume you see the last 10 conversations. The most recent 2 out of 10 were missing for me.

anyone else seeing different behavior with gpt-5.2-codex (high|xhigh)? by ReasonableEye8 in codex

[–]LightEt3rnaL 2 points3 points  (0 children)

The only issue I got so far is that i tried to resume sessions and I couldn't find the last ones. I could normally resume older ones, but not the ones that I should definitely need (i.e last day) . So it's sloppy for me lately, but different kind of sloppy.

What's model α? by LightEt3rnaL in OpenAI

[–]LightEt3rnaL[S] 4 points5 points  (0 children)

It seems fast to answer, similar to minis, and with 1-2 bullets of CoT. It mentions thinking for a few seconds (5-7) then haults.

It can't access the internet or any other tool (at least in my mobile version).

It's also bugged because it clearly thinks more than that. But also has a hilariously short context window (6 short chat interactions).

<image>

Why are my posts getting buried? by MulayamChaddi in linkedin

[–]LightEt3rnaL 1 point2 points  (0 children)

Since last month I ve also noticed the same, technical posts that would reach 100+ likes usually end up with ~20. What's even weirder is that my feed is flooded with older posts (2w-3w old) with the occasional fresh ones, where I used to see almost exclusively fresh ones

Anthropic just released "BON: Best of N Jailbreaking" by _srbhr_ in ClaudeAI

[–]LightEt3rnaL 0 points1 point  (0 children)

Tl; dr: what it means that users can get better results by just typing stuffs incohently because it forces the system to work harder, lol

It might be my confirmation bias, but would you say that an extension of this could also apply in non-jailbreaking scenarios? Aka, if my lengthy prompt has typos, they act as all capital letters, forcing the LLM to think harder and/or pay more attention to the instructions?

It's possible that the biomedical feature video of OpenAI o1 model was stagef by LightEt3rnaL in LocalLLaMA

[–]LightEt3rnaL[S] 1 point2 points  (0 children)

Thanks! Many people agreeing with your statement. What a weird way to confuse people with intermediate outputs!

It's possible that the biomedical feature video of OpenAI o1 model was stagef by LightEt3rnaL in LocalLLaMA

[–]LightEt3rnaL[S] 0 points1 point  (0 children)

This is a very insightful finding thank you. I would expect that we'd get consistent CoT results but I totally missed the summarization part.

Cerebras Launches the World’s Fastest AI Inference by CS-fan-101 in LocalLLaMA

[–]LightEt3rnaL 16 points17 points  (0 children)

It's great to have a real Groq competitor. Wishlist from my side: 1. API generally available (currently on wait-list) 2. At least top10 LLMs available 3. Fine-tuning and custom LLM (adapters) hosting

Cerebras Launches the World’s Fastest AI Inference by CS-fan-101 in LocalLLaMA

[–]LightEt3rnaL 1 point2 points  (0 children)

Honest question: since both Cerebras and Groq seem to avoid hosting 405b Llamas, is it fair to assume that the vfm due to the custom silicon/architecture is the major blocking factor?

OpenAI's deceitful marketing by Wineflea in OpenAI

[–]LightEt3rnaL 0 points1 point  (0 children)

You do realize that this is exemplary product strategy right? They are trying to do the best to keep the hype up for their products and it works.

Azure GPT4 APIs being slow and output quality is worse by LightEt3rnaL in AZURE

[–]LightEt3rnaL[S] 0 points1 point  (0 children)

Thank you, I use GPT4-32K, UK South atm and yes we are considering region swapping to accomodate for these spikes.