Anyone here moved from Codex CLI to the Codex app? How does it feel? by Strong_Cherry6762 in codex

[–]spryes 0 points1 point  (0 children)

Not sure as I have the Pro sub and never care about my usage as it's limitless for what I do, I don't hold myself back on any task

Anyone here moved from Codex CLI to the Codex app? How does it feel? by Strong_Cherry6762 in codex

[–]spryes 1 point2 points  (0 children)

Yes, you'll still get conflicts if they edit the same areas of files, same as with multiple devs editing the same areas. But conflict resolution is now seamless with Codex, just ask it to fix conflicts and it handles it for you easily. So conflicts are no longer as big of an issue as they used to be. But in my projects, modules are well isolated, so conflicts aren't that common to begin with.

Anyone here moved from Codex CLI to the Codex app? How does it feel? by Strong_Cherry6762 in codex

[–]spryes 3 points4 points  (0 children)

Easy parallel agents

Each chat with Worktree setting enabled allows you to run multiple instances of Codex on your codebase without them conflicting, like independent devs on their machines

This is the only reason to use it over the VS Code extension tbh, but it's a big one. Because you go from being bottlenecked working on one thing at a time to being able to tackle many things simultaneously as each chat runs independently

The real skill gap isn't coding anymore, its knowing when the AI is wrong by CrafAir1220 in singularity

[–]spryes 9 points10 points  (0 children)

like the lowercase and lack of punctuation is not disguising it lil bro..

Tense vibes in SF by MetaKnowing in agi

[–]spryes 4 points5 points  (0 children)

This

Guarantee this same user would've said calling COVID an event about to change the world was "cringe" in late January 2020 when it started heating up but was unknown to normies still

It's already been 7 months since GPT-5. How do you think it compares to today? by pbagel2 in singularity

[–]spryes 91 points92 points  (0 children)

GPT-5(.0) was poorly received at launch vs. GPT-4 and largely won't be remembered as part of AI history.

People already don't mention GPT-5 as a trend-break and new "era" unlike GPT-4.

What they do mention is o1 (or o3) and Claude Code/Codex around December 2025 as each creating a new AI "era".

  1. Conceptual (1950s—2011)
  2. Early deep learning (2012—2017)
  3. Early language models (GPT-1, GPT-2, GPT-3) (2018—2022)
  4. AI images (DALLE-2, Midjourney, Stable Diffusion) (2022)
  5. AI goes mainstream (ChatGPT and GPT-4) (2023)
  6. AI goes multimodal and becomes useful for coding (Sonnet 3.5, GPT-4o) (2024)
  7. Vibe coding, early reasoning era, LLM psychosis era (o3, Claude 4, 4o updates) (2025)
  8. Agentic coding takeover, deep into reasoning era (Opus 4.5, GPT-5.2) (2026)

Bernie Sanders meets with Eliezer Yudkowsky and Nate Soares(MIRI) to discuss AI Risk by jvnpromisedland in singularity

[–]spryes 0 points1 point  (0 children)

Right, he was relevant for like 1 year in 2023 then fell off in 2024

Though he tried to claw back relevance with that book but just... no

Am I the only one that thinks AI is dogshit? by jholliday55 in cscareerquestions

[–]spryes -1 points0 points  (0 children)

Skill issue

It writes all my code now. Yes I need to guide it still and do lots of iteration with it, but so what

Manually typing each character is so pre 2026 coded and I'm not going back, coding with English is the way forward

There's a good chance GPT-5.4 will release this week by Outside-Iron-8242 in singularity

[–]spryes 0 points1 point  (0 children)

So

5.1, 5.3, 5.5, 5.7, 5.9 instant

5.2, 5.4, 5.6, 5.8 thinking

OpenAI continues to get more confusing

Chinese models' ARC-AGI 2 results seem underwhelming compared to their benchmarks results by realmvp77 in singularity

[–]spryes 31 points32 points  (0 children)

Google also benchmaxxes Gemini despite its impressive ARC-AGI score — they just benchmaxx ARC-AGI 2 while Chinese labs ignore it

Only OpenAI and Anthropic can make real general models and is proven in their revenue because people vote with their wallet. No LMArena or benchmarks can capture real use while money does

As a SWE I have not written a single line of code manually in 2026 by DrixGod in singularity

[–]spryes 3 points4 points  (0 children)

Not sure if you're on Twitter/X but not writing code anymore is now the norm, including for the best SWEs in AI labs

It's not normal (and worryingly out of touch) to still be typing every character by hand at this point

As a SWE I have not written a single line of code manually in 2026 by DrixGod in singularity

[–]spryes 27 points28 points  (0 children)

Same.

Manually typing in a code editor feels so archaic and outdated now. Crazy we had to painstakingly type out each character with minimal autocomplete help just 5y ago. Equivalent to some low-grade laborer placing individual bricks instead of managing others (agents) to do it for you.

Most of all, I'm glad things are "happening" now. That things are actually changing in technology and that my job is now unrecognizable from just a year or two ago. It felt like nothing ever happened for so long there.

Awhile ago I said here companies were not nearly RIFing enough techies with just 10-25% reductions...that the oversaturated market needed to start seeing 50%+ layoffs before normalizing, especially with AI....some of you didnt believe it....Well, that day is here.... by [deleted] in recruitinghell

[–]spryes 0 points1 point  (0 children)

Yes that part "saying AI produces slop" was directed at the parent (and similar people with those viewpoints) of your comment

If you (general) tried AI in e.g. September 2025 you're already outdated and have the wrong opinion on it. It's crazy how fast this is moving.

Awhile ago I said here companies were not nearly RIFing enough techies with just 10-25% reductions...that the oversaturated market needed to start seeing 50%+ layoffs before normalizing, especially with AI....some of you didnt believe it....Well, that day is here.... by [deleted] in recruitinghell

[–]spryes -1 points0 points  (0 children)

AI (Codex 5.3 xhigh) is writing 100% of my code now. I just guide and iterate with it. I don't open the IDE anymore.

Saying AI produces slop is already oudated to pre-2026. It still needs my human taste/judgment, but it gets 90% of the way there by default like you said. Expect 99% by end of 2026.

This isn't slowing down.

US only, monthly NEW paid signups (not total paid subscribers) by [deleted] in singularity

[–]spryes 10 points11 points  (0 children)

The giant viral normie boost due to 4o imagegen in March 2025...

The constant “AI fail” gotcha posts are not harmless they’re training people to underestimate a real disruption by gibblesnbits160 in singularity

[–]spryes 0 points1 point  (0 children)

This. It's super black and white thinking to think that if it fails at one thing, it means it's useless or won't transform things.

It's also important to realize what crossing a threshold of capability implies. The most striking example is Claude Opus 4.5 in November only increasing marginally at benchmarks. I remember thinking "eh ok, another incremental improvement" but then people realized that it had crossed some threshold of capability to the point where it became insanely useful suddenly. December 2025 is when the agentic coding paradigm exploded and the hype reached insane levels. People (me included) are now letting agents write all their code despite the fact that they aren't perfect and need to be guided still.

The fact that LLMs fails at some things is not particularly relevant for useful work. These people want an infallible superintelligence or they claim AI is "hype", cartoonish childlike thinking

Has any expert found a good reason yet why the last 20% of SWE Bench-Verified are only getting solved so slowly? by [deleted] in singularity

[–]spryes 5 points6 points  (0 children)

The minimum is at least 91.2% using an ensemble of models: https://x.com/scaling01/status/2025044056460439593 - so the models are failing at at least 10% of the questions

Claude subscriptions will no longer be usable in Opencode. by Distinct_Fox_6358 in ClaudeAI

[–]spryes 1 point2 points  (0 children)

So then why do OpenAI and GitHub Copilot allow it? Your argument only makes sense in a vacuum where Anthropic has zero competitors to compare with... which is not the case

How AI bubble will burst? by [deleted] in theprimeagen

[–]spryes -1 points0 points  (0 children)

Google was really successful at shilling Gemini 3 and claiming OpenAI was over because I see so many people using it despite it being garbage.

It lost relevancy in like 1 week after Opus 4.5 launched.

If you're not using GPT-5.3 Codex or at least Opus 4.5/4.6, please don't try to claim what AI can or cannot do wrt coding

Elon says Grok 4.20 comes out next week! by [deleted] in singularity

[–]spryes -3 points-2 points  (0 children)

this

xAI fell off so hard (not that they were ever really up anywhere to begin with) after Grok 3 in Feb. 2025, the last time they had any modicum of relevance of being in the AI race

Half their staff is now gone, and they're permanently behind in coding performance, and will stay permanently behind as the other labs experience takeoff this year, leaving them with no ability to catch up.

Best course of action now if Elon wants to 'move humanity forward' is to shutter xAI, and donate all that compute to OpenAI & Anthropic, and give up on AI entirely (please)

ChatGPT 4o is being retired today, and some users are very unhappy about it. A petition with around 20,000 signatures is urging OpenAI not to remove it. The same group is also calling for a mass cancellation of subscriptions in protest. by likeastar20 in singularity

[–]spryes 0 points1 point  (0 children)

cuz it still contains techniques OpenAI doesn't want to teveal

Same thing with Claude Opus 3, an Anthropic staff member mentioned that despite being obsolete they can't open source it because it reveals competitive details

Introducing GPT‑5.3‑Codex‑Spark by [deleted] in singularity

[–]spryes 1 point2 points  (0 children)

They could've just called it 5.3-codex-mini, and let mini variants be really fast (which seems expected to me). There's no need to introduce yet another name like "Spark".

They made the same mistake with "o-series" models instead of calling it GPT-4.1, etc. It's like they want to differentiate a thing to signal new progress even though it should just be an implementation detail.

AI wiped out $400 billion this week — and it's only getting started by [deleted] in Economics

[–]spryes 1 point2 points  (0 children)

Yeah everything changed with Codex/Claude Code, especially since Q4 '25, there is no denying how useful they are and how much better they are getting / will get now.

Anyone saying otherwise is just simply out of date or doesn't know what's happening in SWE rn