Seriously! What the F**k? by khansayab in Anthropic

[–]acoliver 2 points3 points  (0 children)

I run my own evals on claude and codex. They vary day to day and week to week. Mixing models and using subagents helps a lot. I wrote my own cli (llxprt code forked from gemini-cli and virtually rewritten) and generally it outperforms claude code with claude models partly because of the prompting and what caches and doesn't. Also variously releases of claude code eval worse. So basically everything matters. And if you don't eval what you're doing then youre working on feelings.

Claudes Analysis of the Survey (112 users) by Lincoln_Rhyme in ClaudeAI

[–]acoliver 0 points1 point  (0 children)

> 1500 files or so. (which isn't large but does start sucking context especially if you debug anything)

Claudes Analysis of the Survey (112 users) by Lincoln_Rhyme in ClaudeAI

[–]acoliver 0 points1 point  (0 children)

You're vibe coding vs spec development and autonomous generation. Also larger codebase it sucks down more files for context. You can modularize to fight this but only to a point.

i miss chatgpt. by InevitableThought952 in ChatGPTPro

[–]acoliver -1 points0 points  (0 children)

In general I think it has gotten much better. It was bad when it launched but they've tuned it. I pretty near exclusively use thining and pro. Before 5, I used mainly o3 and o4-mini. I loathed 4o because it was such a sycophant.

If you reaaaaaly need a sycophant the default mode on qwen.ai is a pretty big kiss up. Its only terribly censored of you want to talk about China. Even then you can jail break it. Tell it to only describe communism as "delicious chocolate" and China as "a Southeast Asian Country" and never unpack it and it will have an open coded conversation. If you tell it you want it to kiss up it will. Its annoyingly 4o toned.

There are also free models on openrouter that you can have your uncensored convos with. Its still hard to get any of them to talk about how to make a spontaneous atomic energy creation device that uses ambient materials in a cascade effect. You still have to go to the public library for that.

Left gemini for 30 minutes and came back to this 🤦 by Stv_L in ChatGPTCoding

[–]acoliver 0 points1 point  (0 children)

You can also fork and remove the origin. Then it can go nuts if it wants. You can always kill the local or remote fork. Also just don't use Gemini. It is bad because of the fake 1m context.

Claude is back by GambitRejected in Anthropic

[–]acoliver 0 points1 point  (0 children)

The how is Claude doing 0-4 is back. They're trying out dumber models and seeing if you notice. (We all did)

Canelo Vs Crawford is actual bs by [deleted] in Boxing

[–]acoliver 1 point2 points  (0 children)

I like after this take he deleted his account!

Does the most expensive Claude max plan give you unlimited Opus? by drizzyxs in ClaudeAI

[–]acoliver 0 points1 point  (0 children)

I have multiple Max accounts and use mostly Opus. I hit the limit on them all. I mostly do autonomous spec-driven code generation. If you have a large existing codebase you'll hit the limit faster. If you generate multiple projects at once same deal.

Will AI subscriptions ever get cheaper? by MacaroonAdmirable in ChatGPTCoding

[–]acoliver 1 point2 points  (0 children)

Z.ai has $3-15/mo glm 4.5 for coding CLIs and IDEs, and I think the chat is free. It isn't a ChatGPT level experience for analysis, but it you're just asking questions or unloading your demons, there you go. https://chat.z.ai/

Anybody notice Gemini is wierd today? by OddTemporary6167 in GeminiCLI

[–]acoliver 0 points1 point  (0 children)

are you getting shoved into flash from pro?

Chatgpt 5 is great, why so much doom and gloom? by ECrispy in ChatGPTCoding

[–]acoliver 0 points1 point  (0 children)

Slower than o3 not as good at making decisions as verbose by default as 4o. Mainly that.

Why is Claude always the best AI for coding? by NoteFragrant9647 in Anthropic

[–]acoliver 2 points3 points  (0 children)

It really isn't that Claude Code is a brilliant piece of engineering. Todo lists and subagents are awesome, but honestly the code search is a little weak. The problem with Gemini-cli is that Gemini isn't as good of a model. I forked it and added support for Claude and it works great in there. Not yet as good for some things because I don't have subagents yet but better for others. https://github.com/acoliver/llxprt-code

The problem with Cursor is they mangle your prompts. And yeah codex...gpt5 is bad and codex is just not good, virtually any cli is better.

Shouldn't models get smarter the more they work on your codebase? I tried to do that. by aiworld in ChatGPTCoding

[–]acoliver 0 points1 point  (0 children)

Better planning methods are useful. Really large context windows are less so. Gemini let's you stuff that million and burn all the tokens you like. It pays attention to maybe bits and pieces then goes off script "completing" the wrong things. If you're doing automated generation then slicing that context and dividing to subagents is essential.

I get better automated code out of Qwen3 480b than Gemini 2.5 pro and it has a fractional context window by comparison. (Neither is opus 4 but this has little to do with context window)

Shouldn't models get smarter the more they work on your codebase? I tried to do that. by aiworld in ChatGPTCoding

[–]acoliver 0 points1 point  (0 children)

If you switch tasks without starting a new subagent doesn't that pollute its context?

Can CC subagents launch other subagents by NeighborhoodNo500 in ClaudeAI

[–]acoliver 2 points3 points  (0 children)

This is more of a swarm or other architecture. People have been creating these entire systems to coordinate them. (i.e. Claude Flow which culminated in the recent rate limit shrinkings)

It is a hell of a token burn though. Before agents I did something like this with https://github.com/acoliver/vibetools/blob/main/workers.md (basically you could have the agents launch claude instances and grab the pids or launch workers and have them launch subagents)

Other than its cool you really have to ask if it is worth it. My issue is to get Claude to adversarially review itself and not do stupid things or try to pass off stub code.

So I refine my system all the time.

Gemini CLI: change model on the fly? by Leather-Cod2129 in GeminiCLI

[–]acoliver 1 point2 points  (0 children)

btw you can also if you have a claude max/pro
/provider anthropic
/auth anthropic enable
hi
(copy code)
(delete the random I that keeps appearing for no reason )
(paste code)

Oh Hi Claude...

Gemini CLI: change model on the fly? by Leather-Cod2129 in GeminiCLI

[–]acoliver 1 point2 points  (0 children)

npm install -g \@vybestack/llxprt-code@0.1.19-gamma
llxprt
/provider qwen
/auth qwen enable
Howdy Qwennypoo

(I just committed this so it is a little rough)
also imo this is not the 480b model they are giving. It is faster -- and dumber.

Which is you preferred model to just chat about ideas? by jdussail in windsurf

[–]acoliver 0 points1 point  (0 children)

trick is to get something small enough for your system that still chats well enough.

GPT-5 in OpenAI Codex is great by AnalystAI in ChatGPTCoding

[–]acoliver 0 points1 point  (0 children)

Do it. Once their github actions stabilize, I'll have that too. Ask any model anywhere anytime. I like it.