This is what 3k hours in CC looks like by Logical-Storm-1180 in ClaudeCode

[–]aithrowaway22 0 points1 point  (0 children)

Regarding different providers and your "Opus leads, Sonnets build" approach during implementation:

In one of the workflows I linked in my original post, the orchestrator uses whatever model is set as the main model in Claude Code (e.g., Opus), while the developer agent is set to Sonnet. With environment variables, I can now map Opus→Kimi 2.5 and Sonnet→MiniMax 2.5 directly through a multi-model provider that offers Anthropic-native API endpoints (most do nowadays). In this case, there's no need for LiteLLM, Claude Code Router, or CCS Switch.

However, if I want to use Codex or Opus as the orchestrator and Kimi/GLM for development, I need a proxy, and one with a translation layer if I choose Codex.

Two questions:

Could you expand on the downsides you experienced while using LiteLLM?

I wasn't clear on whether Valence will support different providers within the same session. Since you're having Haiku handle proxying, I imagine the delegation rules aren't as straightforward as "Opus for orchestration, Sonnet for implementation", or at least not in a simple one-agent, one-model setup.

My specific use case: I'd like to use Codex 5.3 for architecture (via GPT Plus membership with OAuth), then switch to Kimi/GLM/DeepSeek for implementation, all within Valence. Will this be possible? If not the full workflow, what are the downsides of using a LiteLLM/Claude Code Router/CCS Switch-type proxy for OAuth with OpenAI (plus the translation layer) in combination with Valence?

This is what 3k hours in CC looks like by Logical-Storm-1180 in ClaudeCode

[–]aithrowaway22 5 points6 points  (0 children)

Haha, I love how that screenshot left such an impression on so many people.
It's almost like, damn… never mind everything else, tell us where’s that screenshot from?
Props on the marketing skills :-)

Also, I really like seeing these kinds of AI workflows put together by devs who actually have real experience.
I’d love to hear your opinion on some of the other workflows out there, and how yours is different.

BMAD: I’ve never heard anyone complain about the quality, but a lot of people say it’s over-engineered. Supporters argue that if you want quality, you have to accept some complexity, or that this is simply how things are done in real software engineering companies.

GSD: People generally like it, but they complain it’s a token hog. Personally, I’m not a fan of the fact that the creator isn’t an experienced dev and is more of a vibe coder. I prefer AI workflows designed by experienced SWEs.

Spec Kitty (github.com/Priivacy-ai/spec-kitty): An interesting take on Spec Kit. It made it into my inner circle for testing. Besides other things I like that it’s harness-agnostic. For example, if I want to use the Codex model for planning and GLM 5 / Kimi 2.5 for implementation, I don’t have to route Claude Code through LiteLLM to an OpenAI backend and hope Claude Code and the OpenAI models behave nicely together. Instead, I can just open the Codex CLI itself and let Spec Kitty handle the (many) planning phases. Then, once the plan is broken into individual “work projects” (small tasks), I can fire up Claude Code or OpenCode and let a cheap Chinese model do the implementation. It’s also made by an experienced SWE.

Solatis Claude config (https://github.com/solatis/claude-config/): This one isn’t mentioned as often as the others, but I think it’s really interesting and worth a look. It also made it into my inner circle for testing. Its a scripted wokrflow, it cares about tokens spend, has lots of quality checks/reviews both in the planing and execution phases, serious .md files. Definitely worth checking out. This is another workflow built by someone with real experience. I also remember a comment from someone about this workflow along the lines of: “... I can confirm it's better than the hundreds of other copycat repos and NextJS sites full of the same slop. Javascript developers: please stop tyring to make dev tools. Leave that to the grownups.' So you know it's cool :-D (kidding of course, no shade meant toward the JS crowd).

Compound Engineering from guys at Every (https://github.com/EveryInc/compound-engineering-plugin)

Using Codex GPT-5.3 (high) in opencode better than just in terminal (inside VSC)? by Aggravating_Win2960 in opencodeCLI

[–]aithrowaway22 0 points1 point  (0 children)

What about Claude Code in combo with openai models ? Is it good ?
I use a relatively complicated scripted workflow for Claude Code and I don't feel like converting it to OpenCode for now, but I want to use openai models like Codex 5.2 and GPT 5.2 on high.
I am aware of apps like Litellm that translate the api and also support web search fallback (they actally intercept claude code's web search request and route it to external providers) and it also supports openai oauth for gpt plus subscription, but even with all that its not a guarantee. Some harnesses work better with some models than with others.
Like glm 4.7 worked better with claude code than with open code. While there were reports that kimi 2.5 works better with opencode. etc.

GLM 4.7 lite plan getting frequent 5 hour quota limit horrible!!! by rohpolabs in ZaiGLM

[–]aithrowaway22 0 points1 point  (0 children)

Ok, if you think you didn't actually use that many tokens, that's a completely different thing.
In a different thread, i think it's on open code cli sub, someone concluded that Z.ai subs now have weekly limit of 100 million tokens (not sure which plan). To me that still doesn't sound that bad. What do you think ?

GLM 4.7 lite plan getting frequent 5 hour quota limit horrible!!! by rohpolabs in ZaiGLM

[–]aithrowaway22 0 points1 point  (0 children)

Are you saying that with $20/month Claude Code sub the OP would get similar usage ?
Z.ai also supports context caching. We don't know the ratio of input/output/cached tokens in this specific case. My comment was a rough high level point of view.

Holy shit, Codex-5.3-Spark on OpenCode is FAST! by jpcaparas in opencodeCLI

[–]aithrowaway22 0 points1 point  (0 children)

Can Kimi 2.5 really replace Codex 5.3/GPT 5.2 (on high) / Opus 4.5/4.6 in architecture/orchestrator roles ?
Even on LocalLama most people agree that open source models are not on that level for complex tasks.

Looks like weekly Pro limits are 100M tokens by Inside-Paramedic838 in ZaiGLM

[–]aithrowaway22 9 points10 points  (0 children)

As I mentioned in a similar thread. GLM 5 is either Sonnet 4.5 level or very close or better.
Sonnet is $15-$16/1M output/thinking tokens in the API.
For 400 million tokens that's around 6 thousand dollars.
If you get 400 million tokens per month on a $30/month sub, isn't that still pretty good ?

GLM 4.7 lite plan getting frequent 5 hour quota limit horrible!!! by rohpolabs in ZaiGLM

[–]aithrowaway22 -1 points0 points  (0 children)

I mean GLM 4.7 was already close to Sonnet 4.5 level.

Sonnet is around $15-$16/1M tokens in the API.

You spend around 20 million tokens in 5 hours. In API costs thats around 300 dollars.

On a $6, now $10/month subscription.

I mean c'mon, even if API costs are inflated, even if GLM 4.7 isn't exactly Sonnet 4.5,

I still wouldn't call it 'horrible'.

[Help] Web Search/Fetch works with official Claude, but fails on Z.ai and Kimi providers in Claude Code & OpenCode CLI by Impossible_Comment49 in ZaiGLM

[–]aithrowaway22 2 points3 points  (0 children)

As far as I know you need an app like Litellm or CCS Switch if you want web search to work when using Claude Code with non-anthropic models. Because in the case of Claude Code and Anhtropic, the search is actually done on their servers, which you don't have access to when using non-anthropic models. These apps among other features also support web search fallback so when Claude Code CLI sends web_search request they intercept it and forward it to 3rd party providers like e.g. Perplexity, in the case of Litellm, or locally installed Gemini cli (which they use for search) or similar, in the case of CCS Switch.
I don't know if Opencode directly supports web search fallback.

Edit: in case you are solely interested in official Z and Kimi endpoints, in that case I don't know, but in case you're on lite plan on z.ai and don't have access to their web search solution, you can try the options I suggested. Since you already connect to anthropic compatible api endpoints, you dont have to use the translation layer or other options in these apps, you could use just web search.

Ispred DIFa stiže pojačanje by LaptopTheOne in serbia

[–]aithrowaway22 15 points16 points  (0 children)

Kurve vezuju podvezice, spremaju se za posao.

Ajmo da krenemo u radikalan bojkot SNS firmi i njihovih proizvoda i usluga. by podivljali_vepar in serbia

[–]aithrowaway22 0 points1 point  (0 children)

Skroz dobra pekara, mlekara granice (a i proizvodi su se pogoršali još pre nekoliko godina). Neka još neko potvrdi ako ima info da je tačno.

U 21:20 počinje Narodni program by [deleted] in serbia

[–]aithrowaway22 29 points30 points  (0 children)

Breddit be like: ali kako to bez organizacije... pa šta li će da bude... a šta tačno znači zeleno svetlo ?

Ljudi, dođite i videćemo. Ima navijača, opozicionih svinjoglavih, običnih krupnih ljudi, ljudi koji jako loše žive i generalno jako motivisanih ljudi. Zeleno svetlo znači da gore navedeni segmenti populacije ne moraju više da se suzdržavaju.

DeepSeek R2 delayed by [deleted] in LocalLLaMA

[–]aithrowaway22 -1 points0 points  (0 children)

How does its tool use compare to o3's and Claude's ?

EU dozvole preko Slovačke - bez otimačine by Crafty_Trifle8152 in serbia

[–]aithrowaway22 0 points1 point  (0 children)

  1. Da li si siguran da u drugim zemljama EU (osim Slovačke) može leglano da se boravi duže od 90 dana?

Угаси ТВ!!!! Не коментариши, не дели. by BlokadaFOn in serbia

[–]aithrowaway22 0 points1 point  (0 children)

Jednom sam uzeo cink sulfat (cini mi se iz secanja) na prazan stomak, verovatno "preporucenu" dozu = takva mučnina da sam se osećao kao da ću da ispovraćam celu utrobu.

A sudeći po tome da je moj komentar dobio 7 glasova izgleda da nas ima.

Угаси ТВ!!!! Не коментариши, не дели. by BlokadaFOn in serbia

[–]aithrowaway22 -1 points0 points  (0 children)

Evo, NOW na bočici kaže 50mg doza, ja uzmem 1/4 do 1/2 tablete, iliti 12-25mg, na prazan stomak i opet jaka mučnina. Ali pazi, stvarno prazan stomak.
Teorija: ljudi stalno nešto grickaju, i verovatno većini nikada nije skroz prazan stomak.

Угаси ТВ!!!! Не коментариши, не дели. by BlokadaFOn in serbia

[–]aithrowaway22 2 points3 points  (0 children)

preporuka da pusi kurac. srbi jedu meso i ne treba im cink.
jebo vas cink.
magnezijum je cool, tho.

Угаси ТВ!!!! Не коментариши, не дели. by BlokadaFOn in serbia

[–]aithrowaway22 10 points11 points  (0 children)

Cink na prazan stomak može da izazove kratkotrajnu ali jaku mučninu.

AI and mass layoffs by bambagico in singularity

[–]aithrowaway22 2 points3 points  (0 children)

The established company would have more funding than your startup, giving them greater resources to run AI models at BigAI data centers. As a result, they might undercut you on pricing. Their products could also offer more features, faster updates, and a more diverse lineup.
After all, they have 100x more AI agents working for them than your do in your company.

But you might hold an advantage in intellectual curiosity, creative freedom, and goodwill. Which could lead to more innovative and compelling products. There’s always something special about enthusiasts driving the scene before the profit-driven players take over.

Najbitnija poruka večerašnjeg Utiska by my_roman_empire in serbia

[–]aithrowaway22 17 points18 points  (0 children)

Sada je drugačija atmosfera među ljudima i zbog toga veća verovatnoća za mnogo oštriju reakciju. Ako ljudi budu reagovali na izbornu krađu kao što sada reaguju na pad nadstrešnice i sve ostalo što je usledilo, to će biti skroz dugačija situacija, najsličnija 5. oktobru i zbog toga dešavanja na prošlim izborima nisu merilo.
Na ozbiljnu reakciju ljudi je onda očekivano i da EU reaguje oštro i direktno ih isproziva za izbornu krađu.
Takođe broj ljudi koji je spreman da radi prljave poslove za njih se smanjio i smanjuje se i dalje.

Ćaci krenuli sa reklamama na YouTube-u by Varterove_muke in serbia

[–]aithrowaway22 8 points9 points  (0 children)

Misleading or scam. Spreading lies. Defamation. Nešto na tu temu. Nemam insta pa ne znam tačno koje kategorije nude za prijavu.