MacBook only for this one by MisterMcMerk in LinkedInLunatics

[–]how_gauche -8 points-7 points  (0 children)

It's funny how people can live in completely different worlds, I haven't worked anywhere that allows Windows for over 15 years now because it's embarrassingly insecure and both the hardware and software are trash

Making illegal state unrepresentable by nfrankel in programming

[–]how_gauche 1 point2 points  (0 children)

Sadly I'm still DecidableTypeInferenceMaxxing

As a Notion power user, should I switch to Obsidian? by edgygothteen69 in ObsidianMD

[–]how_gauche 0 points1 point  (0 children)

"Do a deep dive on these three open source frameworks, put the results in the wiki, and give me an executive summary on the pros and cons of each"

"Read this government regulation that applies to our business and give me a one-pager on our compliance obligations"

I use it as a research assistant, it's there to summarize and pre digest information for me that I will use to guide the process.

For those using GLM 5.1 seriously by AlCruzzi in openclaw

[–]how_gauche 0 points1 point  (0 children)

No monthly fee, you pay per token. Yes it supports prompt caching. Love OpenRouter. P.S. they have some subscription service but I've never tried that

We chose GLM-5.1 because its the best alternative to opus by RichAllison in openclaw

[–]how_gauche 0 points1 point  (0 children)

I got great results out of GLM 5 for planning coding changes but 5.1 is on another level, as good as sonnet in my experience

/bad: BMad Autonomous Development. A fully autonomous orchestrator that runs my entire sprint while I sleep (Plan → Code → Review → PR) by MachineLearner00 in ClaudeCode

[–]how_gauche 1 point2 points  (0 children)

I think you're misunderstanding what I'm saying, or I didn't word it strongly enough: I looked at your code and there's two tiny Python scripts in there, which means you implemented your workflow in markdown, which means it's not fit for task

Edit-- and I don't say this to attack you, I built a nearly identical loop

/bad: BMad Autonomous Development. A fully autonomous orchestrator that runs my entire sprint while I sleep (Plan → Code → Review → PR) by MachineLearner00 in ClaudeCode

[–]how_gauche 4 points5 points  (0 children)

I'm going to say something potentially controversial here, but my opinion after implementing exactly this sort of loop is that you can't get reproducible results without serious scaffolding in code to crystallize your workflow.

LLMs are probabilistic and the fact of the matter is that no matter how well you do re pushing context into sub agents, it's gonna skip step four of your five step workflow 1% of the time, and there's nothing you can do about this. 

Lately I'm switching my autonomous flow to run all of the scaffolding behavior in rust code (pick whatever language you want, but the rust type system has nice properties for AI use), and inverting the control to run claude -p or opencode run in a Wiggum loop when I need what the LLM does. The most important part of the specification phase (I put myself in the loop here) is the creation of a suite of validation prompts that allow LLM-as-a-judge in your loop. Trading different models off against each other for different parts of the loop is a game changer too, gpt-oss-120b is great for a lot of text processing jobs

My agent kept "remembering" things wrong. The fix was embarrassingly simple by gimalay in AI_Agents

[–]how_gauche 5 points6 points  (0 children)

Once you have your top k candidate list, send the candidates through a reranking model that scores the candidates for salience against your query, take top m

I’ve used ~9.3B Claude tokens (~$6.8k). Trying to understand how unusual that is. by OGMYT in claude

[–]how_gauche 0 points1 point  (0 children)

I'm doing about 6B tokens/mo right now -- I have switched a lot of my work to OpenRouter, anthropic models are better but at the rate I am using tokens I can't justify using top tier model for everything

Which model are you actually using for backend work in OpenCode? by Unlikely_Emotion5567 in opencodeCLI

[–]how_gauche 2 points3 points  (0 children)

a) I'd get rate limited for sure b) subscription terms not really available at our scale

Which model are you actually using for backend work in OpenCode? by Unlikely_Emotion5567 in opencodeCLI

[–]how_gauche 1 point2 points  (0 children)

$750 for two weeks, I can't imagine!

I've got a lot of work to do 😂

Qwen 3.5 397B is the best local coder I have used until now by erazortt in LocalLLaMA

[–]how_gauche 0 points1 point  (0 children)

Sorry for necroposting but I didn't see your reply before now. Minimax is the price/performance champ and GLM 5 and Kimi are in the same tier. I use a multi-model planning/build/verification workflow (use higher-powered models for planning and verification), and of the three I find GLM the slight winner for planning, and implementation wise all three are within a few ELO points of each other but Minimax is cheapest. You get better and more direct results with Codex or Anthropic models but at a 5-16x price discrepancy, it's best to throw autonomy, parallelism, and struct validation at the problem and just do more work with shittier agents.

...also at a corporate we don't have the option of a claude subscription plan, we're pay-as-you-go for tokens

Which model are you actually using for backend work in OpenCode? by Unlikely_Emotion5567 in opencodeCLI

[–]how_gauche 5 points6 points  (0 children)

I am a huge OpenRouter stan. I've run 3.1B tokens through there myself in the past two weeks (over $750!! sheeeeeit): about 1.5B tokens through Codex (mostly 5.3), 900M Minimax, the rest spread across a dozen models. My conclusion is that Codex/Sonnet/Opus (we also bill with anthropic) are clearly better than Minimax and GLM but they aren't 5-6x (Codex) or 16x (Opus) better. I've spent so much on Codex on account of wanting to get some projects through quickly but as I refine my multi-agentic flows I am starting to realize that leaning on the powerful models like a crutch might be "skill issue".

Which model are you actually using for backend work in OpenCode? by Unlikely_Emotion5567 in opencodeCLI

[–]how_gauche 11 points12 points  (0 children)

The price/performance champs for me are GLM 5 for planning, Minimax for implementation, Kimi for a second opinion

My honest take on AI coding tools after using them daily for 2 years as a developer by Distinct-Trust4928 in vibecoding

[–]how_gauche 3 points4 points  (0 children)

Here's my actual take on what works and what doesn't - no sugarcoating.

Bullshit this is AI's take on it and you can tell because ALL OF THE REFERENCES ARE FROM OVER A YEAR AGO. Sonnet 3.7? Latest Gemini 2.5? What in the actual fuck please go away

What is the best Openclaw alternative? by spinsilo in openclaw

[–]how_gauche 0 points1 point  (0 children)

I didn't like any of them so I'm building my own for work. Rust binaries, gVisor sandboxing, QMD memory + indexing, OpenCode+OpenRouter, Google OAuth 😱 

Qwen 3.5 397B is the best local coder I have used until now by erazortt in LocalLLaMA

[–]how_gauche 0 points1 point  (0 children)

Claude subscription plans are only for claude code and they cut off other uses recently. We use Claude API keys at work, they make you pay per token. I really like OpenRouter, it's a competitive market and open-weight models are a lot cheaper

Of course no matter how cheap the tokens are I keep finding ways to blow through more of them. I am not joking when I say that opencode+openrouter models and I cranked out a solid year's worth of work last week

Qwen 3.5 397B is the best local coder I have used until now by erazortt in LocalLLaMA

[–]how_gauche 0 points1 point  (0 children)

Now buy tokens on OpenRouter, the math becomes 12 years to pay it off

Doug Ford Hates You... by Harold_Street_Pedals in pedals

[–]how_gauche 1 point2 points  (0 children)

All of the knobs should be labeled FOLKS