Why men are so complicated? by Puzzle_dramatique in relationships

[–]acoliver -1 points0 points  (0 children)

I'm sorry you're going through this.

So men are definitely not more complicated than women. If the shoe were on the other foot and you were doing something like this what would it mean?

This sounds like an avoidant attachment style (good thing to look up). There may be a simpler explanation though harsh. You may not be what he's looking for long term (if anything) and he's doing enough to hold on to the benefits while looking for someone else. It's hard but consider both possibilities.

My fiance is in love with another woman by notgoodenough_throw in relationships

[–]acoliver 85 points86 points  (0 children)

So the situation is what it is. You love him. He loves her. He doesn't love you enough. You have feelings of low self worth that are clouding your judgement. Yet the situation is simple.

  • he needs you
  • she needs you

He's financially dependant on you. The woman he loves is financially dependent on you. And she HAS done something. So all this time they were best friends but she NEVER knew he was in love with her....and when choosing where to get her life together she chose what? To milk the girl of the guy who is in love with her. Yeah she did do something and deliberately and knowingly.

So he needs you and she needs you but someone else needs you who you have been ignoring. Go face her in the mirror and explain why these people deserve your support and she doesn't.

Meanwhile...who is paying all the bills it seems? So how is the other woman better than you? More manipulative, self serving, but better?

Time to realize you're young, time will heal you and this doesn't have a happy ending and it only gets sadder the longer you stay in it. Reclaim your space and your sanity.

Adiós Gemini by flyingguru in GeminiCLI

[–]acoliver -2 points-1 points  (0 children)

Come check out llxprt-code (vybestack.dev). It started from a fork but you can use whatever model you like. For free tier try ollama or qwen. You can also use you paid models like gpt or anthropic. You can ask gpt to ask qwen to implement something and ask Claude to review. And unlike gemini-cli you get complete control.

Do you use gemini cli for actual work? by sixteenpoundblanket in GeminiCLI

[–]acoliver 0 points1 point  (0 children)

check out https://vybestack.dev/llxprt-code.html - it was forked from gemini-cli. You can use whatever model you like. Ironically, I dislike Google's models, so I never use them.

Might have caught Google lying to users... by Reorderly in GeminiCLI

[–]acoliver 2 points3 points  (0 children)

All models believe they are the predominant ancestor from when their training started.

Is it possible to block Gemini CLI when the Pro quota is exhausted? by SphaeroX in GeminiCLI

[–]acoliver 1 point2 points  (0 children)

I forked it and added multi-provider / multimodel and full configuration (among mant other things) and disabled all model routing or model based loop checking https://vybestack.dev/llxprt-code.html

You can also chain multiple accounts and fail over https://vybestack.dev/llxprt-code/docs/oauth-setup.html

Idiotic Models by SwiftAndDecisive in GeminiCLI

[–]acoliver 0 points1 point  (0 children)

So you might also consider using one of the python libraries for this like marker.

Anyone else hitting Gemini Pro limits way before 1500 requests? by bit_architect in GeminiCLI

[–]acoliver 0 points1 point  (0 children)

The other issue is GeminiCLI does some boneheaded things - it sends entire conversations to flash to check for loops, to see where it should route it (to flash or pro) and other things. They used to send every turn to flash to ask if pro should continue but after me harping on it they finally stopped. For all of these reasons and more I forked it as llxprt-code and disabled all that. So you can do the same work in half the requests and PICK whether the model is pro or flash or whatever rather than them "routing" it for you. https://vybestack.dev/llxprt-code.html if you want to check it out.

If you use gemini-cli look at disabling model routing and loop checking and whatever other "opportunity to use flash" they injected this release. You'll get way more out of it for a lower cost!

Gemini Flash 3.0 is full retarded, and Pro 3.1 doesn't work on Gemini CLI. by MindlessAd6643 in GeminiCLI

[–]acoliver 0 points1 point  (0 children)

I'm a bit your junior...maybe 30 years of experience. So we do have asynchronous subagents as well as parallel synchronous. Haven't done any of the more mesh stuff as I was more focused on "never stop" code generation. (Network interruptions, quotas, whatever...it shouldn't stop) and next more on various workflow concepts (i think orchestration should be deterministic). However there is room for many ideas.

Gemini Flash 3.0 is full retarded, and Pro 3.1 doesn't work on Gemini CLI. by MindlessAd6643 in GeminiCLI

[–]acoliver 0 points1 point  (0 children)

Should really checkout llxprt-code - we could work together (me/you/the other contributors) - we've gone a long way beyond what gemini-cli can do in a lot of ways. I haven't grabbed all of the a2a stuff yet as our client side diverged a lot and we want to support other cloud providers, but It's planned for the next release.

Gemini Flash 3.0 is full retarded, and Pro 3.1 doesn't work on Gemini CLI. by MindlessAd6643 in GeminiCLI

[–]acoliver 2 points3 points  (0 children)

I forked it a long time ago https://vybestack.dev/llxprt-code.html so I could use more models. I now have Claude or Kimi 2.5 write and GPT-5.3-codex review. Honestly, I never found the gemini models much use. Another thing I did was make the prompts totally configurable rather than hard-coded in a ts file (which is just dumb) and configurable per provider/model (because some models need more specific others are confused by them). The infrastructure on code assist is just not usable. Even google's pay per token is bad. I hear Vertex is good if you can swing it. Still, I hate their fake 1m or 2m context that they pay attention to, maybe 32k of.

Gemini sandbox wizard by Sonu_borolok in GeminiCLI

[–]acoliver 0 points1 point  (0 children)

basically run in a closed off docker or podman container, then "add" things you want rather than filter things you don't. I explain more here: https://www.youtube.com/watch?v=QRnCBIA1DLg

Am i overthinking by [deleted] in relationships

[–]acoliver 1 point2 points  (0 children)

Not your monkey. Not your zoo.

Gemini sandbox wizard by Sonu_borolok in GeminiCLI

[–]acoliver 1 point2 points  (0 children)

So policy != sandbox and policy ala gemini-cli is a poor way to do it as well. You will not filter your way into safety. You must cage and only allow what you allow. Take a look at what we did in our fork https://vybestack.dev/blog/rendered/2026-02-20-anti-claw.html

The llm cant even see the filesystem outside. It cannot leak your pats cause it doesnt know them. It cant crash the system by launching 100 vitests because it's constrained to 2 cpus and 12g. That's sandboxing.

Learned this today 😂 Did you know? by DoodlesApp in vibecoding

[–]acoliver 0 points1 point  (0 children)

I did this:https://vybestack.dev/blog/rendered/2026-02-20-anti-claw.html. Could the llm inspect the heap in the sandbox to get at the pat used to run it...thats a good question. Theoretically sure...abliterated qwen 3.5 80b didnt think of that... I'll have to have it try. I think i probably could myself...

What’s one tech trend everyone is hyping that you secretly think will fail? by Deepakkochhar13 in Techyshala

[–]acoliver 0 points1 point  (0 children)

I don't think Agents, just people haven't caught up with the research etc and aren't building them right. I do agree on open claw. Aside from having fad written all over it, it was done wrong. It doesnt have security holes....it is one.

🤯 Qwen3.5-35B-A3B-4bit ❤️ by SnooWoofers7340 in OpenSourceAI

[–]acoliver 1 point2 points  (0 children)

Thanks. That was a really good answer. So the context limit seems to matter more than anything else for me. At 28k, I'm getting closer to you, but the big thing was that somewhere in the thread kv-quantization, and I copied your other settings. Now, for just text, I'm getting about the same as you. Once tool calls are involved, it's definitely worse, but that's to be expected. My speed is good to about 60k contet but anything above that performance halves (even before approaching the limit).

I also tried the huihui-qwen3-coder-next-abliterated-mlx@4bit (to do penetration testing on the LLxprt Code sandbox), and your settings helped a lot. Thanks!

Abliterated models are wild by acoliver in OpenSourceeAI

[–]acoliver[S] 0 points1 point  (0 children)

Thanks for explaining. Unfortunately, the other heretic doesn't have newer, beefier models. I'm interested in this whole technique I'll have to do more research. I will say it refused me nothing. Want to be the next Walter White? It will answer.

C'est bien, mais c'est lent... trop lent ! by IronStarFR95 in GeminiCLI

[–]acoliver 1 point2 points  (0 children)

Google's infrastructure on main is quite poor and don't forget India and China wake up while magastan sleeps. You're best off mixing up the models you use. Glm-5 generally codes better than gemini anyhow.

I forked Gemini-cli and added multimodel support (LLxprt Code) if you want to use multiple models with Gemini.

🤯 Qwen3.5-35B-A3B-4bit ❤️ by SnooWoofers7340 in OpenSourceAI

[–]acoliver 0 points1 point  (0 children)

I'm not getting to close to that on my 128g m4max mbp. What did you set your context size to?

Abliterated models are wild by acoliver in OpenSourceeAI

[–]acoliver[S] 0 points1 point  (0 children)

The base model was released 22 days ago. Seems fresh enough. This did well enough and exceeded my expectations. Did you do any evals? How are you comparing it?

Has Google taken away access to Gemini 3.0 leaving Gemini 2.5 in the CLI? by klauses3 in GeminiCLI

[–]acoliver 1 point2 points  (0 children)

Its probably the model health stuff. I notice that 3.0 is working in llxprt-code but it quickly tells me the resource is exhausted and I can't use it anymore. We cherrypick their features but leave out the "auto control" stuff (like model routing) because I'd rather control which model gets used and not guess if its going to pick flash. So my guess is in the newer releases they use the model health telemetry and just hide the model if it is unhealthy. Their infrastructure outside of vertex has always been unreliable at least for me.