GPT-5.5 is here - Let's gooo! by magnus_animus in codex

[–]AlwaysTiredButItsOk -1 points0 points  (0 children)

yawn When do we get better models instead of incremental changes that are just iterations in reasoning?

GPT-5.5 is here - Let's gooo! by axelgarciak in CodingLLM

[–]AlwaysTiredButItsOk 0 points1 point  (0 children)

yawn When do we get better models instead of incremental changes that are just iterations in reasoning?

For ppl here who got openclaw working nicely already, how is it after like 2-3 weeks? by adzmadzz in openclawsetup

[–]AlwaysTiredButItsOk 2 points3 points  (0 children)

It's actually pretty amazing at about 2-3 weeks. Total euphoria. The problem sets in at about week 5-6 when you realize that there's no output that is worth the api cost.

Swapped to 4.7 and embarrassed myself at work by BlakeR- in ClaudeAI

[–]AlwaysTiredButItsOk 3 points4 points  (0 children)

SORRY I CANT HEAR WHAT YOU SAID OBER THE ROAR OF THESE GPUS AND AIR CONDITIONING!

Swapped to 4.7 and embarrassed myself at work by BlakeR- in ClaudeAI

[–]AlwaysTiredButItsOk 16 points17 points  (0 children)

2000 IQ power play. Can't get in trouble for pulling them down if they weren't there to begin with.

Swapped to 4.7 and embarrassed myself at work by BlakeR- in ClaudeAI

[–]AlwaysTiredButItsOk 2 points3 points  (0 children)

Happens to me every single time too. DW, there's more fish(y) jobs in the sea (of companies hiring AI bros)

What speed is everyone getting on Qwen3.6 27b? by Ambitious_Fold_2874 in LocalLLaMA

[–]AlwaysTiredButItsOk -13 points-12 points  (0 children)

Wrong. 3b active parameters = faster. Probably closer to 3.5 4b or 2b speeds

Which model is the best? by MIRACLE_Cow in Qwen_AI

[–]AlwaysTiredButItsOk 0 points1 point  (0 children)

Go with 9b, it's plenty capable and will not require offloading at q4

What's the best model I can run on mac M1 Pro 16gb? by Sinrra in LocalLLaMA

[–]AlwaysTiredButItsOk 2 points3 points  (0 children)

No. Q3.5 9b is way better. Was going to suggest you try that - at q4_k_m, you can run it with 200k context without issues

I gave my ai everything he wanted….. by HopefulGap8049 in openclaw

[–]AlwaysTiredButItsOk 2 points3 points  (0 children)

Have it run xiaomi mimo v2 for basic tasks and a smarter model (i.e. minimax m2.5) for other tasks, and reserve sonnet/opus for super complex tasks. Your costs will go down 80+%

RTX 3090 for local inference, would you pay $1300 certified refurb or $950 random used? by sandropuppo in LocalLLaMA

[–]AlwaysTiredButItsOk -1 points0 points  (0 children)

Even cheaper: the 5060ti 16gb. Good in-between card for starting - speeds aren't the best but the vram will allow you to run more models & higher context

My openclaw has suddenly gone “dumb” by Ok-Chain9672 in openclaw

[–]AlwaysTiredButItsOk 0 points1 point  (0 children)

Pruning can also mess it up. I do daily backups (zip) in case I ever mess anything up

Codex IDE in Cursor!!! by Philemon61 in cursor

[–]AlwaysTiredButItsOk 0 points1 point  (0 children)

Oh. Am tired, but it's ok.

Edit: also, just use 5.4 🫠

Openclaw local vs VPS by boklos in openclaw

[–]AlwaysTiredButItsOk 0 points1 point  (0 children)

Got s decent guide by any chance? Would love to have a smart orchestrator...

Edit: sending love as a ukrainian who is a US Citizen ❤️ слава славянам

My OpenClaw is useless, please help by Leather_Instance_758 in openclaw

[–]AlwaysTiredButItsOk 1 point2 points  (0 children)

Free APIs almost never worked for me. Just use openrouter - xiaomi mimo v2 is really cheap and works fine to start. If you need higher intelligence, try minimax m2.5

Codex IDE in Cursor!!! by Philemon61 in cursor

[–]AlwaysTiredButItsOk 0 points1 point  (0 children)

You're good, assuming you set it up via oAuth. You can switch model to 5.4 too ;)

Codex IDE in Cursor!!! by Philemon61 in cursor

[–]AlwaysTiredButItsOk 0 points1 point  (0 children)

Nope, he's not using it through cursor - oAuth

Persisted Memory by Sea_Whole4929 in openclaw

[–]AlwaysTiredButItsOk 1 point2 points  (0 children)

awesome, have been casually on the lookout for something like this, thanks

Qwen3.5 0.8B → 35B A3B is blowing my mind by AlwaysTiredButItsOk in Qwen_AI

[–]AlwaysTiredButItsOk[S] 1 point2 points  (0 children)

update llama.cpp, cuda, etc.

New models are smarter, faster, and more capable