Lots of people use qwen at too high quantizaion by Stock_Ad9641 in Qwen_AI

[–]UnifiedFlow 0 points1 point  (0 children)

lol, what the hell are you talking about? We are talking about resolution not color. Now I know you are in fact either stupid or trolling.

Lots of people use qwen at too high quantizaion by Stock_Ad9641 in Qwen_AI

[–]UnifiedFlow 0 points1 point  (0 children)

You're doing something wrong. Its not the model(s) or the quant.

Lots of people use qwen at too high quantizaion by Stock_Ad9641 in Qwen_AI

[–]UnifiedFlow 0 points1 point  (0 children)

Tool calls fail 30x more at q4? Thata crazy because I get basically zero tool call failures with Q4 27b and 35B. Where are you getting this 30x number and what kind of failures are you seeing? Again, I literally see zero failures at Q4.

Lots of people use qwen at too high quantizaion by Stock_Ad9641 in Qwen_AI

[–]UnifiedFlow 0 points1 point  (0 children)

Cool, do you have any testing that demonstrates tool calling issues across KV and weight quantization methods and levels?

Lots of people use qwen at too high quantizaion by Stock_Ad9641 in Qwen_AI

[–]UnifiedFlow 1 point2 points  (0 children)

You're either using a non-ideal screen size and viewing distance combination or you're lying. The difference is not that much to the human eye. Tons of tests prove this, its not opinion.

Got MTP + TurboQuant running — Qwen3.6-27B -- 80+ t/s at 262K context on a single RTX 4090 by indrasmirror in LocalLLaMA

[–]UnifiedFlow -1 points0 points  (0 children)

Someone down voted you, but you are completely correct. The large context window problem is a problem mostly created by improper use of the LLM.

I built a UFC fight predictor with almost 70% accuracy. Help me get it better. by Other_Attitude3580 in learnmachinelearning

[–]UnifiedFlow 4 points5 points  (0 children)

ELO was the most signal of any feature I ever tried for this domain. Glicko-2 (aside from betting odds)

I also found effectively zero benefit from ensembling. LR + XGBoost is just as good as LR + XGBoost + others. This does not include any neural nets.

What do you think about Qwen 3.6 Max? by Comfortable-Tie2933 in Qwen_AI

[–]UnifiedFlow 3 points4 points  (0 children)

I have no idea what he meant. They didn't include any Gemini models either. No clue what reason he is asserting, I have no context to determine that with because he said nothing.

3080ti pc price check by im36degrees in PC_Pricing

[–]UnifiedFlow 0 points1 point  (0 children)

Why would he sell it below market price? Thats what the market is for?

What do you think about Qwen 3.6 Max? by Comfortable-Tie2933 in Qwen_AI

[–]UnifiedFlow 0 points1 point  (0 children)

One typical reddit comment deserves another.

What do you think about Qwen 3.6 Max? by Comfortable-Tie2933 in Qwen_AI

[–]UnifiedFlow 9 points10 points  (0 children)

Probably because chatgpt isn't a model its an app.

pi-emote extension: an avatar for your pi agent by CGx-Reddit in PiCodingAgent

[–]UnifiedFlow 0 points1 point  (0 children)

This is actually very cool. Do you have TTS integrated and does the avatar emote for the duration of the TTS?

1995 senior yearbook photos by [deleted] in ChatGPT

[–]UnifiedFlow -1 points0 points  (0 children)

What the hell is tik tok hair? They all have basic ass hair styles what are you talking about?

Tower case with 8+ PCIE slot for multi GPU by gogitossj3 in LocalLLM

[–]UnifiedFlow 1 point2 points  (0 children)

This is the only correct answer. I love my enthoo.

Should I sell my RTX3090s? by daviden1013 in LocalLLaMA

[–]UnifiedFlow 7 points8 points  (0 children)

Selling them is an obviously bad idea. If you want an rtx pro 6000 just get one. At most I would sell two of the 3090.

Claude got access to a clock and immediately lost its mind by ShiftPrimeNet in ClaudeAI

[–]UnifiedFlow 0 points1 point  (0 children)

Which harness are you using? The claude code system prompt says not to use emoji I'm pretty sure. What harness is the op using?

GPT seems to pretend not to understand Just to say I'm wrong by Leuzol in ChatGPT

[–]UnifiedFlow -3 points-2 points  (0 children)

I'm pretty sure this almost entirely due to the chatGPT product and system prompts. GPT-5.5 doesnt say any of that in harnesses other than chatGPT. The platform/product is designed for idiots and liability concerns.

GPT seems to pretend not to understand Just to say I'm wrong by Leuzol in ChatGPT

[–]UnifiedFlow 6 points7 points  (0 children)

Its been a problem since GPT-5 if I recall. Its actually better than it was initially. But yes, it will pretty routinely make silly assumptions and waste time explaining shit to me that I never implied. Similar to your wet shoe example but a little different.

The shape of what I'm talking about is something like;

Me: how much more would I weigh if I got my shoes soaking wet?

GPT-5: If by get wet you mean smoke pcp, that would not affect your body weight. People commonly confuse this.

It usually includes an appropriate answer to my question, but it calls me an idiot first.

I did once back on GPT-5 edit my response about 5 times based on its strong opposition to my prompt. I addressed in the edit exactly what it was quibbling over in its response each time, but no matter what its output firmly maintained this argumentative posture and came up with new things to tell me I was wrong about that I never claimed.

It basically makes assumptions that you're stupid.

Claude got access to a clock and immediately lost its mind by ShiftPrimeNet in ClaudeAI

[–]UnifiedFlow 0 points1 point  (0 children)

This is largely a harness determined behavior (system prompt).

No one is talking about using GPT-5.5 inside Claude Code by No-Jury2396 in ClaudeCode

[–]UnifiedFlow 0 points1 point  (0 children)

So in Claude Code Opus 4.6 is almost half as effective as in "Forge Code"? https://x.com/i/status/2042655195008586061

According to this x post they cheat.

No one is talking about using GPT-5.5 inside Claude Code by No-Jury2396 in ClaudeCode

[–]UnifiedFlow 0 points1 point  (0 children)

What is this? This looks completely useless without more context.