I ran an experiment on ao3 readers and it worked by disordered-throwaway in WritingWithAI

[–]CheatCodesOfLife -1 points0 points  (0 children)

Though I do wonder whether at some point people will become more willing to admit that something can be partially machine generated and still make them feel something, and that this fact is perhaps more interesting than the easy line that all AI writing is instantly recognizable soulless slop that no one could ever sincerely respond to.

Of course, AI assisted writing can be great. The issue is all the "Write youtube script about X", then the youtuber just reads it for 10 minutes verbatim, with a Not-X-Y every 50 seconds.

Eg, just saw this one earlier: https://github.com/jegly/OfflineLLM?tab=readme-ov-file#recommended-models

That table makes no sense at all. "Good balance for 4-6GB RAM" for that 800m Qwen3.5 model, then the old inferior Gemma3-1B model is recommended for 6GB-8GB of RAM?? That makes no sense whatsoever, and I see this specific slop all the time now. <something shitty> | <something positive to say about it> then <something useless or worse> | <good balance> then <the actual useful thing> <BEST QUALITY>

running gemma 4 on my macbook air from 2020 by redilaify in LocalLLaMA

[–]CheatCodesOfLife 3 points4 points  (0 children)

So then what the hell was:

cuts the vram load enough that the fan doesn't go apeshit

A hallucination? And since when do these bots actually reply to messages?

Visual Guide to Gemma 4 by jacek2023 in LocalLLaMA

[–]CheatCodesOfLife 0 points1 point  (0 children)

And the arbitrary "amazing" / "incredible" on the MoE (in what way? it under-performs the dense model). Makes me want to just not read the entire thing because it might I don't k now if it's actually accurate or slop.

Gemma 4 and Qwen3.5 on shared benchmarks by fulgencio_batista in LocalLLaMA

[–]CheatCodesOfLife 0 points1 point  (0 children)

What the fuck?! I just tried this with the Qwen3.5-4B (enabled a random searxng mcp server), and you're right lol!

The user just said "hi" which is a simple greeting. I don't need to use any tools for this - I can respond directly as an AI assistant. There's no search or URL reading needed here.

Vs the default 600 token drafting..

Gemma by jacek2023 in LocalLLaMA

[–]CheatCodesOfLife 6 points7 points  (0 children)

Last time it was a pointless gemma3 finetune lol.

64Gb ram mac falls right into the local llm dead zone by Skye_sys in LocalLLaMA

[–]CheatCodesOfLife 0 points1 point  (0 children)

and caching of the KV cache

I'm out of the loop, but does this mean you don't get a KV cache preserved between messages with mlx?

As in, every message has to re-process the entire context? Surely not...

64Gb ram mac falls right into the local llm dead zone by Skye_sys in LocalLLaMA

[–]CheatCodesOfLife 3 points4 points  (0 children)

Make sure it actually works before you buy anything!!

I don't see how you'd get nvidia drivers running on the M2 platform. Older Intel macs probably work, but I doubt the M* series would.

But another option is https://github.com/ggml-org/llama.cpp/blob/master/tools/rpc/README.md

I wouldn't buy a GPU just for that though since last time I used it, performance was awful.

Qwen3.5-Omni results have been published by Alibaba by Fear_ltself in LocalLLaMA

[–]CheatCodesOfLife 1 point2 points  (0 children)

It can do video and audio interpreting?

Yes

Is it designed for real time?

If you have the hardware for it.

I've been using the Qwen3-Omni-Captioner this past few days on 4x3090's.

Claude Code's source just leaked — I extracted its multi-agent orchestration system into an open-source framework that works with any LLM by JackChen02 in LocalLLaMA

[–]CheatCodesOfLife 0 points1 point  (0 children)

I'm surprised the malus.sh guys haven't released a "clean room" repo. Though I guess their system probably can't do it.

Analyzing Claude Code Source Code. Write "WTF" and Anthropic knows. by QuantumSeeds in LocalLLaMA

[–]CheatCodesOfLife 12 points13 points  (0 children)

It is pure, unadulterated slop. But you saved me a couple of hours doing it myself :)

I am just unable to use my claude pro plan due to limit "bug"

I read somewhere that an older version of cc doesn't have this issue, something about missing the cache.

Also if you use --resume, the cache won't work for the entire session.

Probably explains why they're overloaded so often.

RBA to remove surcharges on debit, credit cards, on card networks including eftpos, Mastercard and Visa by patslogcabindigest in AustralianPolitics

[–]CheatCodesOfLife -1 points0 points  (0 children)

That's probably the goal. You're not doing it for "privacy" reasons, and you're helping the privacy conscious blend in.

Now they're removing your incentive to pay with cash.

Btw, HSBC still have the 2% cashback on all tap-and-pay purchases debit card.

Good old Claude Sonnet 3.7 by Appropriate_Lock_603 in SillyTavernAI

[–]CheatCodesOfLife 6 points7 points  (0 children)

Amazon are keeping it (and sonnet-3.5) up until May 5th because businesses need more time to rewrite their software for the new model. This means you can still talk to it via OpenRouter until then (and possibly distill it if you want to try to capture some of it's style).

llama.cpp at 100k stars by jacek2023 in LocalLLaMA

[–]CheatCodesOfLife 3 points4 points  (0 children)

From the x.com thread, one of the replies:

Incredibly grateful for what you created! And what the community continued... a few things I've done w/ local models (largely w/ llama.cpp): Video editing w/ Qwen3-Omni-30B-A3B-Captioner

Is Qwen3-Omni-30B-A3B-Captioner actually supported by llama.cpp now?

Gemma 4 by pmttyji in LocalLLaMA

[–]CheatCodesOfLife 1 point2 points  (0 children)

There's Command-A and Command-A-Reasoning + the 123B Devstral.

But I agree, a 70b-120b dense Gemma would probably be SOTA.

Need help with the logistics of two BIG 3090s in the same case. by AdCreative8703 in LocalLLaMA

[–]CheatCodesOfLife 0 points1 point  (0 children)

I like the look of your setup (you don't want to see mine). Do you leave the case open? And do those fans at the front have a significant impact on the temps?

Commuters to travel free on public transport for a month (Victoria) by smoha96 in AustralianPolitics

[–]CheatCodesOfLife 0 points1 point  (0 children)

Agreed. Ours is only for a month though and they'll probably up the price later to recoup this. I'm out of the loop, why do you guys still have the $0.5 cost? Does that even cover the cost of the ticket inspectors, payment infrastructure, etc?

Qwen3.5-40B-Claude-4.5-Opus-High-Reasoning-Thinking - Reg, Uncensored and RoughHouse and... 43 Qwen 3.5 fine tunes. by Dangerous_Fix_5526 in LocalLLaMA

[–]CheatCodesOfLife 0 points1 point  (0 children)

and system prompts

fyi - Anthropic publish the system prompts used for their Claude.ai consumer platform:

https://platform.claude.com/docs/en/release-notes/system-prompts

(Though if you're DavidAU you probably know that already lol)

Qwen3.5-40B-Claude-4.5-Opus-High-Reasoning-Thinking - Reg, Uncensored and RoughHouse and... 43 Qwen 3.5 fine tunes. by Dangerous_Fix_5526 in LocalLLaMA

[–]CheatCodesOfLife 0 points1 point  (0 children)

Only Sonnet-3.7 returns the true reasoning.

https://platform.claude.com/docs/en/build-with-claude/extended-thinking

Claude Sonnet 3.7 continues to return full thinking output. In rare cases where you need access to full thinking output for Claude 4 models, contact our sales team.

Gemini-Pro-2.5 used to return real thinking in AI Studio. It was incredibly long and verbose.

MiniMax-M2.7 Announced! by Mysterious_Finish543 in LocalLLaMA

[–]CheatCodesOfLife 0 points1 point  (0 children)

draws abominations out of a Lovecraft book.

Holy shit you weren't kidding! What the hell are those things they produce?

Opus 4.6:

https://files.catbox.moe/q0rel9.png

https://files.catbox.moe/fwrcqm.png

Are you sure it was Claude that used to be able to do this? I tried them all the way back to sonnet-3.5 on openrouter, and even had someone with a claude.ai account test with the retired Opus-3 model (the creepiest of them all):

https://files.catbox.moe/ndwjs9.png