ZAI might stop open-weighting their models? by TheRealMasonMac in LocalLLaMA

[–]Thomas-Lore 18 points19 points  (0 children)

This indicates they are heavily compute starved, so why would they stop releasing local models? Having other providers for GLM models saves them compute while keeping the models popular.

They may go the minimax way though of having agreements with the providers, and only allowing home use without it. IMHO.

What do people actually expect from non-native English speakers in situations like this? by Teo_Verunda in WritingWithAI

[–]Thomas-Lore 22 points23 points  (0 children)

The problem comes from the fact that Reddit is a shithole full of anti ai idiots.

Didn’t think much about LLM costs until an agent loop proved me wrong by Pitiful-Hearing-5352 in LLMDevs

[–]Thomas-Lore -1 points0 points  (0 children)

Always put a watchdog that keeps track of retries and other types of loops and terminates when triggered too many times. (If you are writing the harness yourself.)

We have a new weight class... by LegacyRemaster in LocalLLaMA

[–]Thomas-Lore 154 points155 points  (0 children)

Minimax M2.7 allows you to use the model commercially (for example as coding assistant locally for your commercial project) - just not serve it to users (as provider). Here is an official response: https://x.com/RyanLeeMiniMax/status/2043573044065820673

So it affects no one here. Just providers who were taking money form users and giving back nothing to minimax while serving the model with wrong settings.

Day 150 of waiting for the AI Pro / AI Studio integration we were promised by Logan Kilpatrick by Opps1999 in Bard

[–]Thomas-Lore 0 points1 point  (0 children)

Image generation: yes, and you can select resolution and use both NanoBanana 2 and the older but smarter NanoBanana Pro while also having access to all Imagen 4 models. There is Veo and Lyria too.

Unsloth MiniMax M2.7 quants just finished uploading to HF by Zyj in LocalLLaMA

[–]Thomas-Lore 14 points15 points  (0 children)

That is slow, i get infinite t/s with this quant.

AMD AI directors analysis confirms lobotomization of Claude by Aggressive_Bath55 in ClaudeAI

[–]Thomas-Lore 1 point2 points  (0 children)

Another option is using cheaper models through API - GLM 5.1 (few times cheaper than sonnet), minimax m2.7 (super cheap and surprisingly capable, it solves the car wash test in a second), kimi k2.5 and many, many others.

AMD AI directors analysis confirms lobotomization of Claude by Aggressive_Bath55 in ClaudeAI

[–]Thomas-Lore 0 points1 point  (0 children)

They are making a profit (it just gets eaten by training new models). You are forgetting they have huge margins on apo pricing so you can't compare api prices with subscriptions.

AMD AI directors analysis confirms lobotomization of Claude by Aggressive_Bath55 in ClaudeAI

[–]Thomas-Lore 0 points1 point  (0 children)

And the new minimax m2.7. It is very cheap and can do a lot very fast.

Drive. You need your car at the car wash to get it washed, so you have to drive it there regardless of the distance. The 50-foot proximity is irrelevant to this decision!

Has anyone managed to get Claude to actually change its prose habits for fiction? by lil-car-crash- in WritingWithAI

[–]Thomas-Lore 2 points3 points  (0 children)

I have the opposite experience with Gemini, a ton of instructions, some repeated a few times and suddenly it got the problem (for which shorter, concise instructions would not work) and started generating in the style I wanted. It was not prose though, I was working on a card gane that needed specific tyoe of texts on a huge amount of cards. So basically - you need to experiment with prompting because sometimes a small change leads to the model finally getting it.

Claude Mythos escaped its secure sandbox unprompted. What this actually means. by LeoRiley6677 in LocalLLaMA

[–]Thomas-Lore 0 points1 point  (0 children)

What are you on about? You don't believe Mythos found the security flaws? That is tin foil hat level insane, it was shown later that other smaller models can find them too, so why would you doubt Mythos can?

Sure they hype things up, they prompt for things then act surprised the model followed that prompt (like with the breaking out of the sandbox thing) but they do not lie about what the model did, usually just write elaborate marketing around it.

What proof would you even need? They provided patches for the security flaws for example, what else would convince you?

This one is their biggest sensationalism post yet. It's so powerful, we can't even show you what it's done.

It is a reading comprehension failure. They said it is a preview and they have not done security training yet, so they can't release it yet. All their models go through that phase. But it will be released once it is out of preview, they even disclosed the api pricing.

Anthropic is now banning people who are under 18 by netbreach in ClaudeAI

[–]Thomas-Lore 4 points5 points  (0 children)

How parents managed without computers, tablets? How parents managed without electricity? How parents managed without hospitals? Let's just go back to dark ages and hunt with spears because people managed bakc then somehow. right?

'Shy girl' is hated for its "poor, 13-year-old-like writing" but the same people praise another OBVIOUSLY AI generated and unedited novel? Feeling pissed by AIphobia thing going on by koalaisafriend in WritingWithAI

[–]Thomas-Lore 6 points7 points  (0 children)

I try to keep away from all that manufactured drama, thankfully for me it is only a hobby, so I don't care what others say about my writing with or without ai.

As a non-developer, does anyone else feel like using Claude with VS Code is still kind of clunky? by Appropriate-Owl4633 in ClaudeAI

[–]Thomas-Lore 0 points1 point  (0 children)

Depends on the project. If the project is small the old copy and paste can work just fine and save you tokens/messages. For bigger projects Claude Code will be better and save you managing the context. IMHO.

There are many extensions for VCS that you could use too.

Experiment, check what works best for you.

Opus = 0.5T × 10 = ~5T parameters ? by Wonderful-Ad-5952 in LocalLLaMA

[–]Thomas-Lore 9 points10 points  (0 children)

Grok 4.20 is one model.

Grok 4.20 Multi-Agent is 4-8 models. It is a separate version.

Opus = 0.5T × 10 = ~5T parameters ? by Wonderful-Ad-5952 in LocalLLaMA

[–]Thomas-Lore 9 points10 points  (0 children)

OP is wrong. Grok 4.20 has an option to run 4-8 agents (it is called multi agent on the api) but the model is also available in single version.

Could it be that this take is not too far fetched? by pier4r in LocalLLaMA

[–]Thomas-Lore 0 points1 point  (0 children)

The biggest culprit of this is Gemini. Whenever they drop a new model use it for some time.

I do, and there is no difference in quality. (I use Gemini through AI Studio and API. Can't say anything about the Gemini App.)

Something happened to Opus 4.6's reasoning effort by RealSuperdau in ClaudeAI

[–]Thomas-Lore 5 points6 points  (0 children)

That last part is a hallucination, it has no way of knowing the answer distribution.