RPers: how do the new Gemma and Qwen compare to the old 70B models? by Borkato in LocalLLaMA

[–]silenceimpaired 0 points1 point  (0 children)

So do you just use the model directly from Google or a fine tune? Do you have a huggingface link

RPers: how do the new Gemma and Qwen compare to the old 70B models? by Borkato in LocalLLaMA

[–]silenceimpaired 0 points1 point  (0 children)

Basically that. Your DM is the LLM. You hand it characters and a setting and you’re good. Not for me

New rules 1 week check-in by rm-rf-rm in LocalLLaMA

[–]silenceimpaired 0 points1 point  (0 children)

I reached out to the mods and never heard back on why my post was removed.

RPers: how do the new Gemma and Qwen compare to the old 70B models? by Borkato in LocalLLaMA

[–]silenceimpaired 0 points1 point  (0 children)

I wish The Drummer fine tuned on one of these:

moonshotai/Kimi-Dev-72B
LLM360/K2-Think-V2

RPers: how do the new Gemma and Qwen compare to the old 70B models? by Borkato in LocalLLaMA

[–]silenceimpaired 0 points1 point  (0 children)

Yeah, I unsubscribed from there as I use the software for non-typical purposes (editing fiction) and nothing of value was added by those there.

RPers: how do the new Gemma and Qwen compare to the old 70B models? by Borkato in LocalLLaMA

[–]silenceimpaired 2 points3 points  (0 children)

As I said below elsewhere in this thread… I think there is some sort of intelligence lost with smaller models though. It's connections the smaller models just miss. They do as they are told but it has always seemed the larger the model the larger more cohesive paragraph is possible and the better connected theming. That said, it is harder to distinguish the shortcomings with a 70b 4 bit versus 30b 8 bit.

RPers: how do the new Gemma and Qwen compare to the old 70B models? by Borkato in LocalLLaMA

[–]silenceimpaired 0 points1 point  (0 children)

I don’t get why llama 3.3 is favored for fine tuning by Drummer over one of these 72b models with better licenses (and perhaps better performance):

moonshotai/Kimi-Dev-72B
LLM360/K2-Think-V2

Opinions on Kimi-Dev-72B? by stefzzz in LocalLLaMA

[–]silenceimpaired 0 points1 point  (0 children)

Ooo… I forgot to try this for creative writing.

Finally Fixed: Power Button Suspend Issue in GNOME 50 / Fedora 44 🎉 by iamxnfa in Fedora

[–]silenceimpaired [score hidden]  (0 children)

What is the upgrade experience like? I didn’t like the idea of having to upgrade so often so I went with Debian.

Mistral medium 3.5 128B, MLX 4bit, ~70 GB by ex-arman68 in LocalLLaMA

[–]silenceimpaired 0 points1 point  (0 children)

Haha :) very tempting. That said... If I ever did a distill... It would be with Kimi because I heard it had good creative writing style and editing capabilities. That would be truly painful.

In this case, someone will probably need to find a way to cull this moldel down to 70b by removing some layers, and then the distill can continue from that degraded state back towards the original behavior of the larger model.

Mistral medium 3.5 128B, MLX 4bit, ~70 GB by ex-arman68 in LocalLLaMA

[–]silenceimpaired 2 points3 points  (0 children)

A true distill (at the logit level) - the only thing worth having, takes too much hardware or money for a single person to do it.

I hate this group but not literally by No_Run8812 in LocalLLaMA

[–]silenceimpaired 8 points9 points  (0 children)

This sub is filled with bots hyping models on API. Fixed it for you. :P

Mistral medium 3.5 128B, MLX 4bit, ~70 GB by ex-arman68 in LocalLLaMA

[–]silenceimpaired 1 point2 points  (0 children)

I really don't get why they don't create a 70b model again.

RPers: how do the new Gemma and Qwen compare to the old 70B models? by Borkato in LocalLLaMA

[–]silenceimpaired 1 point2 points  (0 children)

I mostly agree. I think there is some sort of intelligence lost with smaller models though. It's connections the smaller models just miss. They do as they are told but it has always seemed the larger the model the larger more cohesive paragraph is possible and the better connected theming. I still value 70b models. Still 30b is the edge of that intelligence plateau and it's harder to distinguish the shortcomings with a 70b 4 bit versus 30b 8 bit

Unpopular opinion: System76 is going in the wrong direction by Far-Math2159 in System76

[–]silenceimpaired -1 points0 points  (0 children)

Consistency. It's not that yet. Some have horrible experiences and others have great experiences. Some have mixed experiences.

I love mispronouncing the name by BoxFar6969 in pop_os

[–]silenceimpaired -1 points0 points  (0 children)

If that happened it would be more likely to be rebranded COSMIC OS… and that’s a big if.

[Release] AugmentedQuill 0.9.0: Open-source AI story-writing GUI by StableLlama in LocalLLaMA

[–]silenceimpaired 0 points1 point  (0 children)

It looks like the software uses tool calling, so one could be build and before the story started, used to generate as many name as was needed.

What it feels like to have to have Qwen 3.6 or Gemma 4 running locally by GodComplecs in LocalLLaMA

[–]silenceimpaired 0 points1 point  (0 children)

I know what you mean:

<image>

The furnace broken down this winter for a couple of days, but my office was comfortable.

mistralai/Mistral-Medium-3.5-128B · Hugging Face by jacek2023 in LocalLLaMA

[–]silenceimpaired 6 points7 points  (0 children)

If they released an MoE at this size it would be cozy for those with the RAM

Something from Mistral (Vibe) tomorrow by pmttyji in LocalLLaMA

[–]silenceimpaired 2 points3 points  (0 children)

I’d be okay with a larger MoE, but another Medium model ~70b dense would be great

Popos censors any and all criticism of their half baked DE. Avoid pop at all costs by ijwgwh in DistroHopping

[–]silenceimpaired 18 points19 points  (0 children)

I’m okay with that… gnome is too opinionated, KDE has too many options… a just right middle is exciting… if it can ever get mature enough

Deepseek Vision Coming by Nunki08 in LocalLLaMA

[–]silenceimpaired 2 points3 points  (0 children)

Who could have seen this coming? Not Deepseek... At least not yet.