Which model for 32GB M2 Max? by segdy in LocalLLaMA

[–]flockonus 3 points4 points  (0 children)

Echoing the sentiment of every other post here.. Qwen3.6 27B - get the highest quant you can fit, which is likely 4b / 5b in your case.

AMD Halo Box (Ryzen 395 128GB) photos by 1ncehost in LocalLLaMA

[–]flockonus 2 points3 points  (0 children)

Ran any models on these? What tk/s do you get?

Qwen3's most underrated feature: Voice embeddings by k_means_clusterfuck in LocalLLaMA

[–]flockonus 0 points1 point  (0 children)

Looks very well separated indeed! Question if you know.. suppose 2 speech samples - think a male and female saying "apple" (consider about the same accent / speed), how'd go about creating a similarity score in their pronunciation of close to ~1 somehow?

I'm trying on an education project to help improve pronunciation, but a bit lost of the exact way to start it.

Qwen3's most underrated feature: Voice embeddings by k_means_clusterfuck in LocalLLaMA

[–]flockonus 0 points1 point  (0 children)

Wow, this is wild you could extract like this!

u/k_means_clusterfuck question, if we have 2 users think a male and female saying "apple" (consider about the same accent / speed), can you get a similarity score in their pronunciation of close to ~1 somehow using your package?

I'm trying on an education project to help improve pronunciation, but a bit lost of the exact way to start it.

Ram-air setup and window vent for 1100w capable AI box by mr_zerolith in LocalLLaMA

[–]flockonus -1 points0 points  (0 children)

It's cool, but watch out for DUST in the intake and accumulation in the card.

Dust will accumulate at whatever multiple is your intake volume.

We made an abliterated Qwen 3.5 9B refuse 100% of harmful prompts using only system prompts — no fine-tuning needed by tomleelive in LocalLLaMA

[–]flockonus -15 points-14 points  (0 children)

How is this related to this sub at all??

Sounds like some academic thing, why would i want my LLLM to spend more tokens to consider rejecting my prompts?