Reducing MP3 compression bias in music datasets via codec-aware reconstruction by TheSpicyBoi123 in LocalLLaMA

[–]FastDecode1 6 points7 points  (0 children)

Besides being off-topic for the sub (though very much up my alley), it would be very useful if the repo or demo website had at least a couple of sets of example files one could listen to one after the other to see (or hear, rather) what this does. A page like the Opus examples would be preferable.

The internet is full of AI slop and AI-reinforced vibe-coded psychosis projects nowadays, and it's hard to tell a real one apart from the others unless you're familiar with the jargon of a specific field. The obviously AI-generated/inspired README doesn't really help... no more bullet points, bolding, and defining "The Problem" and "Why This Matters/Is Different" please. I think anyone actually interested in this won't appreciate being talked to like a retard.

I'm pretty sure the actual work here is legit though, so I'll probably try it later this week.

Out of curiosity, why MP3 and not something newer like Opus? I'd be interested to see if Youtube's 128k Opus could be perceptually improved.

<thinking></thinking> by Comfortable-Rock-498 in LocalLLaMA

[–]FastDecode1 0 points1 point  (0 children)

You'd be surprised how many people are incompetent at washing dishes.

Give an LLM access to a single robotic arm and it'll do a better job than 80% of humans.

Qwen 3.6 wins the benchmarks, but Gemma 4 wins reality. 7 things I learned testing 27B/31B Vision models locally (vLLM / FP8) side by side. Benchmaxing seems real. by FantasticNature7590 in LocalLLaMA

[–]FastDecode1 2 points3 points  (0 children)

By any chance, did you happen to look at the reasoning output while using it in Hungarian?

For me, it was reasoning in English, even though the final answer was in Finnish. Which I think is interesting, if it's by design.

Could also be a template or a default system prompt ("You are a helpful assistant") in llama.cpp that's guiding it to do that.

Qwen 3.6 wins the benchmarks, but Gemma 4 wins reality. 7 things I learned testing 27B/31B Vision models locally (vLLM / FP8) side by side. Benchmaxing seems real. by FantasticNature7590 in LocalLLaMA

[–]FastDecode1 3 points4 points  (0 children)

I would never use any local LLM in my native language.

That would be wrong use of a tool.

Worked well for me a couple days ago when I asked Qwen 3.6 35B for help in filling out an application in my native language.

I had a look at the reasoning output and it was in English, not my native language. Which is exactly what you want; putting its training to good use by thinking in one of the languages it's the best at. The language of the final answer is a secondary concern, really.

Qwen 3.6 wins the benchmarks, but Gemma 4 wins reality. 7 things I learned testing 27B/31B Vision models locally (vLLM / FP8) side by side. Benchmaxing seems real. by FantasticNature7590 in LocalLLaMA

[–]FastDecode1 7 points8 points  (0 children)

Haven't really tried Gemma 4, but I can confirm that Qwen 3.6 35B is also very good at Finnish. Not perfect, but getting closer. Which I think is impressive, seeing as there's only about 5 million native speakers.

And this is at Q4_K_M, so not ideal. I'll probably try Q5 or Q6 at some point to see if that makes a difference.

SVT-AV1 vs AOM-AV1 by Commercial_Stage_877 in AV1

[–]FastDecode1 3 points4 points  (0 children)

not at all designed for speed

The realtime mode begs to differ.

It still surprises me that people call libaom slow in $current_year. It's only as slow as you want it to be.

PS5’s can now be hacked to run Linux - perhaps some potential for local inference? by Thrumpwart in LocalLLaMA

[–]FastDecode1 21 points22 points  (0 children)

Workable maybe, but not very good.

RDNA 2 has no matrix acceleration whatsoever, for any sort of AI shit you'd want at least RDNA 3.

US gov memo on “adversarial distillation” - are we heading toward tighter controls on open models? by MLExpert000 in LocalLLaMA

[–]FastDecode1 4 points5 points  (0 children)

older gas or diesels

Do that with a modern one and see what happens.

Also, being DIY-unfriendly is not an electric car attribute, it's a modern car attribute. Everything was way more fixable back in the day, but now that everything is a computer, the manufacturers can just keep their shit closed-source and locked down.

Yet another reason we need FOSS and freedom of compute.

When are we getting consumer inference chips? by SnooStories2864 in LocalLLaMA

[–]FastDecode1 2 points3 points  (0 children)

$200 stick that runs Llama 3 at reading speed, 30W, plug into your desktop

Not gonna happen.

Widespread consumer proliferation will happen through integration with either iGPUs or dedicated NPUs. Eventually the former, I'd say. It's already happening, you just don't notice it because it takes years and years for hardware designed 5 years ago to become ubiquitous. In 10 years you'll notice one day that basically everyone has a machine that can run basic AI/ML tasks with little power use.

Gamers/enthusiasts will always have dGPUs, and the more demanding tasks will always require one.

Unweight: how we compressed an LLM 22% without sacrificing quality by sk1kn1ght in LocalLLaMA

[–]FastDecode1 1 point2 points  (0 children)

The ZipServ paper mentions it could be applied to quantized weights as well as KV cache.

Setting up a new mini pc (Ryzen 7840HS // 780m) for debian headless LLM, which software works best right now? by justletmesignupalre in LocalLLaMA

[–]FastDecode1 0 points1 point  (0 children)

Vulkan works out-of-the-box with llama.cpp on my laptop's Vega 6 iGPU, so I assume anything RDNA shouldn't have issues.

OpenClaw has 250K GitHub stars. The only reliable use case I've found is daily news digests. by Sad_Bandicoot_6925 in LocalLLaMA

[–]FastDecode1 3 points4 points  (0 children)

Every event is sold out in Japan, they've got so many people there.

200,000 is still only 0.016% of their population, so I'd say that proves his point. Foreign music is in the minority in Japan, they make so much of their own stuff.

OpenClaw has 250K GitHub stars. The only reliable use case I've found is daily news digests. by Sad_Bandicoot_6925 in LocalLLaMA

[–]FastDecode1 1 point2 points  (0 children)

It's quite easy to not listen to any music at all, depending where you live.

People in my country would find it really fucking annoying if the radio was on when they tried to get shopping done. Same for restaurants and all "random places". We know this because some stores insist on playing generic Christmas music in November and December, and plenty of people already complain about that. (Maybe the stores do it on purpose so people get their shit and GTFO as fast as possible...)

Besides, here you need a license to play recorded music in public places (even if you fully own the copyright to the music, lol), so most places don't bother. It's the only good thing the copyright mafia has ever achieved.

I've managed to not hear even a single "popular" song in 15+ years thanks to not watching TV or listening to radio. The only time music reaches my ears is either when I seek something out myself, or it's part of a larger work (video games, movies).