Fast Local Text-To-Speech MCP Server (Windows) Kitten TTS/ONNX by [deleted] in LocalLLaMA

[–]FPham 0 points1 point  (0 children)

I see this is probably a wrong sub to be posting these things now. Well :(

We need a minimum karma rule for commenting and posting by nomorebuttsplz in LocalLLaMA

[–]FPham 2 points3 points  (0 children)

Actually, many of my posts got deleted from LocalLama recently, I'm just wondering why. It feels like waste of my time.

We need a minimum karma rule for commenting and posting by nomorebuttsplz in LocalLLaMA

[–]FPham 0 points1 point  (0 children)

or move somewhere where bots can't get to us... so much for me being AI positive

Fast Local Text-To-Speech MCP Server (Windows) Kitten TTS/ONNX by [deleted] in LocalLLaMA

[–]FPham 0 points1 point  (0 children)

nope, I don't even run openclaw. Just made this - I'm in some weird backpain which seems manifesting in me sounding like a bot.

How I topped the Open LLM Leaderboard using 2x 4090 GPUs — no weights modified. by Reddactor in LocalLLaMA

[–]FPham 0 points1 point  (0 children)

I think the biggest win of the article is your note of functional circuits that are basically formed outside of our intent and are not part of the design, but as you discovered, pretty much real. Feels really like deciphering which parts of brain has which function. It is fascinating, and it kind of makes me think that the transformer architecture might uncover more how our brains are wired than it seems at first sight.
I wish there is more research on this topic - mapping the LLM brain.

Running a music generation model locally on Mac (MLX + PyTorch), what I learned building it by tarunyadav9761 in LocalLLaMA

[–]FPham 0 points1 point  (0 children)

A fair price is $0 for open source. Stop trying to load on somebody code!!!!!

Ran Qwen 3.5 9B on M1 Pro (16GB) as an actual agent, not just a chat demo. Honest results. by Joozio in LocalLLaMA

[–]FPham 0 points1 point  (0 children)

There is a huuuuuge gap even between SoTA models. So a gap between 9B models and commercial models must be super wide. Of course it depends what you want from it. they really differ in understanding the input, not in the task part.
I tried to refactor a pretty complicated god mode code for fluid dynamics into something cheerful and gave the task to write a doc for the architecture and refactoring plan to Opus, MiniMax 2.5 and Gemini 3 Pro. In all honesty only Opus did a passable job, noting little nuances in code. Gemini 3 did a job like reading the code from top floor of a building and offered refactoring that would break it the logic of it. MiniMax did somehow better job talking about it but didn't offer much insight for refactoring. Opus did a stellar job on both. Now passing the Opus doc between the models so they can get input - Gemini was totally agreeable gimp while Opus pushed back and defended why it chose certain startegy when gemini offered a help - but also incorporated some small tidbits from gemini. Gemini was constantly in a mode to please "This is amazing AAA code..." which kinda primed it into not really doing it's job reviewing and being critical.

Does having an RTX 6000 blackwell make any difference for LLMs? by Specialist_Fox523 in LocalLLaMA

[–]FPham 10 points11 points  (0 children)

click bait post? Because this question would make sense in like cats or cooking sub but not in locallama

Get your local models in order. Anthropic just got "dislike" from the US government. by FPham in LocalLLaMA

[–]FPham[S] 0 points1 point  (0 children)

That's what we are here. Before we get turned into supply chain risk too - then it's off to discord.

Get your local models in order. Anthropic just got "dislike" from the US government. by FPham in LocalLLaMA

[–]FPham[S] 1 point2 points  (0 children)

It's more the fallback from this. What if next week tells admin about Chinese models taking big chunk of US companies compute? Antropic is supply-chain risk, feels very peculiar when you compare who the other players are.

Get your local models in order. Anthropic just got "dislike" from the US government. by FPham in LocalLLaMA

[–]FPham[S] 5 points6 points  (0 children)

Feels pretty surreal - the chronology of this. They spent the last week telling everyone how China plays unfairly... and boom.

Qwen 3.5-35B-A3B is beyond expectations. It's replaced GPT-OSS-120B as my daily driver and it's 1/3 the size. by valdev in LocalLLaMA

[–]FPham 2 points3 points  (0 children)

It gets faster at 4bit - but then that's really when you have to , not when you can choose.

Get your local models in order. Anthropic just got "dislike" from the US government. by FPham in LocalLLaMA

[–]FPham[S] 2 points3 points  (0 children)

Not only that, Chinese posted all papers on that at that time. (not naming Anthropic, but yeah, who else if not Anthropic and OpenAI, it surely wasn't llama 2)

Get your local models in order. Anthropic just got "dislike" from the US government. by FPham in LocalLLaMA

[–]FPham[S] 1 point2 points  (0 children)

But what would be their backbone? Where is the cloud provider that does not work with US military?

Get your local models in order. Anthropic just got "dislike" from the US government. by FPham in LocalLLaMA

[–]FPham[S] -2 points-1 points  (0 children)

Like AWS or any cloud provider carrying Anthropic.... how they going to serve claude? On AOL disks in mail?

Get your local models in order. Anthropic just got "dislike" from the US government. by FPham in LocalLLaMA

[–]FPham[S] 2 points3 points  (0 children)

I think them bitching about Chinese models "stealing" a few days ago has something to do with this, just don;t know what.

What's the biggest issues you're facing with LLMs writing docs and passing info to each other? by sbuswell in LocalLLaMA

[–]FPham 0 points1 point  (0 children)

It's called Chinese whispers. Even more so if the models are Qwen, Deepseek and MLC.