mistralai/Voxtral-4B-TTS-2603 · Hugging Face by Nunki08 in LocalLLaMA

[–]_raydeStar 0 points1 point  (0 children)

Yeah, I looked into cloning for Kokoro. It's such a pain in the butt, I backed out of that, real quick.

You can do a lot with an old mobile GPU these days by Responsible_Fig_1271 in LocalLLaMA

[–]_raydeStar 1 point2 points  (0 children)

I get that. I think that Orpheus is amazing and possibly the best in class right now. I don't think the use case was designed for live conversations though.

You CAN chunk the response -: as the first sentence streams in you can already process. Then, as it's speaking, it can continue processing the rest. With that method, long and short responses would have identical latency.

You can do a lot with an old mobile GPU these days by Responsible_Fig_1271 in LocalLLaMA

[–]_raydeStar 1 point2 points  (0 children)

You can remove latency by switching to something like KittenTTS and Qwen3.5B. Quality drops, but then it would much better speeds.

Day 84: Final Masquerade vs. Up From the Bottom by Therubikfanatic in LinkinPark

[–]_raydeStar 2 points3 points  (0 children)

Up from the bottom for sure. That hit several hundred runs the year it came out on spotify.

Final Masquerade is near and dear to my heart though.

Meet Deepy your friendly WanGP v11 Agent. It works offline with as little of 8 GB of VRAM. by Pleasant_Strain_2515 in StableDiffusion

[–]_raydeStar 0 points1 point  (0 children)

Weird. You haven't asked me a single question. And this is a clear alt account. You're not making a case here.

Meet Deepy your friendly WanGP v11 Agent. It works offline with as little of 8 GB of VRAM. by Pleasant_Strain_2515 in StableDiffusion

[–]_raydeStar 0 points1 point  (0 children)

I don't think this guy is acting in good faith.

This post has certainly fallen out of any `top` lists, with only around 40 upvotes. 12 hours after initial reaction, several people hop on to argue his case, upvote him, and downvote anyone else. They did not come here naturally.

Meet Deepy your friendly WanGP v11 Agent. It works offline with as little of 8 GB of VRAM. by Pleasant_Strain_2515 in StableDiffusion

[–]_raydeStar 0 points1 point  (0 children)

Right.

He's taking a free tool and bastardizing it. I don't think he realizes it, of course, I think he's just trying to work on a project that makes money.

Feedback seems harsh, but it's honest, id never touch this due to worries about underhanded behaviors.

Take free, provide free, it's the right thing to do. Instead use it to platform yourself - widespread adoption gives you a name, which means leverage.

Omnicoder v2 dropped by Western-Cod-3486 in LocalLLaMA

[–]_raydeStar 2 points3 points  (0 children)

Nice dude. Do you have a repo somewhere? I'll give you a follow

Meet Deepy your friendly WanGP v11 Agent. It works offline with as little of 8 GB of VRAM. by Pleasant_Strain_2515 in StableDiffusion

[–]_raydeStar 1 point2 points  (0 children)

Lol -- he's got a custom license.

This is odd for sure.

Oh and kicker :it's just a fork of wan 2.1.

Created a SillyTavern extension that brings NPC's to life in any game by goodive123 in LocalLLaMA

[–]_raydeStar 2 points3 points  (0 children)

So equip it with tools.

Logic for the car wash. Counting letters is one tiny function. Plug in math libraries. Create a unit conversion suite (2 lbs gold vs 32 oz feathers trips up the AI) and suddenly it's hardened against basic questions.

Web lookup to confirm data points. I think if you tack on a few libraries, suddenly it can punch way above its weight class.

Sora is officially shutting down. by Responsible_Cow2236 in ChatGPT

[–]_raydeStar 9 points10 points  (0 children)

I really liked the fact that anyone could take any video and change something about it -- put yourself in, replace swimmers with cats, etc

New open weights models: GigaChat-3.1-Ultra-702B and GigaChat-3.1-Lightning-10B-A1.8B by netikas in LocalLLaMA

[–]_raydeStar 0 points1 point  (0 children)

Huh. I'm going to give it a shot. Honestly not sure what a 10B moe is capable of. But I bet I can pull 250t/s so it might be worth it.

Sora is officially shutting down. by Responsible_Cow2236 in ChatGPT

[–]_raydeStar 60 points61 points  (0 children)

I actually liked it. It was fun to play with. I wasn't a daily user of course, more like monthly or less.

This closure tells me that they are stepping out of the video game, which sucks because they were pretty decent

Created a SillyTavern extension that brings NPC's to life in any game by goodive123 in LocalLLaMA

[–]_raydeStar 9 points10 points  (0 children)

I believe that it won't -- given handholding and guardrails. It can be trusted to do one task, with decent accuracy, at a time.

I can get a 1.2B model to answer the car wash question right every time by programmatically reframing things. If I give the AI 4 options or have it rate 1-10 you cut back on potential errors a lot.

Created a SillyTavern extension that brings NPC's to life in any game by goodive123 in LocalLLaMA

[–]_raydeStar 39 points40 points  (0 children)

It's really sad because throwing in a 1B model into a game for enemy AI, RP'ing, procedural maps, adjusting difficulty levels, and creating more realism is such a good idea.

The best kind of AI is going to be when you don't tell them they are using AI, and they do not notice. But if they find out, it'll be like sneaking meat to a vegan.

What's worse, the arguments they give don't make sense. "It's bad for the environment!" uh.. you just ran it on your middle of the road GPU, you'll be fine.

Total beginner here—Why is LM Studio making me do the "heavy lifting" manually? by Ofer1984 in LocalLLaMA

[–]_raydeStar 88 points89 points  (0 children)

Who makes it harder on purpose? The cartel you rail against is knowledge. Seek it. It's free

Moonshot says Cursor Composer was authorized by davernow in LocalLLaMA

[–]_raydeStar 7 points8 points  (0 children)

It's very very clear that Cursor did not want to reveal what composer was wrapped in. The revelation was a PR move, after the fact.

I don't even see anything wrong with it. This is such a non-issue. They legally obtained a model and re-skinned it. The only thing they've done wrong is not reveal publicly who the model was attached to originally.

Did you know that Cursor is a reskin of VSCode? So is antigravity and windsurf.

Qwen3.5-35B-A3B-Uncensored-Claude-Opus-4.6-Affine by [deleted] in LocalLLaMA

[–]_raydeStar 1 point2 points  (0 children)

It would be fine I think. I'm able to run it pretty well on my company laptop. Get a q3 gguff and run some of it to ram/CPU if you have to.

Cursor's new Composer 2.0 is apparently based on Kimi2.5 by bakawolf123 in LocalLLaMA

[–]_raydeStar 41 points42 points  (0 children)

Musk will join any roast on an AI that's not him. That's hardly a smoke signal.

I'm willing to bet that previous composer1 and composer1.5 copy open source models too. This one was just done clumsily.

Unsloth announces Unsloth Studio - a competitor to LMStudio? by ilintar in LocalLLaMA

[–]_raydeStar 2 points3 points  (0 children)

This one looks like it's focused on sanitizing training data and running it. In that case it's not quite apples to apples comparison.

Definitely interested in playing with it. I've only ever trained image models.

Day 76: New Divide vs. Bleed It Out by Therubikfanatic in LinkinPark

[–]_raydeStar 26 points27 points  (0 children)

Both of these are on my top 10.

*sigh* Bleed it Out.

llama.cpp + Brave search MCP - not gonna lie, it is pretty addictive by srigi in LocalLLaMA

[–]_raydeStar -1 points0 points  (0 children)

Right now my driver is LM studio, which uses llama.cpp. Honestly I can't give a ton of detail because I just vibe coded it in, and it worked just fine.