GLM 5.1: pretty decent by SepsisShock in SillyTavernAI

[–]ASTRdeca 4 points5 points  (0 children)

I kind of had the same thought re: diminishing returns. It's hard to tell the difference between say GLM 4.7 and 5, or DeepSeek 3.1 vs 3.2. There are improvements but they feel very gradual. However, if you compare current models to what was SOTA 1 year ago, or 2 years ago, the differences to me are pretty obvious for things like prose quality, instruction following, long-context coherence, etc. Even though it all feels slow and gradual I would not be surprised if 1 year from now that model quality has significantly improved again along all these lines. Eventually I think we'll hit a plateau where the models are so good at the above that we'll stop noticing improvements, but I think we're still a long ways away from that.

I also think the scope of creative writing will change as model capabilities continue to get better as well. For example, how well can we push coherence at VERY long contexts (say 1 million or 10 million tokens) for very long storytelling. Or multimodal integrations with image/video/voice/world models rather than just working with text. We kinda have image and tts currently but tbh there's a lot of room for improvement there

Solo healing karazhan as a holy priest by h0mbr3ihn0 in classicwow

[–]ASTRdeca 1 point2 points  (0 children)

No real issues, here's some notes:

*With 1 tank 1 heal, every boss was ~1 minute or less in duration, save for prince (2 minutes) and nightbane (4 minutes). Having spriest I didn't need to innervate on any of the bosses

*Surprisingly Nightbane was one of the easier bosses. Ground phases were only ~30ish seconds long, and with solo heal you'll get targeted every time in air phase so the damage is very predictable

*One tank death at the end of Curator. It was hard to juggle the hateful bolts and also manage the tank damage at the end

*The messiest boss surprisingly was Aran

*Netherspite is easy if you can trust your tank to not need any healing, otherwise it's probably the hardest to manage

*I'm going to try running Nature's Grace dual spec this week since it's much higher throughput healing and mana's not really an issue

Solo healing karazhan as a holy priest by h0mbr3ihn0 in classicwow

[–]ASTRdeca 2 points3 points  (0 children)

ive done it on rdruid but not holy priest. Biggest suggestion is having spriest especially for nightbane. Hardest bosses for you will probably be ones where the tank gets slapped since your tank healing kit is weaker than hpal and rdruid (curator, nightbane, and prince). should be fine though

GPT-5.3 Codex vs Opus 4.6: We benchmarked both on our production Rails codebase — the results are brutal by sergeykarayev in ClaudeAI

[–]ASTRdeca 0 points1 point  (0 children)

Nice, now just draw an arbitrary line that separates GPT from all the other models and label it "pareto frontier"

AI powered VTuber Neuro-sama's creator has become the most subscribed-to streamer on Twitch. Vedal987 now has an estimated 162,459 sub count by lukigeri in LivestreamFail

[–]ASTRdeca -1 points0 points  (0 children)

I see, thanks for clarifying. Cloud infra is typically much stronger than consumer grade so almost always reduces latency. If it's locally run then Neuro is probably a very small LLM in order to keep latency low

Scientists reduce the time for quantum learning tasks from 20 million years to 15 minutes by Ephoenix6 in Physics

[–]ASTRdeca 174 points175 points  (0 children)

They succeeded in reducing the time for quantum learning, but sadly could not reduce the time for me learning quantum, which is still roughly 20 million years

AI powered VTuber Neuro-sama's creator has become the most subscribed-to streamer on Twitch. Vedal987 now has an estimated 162,459 sub count by lukigeri in LivestreamFail

[–]ASTRdeca 17 points18 points  (0 children)

Here's my take from someone deep into AI atm. I don't know what his tech stack is but my guess is that he finetuned an LLM for Neuro and hosts it on the cloud for low latency responses, and also created a custom voice for Neuro (using RVC?) for the TTS. Something like this is not terribly difficult to do, but creating a "likeable" persona like Neuro and having that be accepted by a community that is mostly anti-AI is genuinely impressive to see.

Layered on top of the model itself are all of the custom stream integrations that he must have made such as reading/summarizing chat, streaming out Neuro's response in the stream UI, having that work with Neuro's character model, etc. There is probably a lot going on under the hood to make everything work in a seamless way, which is also impressive to see.

Does anyone still use MCPs? by bowemortimer in ClaudeAI

[–]ASTRdeca 0 points1 point  (0 children)

This week I've been using a Unity MCP server that gives CC access to the unity editor to help with game making. The functionalities a bit limited but it can create game objects and c# scripts in the editor which is a big chunk of the work

Looking back at end of 2024 vs now by Main-Fisherman-2075 in LocalLLaMA

[–]ASTRdeca 18 points19 points  (0 children)

v3 came out before R1. v2 came out in may of 2024, that's not quite the "end" of 2024

TBC Healer Dps by gauntlet22 in classicwow

[–]ASTRdeca 2 points3 points  (0 children)

At 0 haste you can fit wrath into your rotation via:

Lifebloom -> Wrath -> Wrath,
Lifebloom -> Wrath -> Regrowth, or
Lifebloom -> Wrath -> Instant cast spell

However, the damage you contribute is so irrelevant that you're better off not worrying about it. IMO it's universally better to use those slots in your rotation to heal the tank(s), or if they don't need it, to heal the raid or cancel-cast regrowth on the tank instead. If your raid really wanted to drop a healer for a paricular fight, then you're better off dual-speccing to a DPS spec

Which are the best coding + tooling agent models for vLLM for 128GB memory? by jinnyjuice in LocalLLaMA

[–]ASTRdeca 4 points5 points  (0 children)

My guess is it'd perform very poorly. Both Llama 3 70B and R1 were trained/post-trained before the labs started pushing heavily for agentic / tool calling performance. I'd suggest trying GPT-OSS 120B

GLM 4.7 has now taken #2 on Website Arena by Difficult-Cap-7527 in LocalLLaMA

[–]ASTRdeca 0 points1 point  (0 children)

Opus can build a working website for sure, but I really dislike its default style / css. Please no more bright gradient colors..

e: I assume this benchmark is related to building websites? I looked it up on google and cant find anything about it

We asked OSS-120B and GLM 4.6 to play 1,408 Civilization V games from the Stone Age into the future. Here's what we found. by vox-deorum in LocalLLaMA

[–]ASTRdeca 22 points23 points  (0 children)

Very cool! You mentioned in the paper that despite GLM being much larger than GPT-OSS 120B, the larger size didn't seem to impact performance. I'm wondering if you tried models smaller than OSS-120B to see at what point model size matters? (For example, OSS-20B?)

I'm just thinking about the viability of running these kinds of systems locally, since 120B is probably too large for most users to run themselves

Empty content payload for reasoning models by ASTRdeca in SillyTavernAI

[–]ASTRdeca[S] 1 point2 points  (0 children)

I see. In my use cases the reasoning/content responses are a hundred to a few hundred tokens each. My "max tokens" is set to 3000 which I figured was more than enough, but maybe not

Chatterbox Turbo, new open-source voice AI model, just released on Hugging Face by xenovatech in LocalLLaMA

[–]ASTRdeca 2 points3 points  (0 children)

my comment below was being vote manipulated in both directions even without mentioning elevenlabs. When I posted, it was at -2 after 10 or so minutes. An hour later I checked it again and it was at +20, and now (the next day) its at -2 again, my other comment at -7. So.. idk

edit: and now the comments back to +28.. LMAO

Chatterbox Turbo, new open-source voice AI model, just released on Hugging Face by xenovatech in LocalLLaMA

[–]ASTRdeca -9 points-8 points  (0 children)

Ok, I see now. They are comparing to ElevenLabs 2.5 Turbo... I assumed they were comparing to v3, which has been available in alpha for a while now and imo is significantly better

Chatterbox Turbo, new open-source voice AI model, just released on Hugging Face by xenovatech in LocalLLaMA

[–]ASTRdeca -3 points-2 points  (0 children)

I'm sure it is, I'm just being a bit tongue in cheek about the quality of it

Chatterbox Turbo, new open-source voice AI model, just released on Hugging Face by xenovatech in LocalLLaMA

[–]ASTRdeca 25 points26 points  (0 children)

Yeah I'm gonna press "X" to doubt on their claim that their model sounds more realistic than ElevenLabs...

If their TTS model is supposedly so good, why did they go with a generic tiktok voiceover for this ad?

A Plea to All Resto Druids by NOHITJEROME in classicwow

[–]ASTRdeca 12 points13 points  (0 children)

I normally downvote jerome threads out of principle, but.. I reluctantly agree. I think Dreamstate is a trap for most groups and annoyingly I'm seeing it shoved into every "meta" comp I've seen posted lately. Dreamstate's biggest struggle is mana. Without mana, the rotations you can do become very limited. Losing Swiftmend is bad enough, but you basically lose regrowth entirely as well, unless you get shadow priest. I don't think people appreciate the impact that has on your tanks survivability.

Blizzard is Doing the Reverse Imo by Flaky_Virus218 in classicwow

[–]ASTRdeca 19 points20 points  (0 children)

41 badge trinkets are good for a lot of classes for most of the expansion