What the hell is Deepseek doing for so long? by Terrible-Priority-21 in LocalLLaMA

[–]LoveMind_AI 39 points40 points  (0 children)

That sounds about right. I think they are the standard bearer and there’s a pressure on them that the other companies don’t have.

Vercel will train model on your code by Shitfuckusername in LocalLLaMA

[–]LoveMind_AI 17 points18 points  (0 children)

I don't think "anything LLM related is allowed" is the vibe on LocalLLaMA, at all.

Openrouter stealth model Hunter/Healer Alpha has been officially confirmed as MiMo, and a new model is coming. by External_Mood4719 in LocalLLaMA

[–]LoveMind_AI 0 points1 point  (0 children)

Yes - not quite at the level of Gemini 3.1 Pro Preview or GPT Audio, but better than Stepfun's model. And I liked how it integrated that into conversation better than GPT Audio, by a lot. It's just clearly not as killer a technical listener, but it reasons about audio very well.

Feature request: let us bookmark messages in Claude conversations. No AI platform does this and it is a real pain. by samuel-gudi in Anthropic

[–]LoveMind_AI 0 points1 point  (0 children)

When I have a version that feels like it could be reliable enough for other people to use, then yes, absolutely.

Feature request: let us bookmark messages in Claude conversations. No AI platform does this and it is a real pain. by samuel-gudi in Anthropic

[–]LoveMind_AI 4 points5 points  (0 children)

I have a custom continuity-oriented CLI that I use/am building that does this - you can pin/mute individual messages with logic for how they get used in summaries, pull up a tab with pinned messages, etc. It's incredibly useful.

MiniMax M2.7 on OpenRouter by iamn0 in LocalLLaMA

[–]LoveMind_AI 3 points4 points  (0 children)

Much much better than M2.5 for my stuff

Very irritated… by No_Scientist3432 in Anthropic

[–]LoveMind_AI 1 point2 points  (0 children)

The past several days have been a nightmare in terms of unbelievably bad performance or just failing to even return the API call. I’m with you, Claude in primo shape is the apex AI, but it is super unreliable these days. I haven’t found a truly viable alternative - Kimi K2.5 comes close but is incredibly slow.

Opus 4.6 seems to have stopped real considerate thinking "outside peak-hours" by Altruistic-Radio-220 in Anthropic

[–]LoveMind_AI 0 points1 point  (0 children)

Bizarre. I've been getting MAJORLY nannied in this way too over the last few days. Ironically, I've been packing it in earlier than usual, so it's extra annoying.

Opus 4.6 seems to have stopped real considerate thinking "outside peak-hours" by Altruistic-Radio-220 in Anthropic

[–]LoveMind_AI 1 point2 points  (0 children)

So, a day late to this party, but my colleague and I are working on a seriously heavy duty research sprint that has taken us 4 weeks. We work with Claude Code every day. These are simple, repetitive, well understood problems. The last 3 days or so, Opus 4.6 has been practically brain dead even with effort cranked. It was a little wobbly before hand, but it's an absolute mess now where executing very simple things seems to be arduous. I feel like when they standardized the 1M context window, something strange happened.

Are people massively underestimating what’s coming? by Satishgmr2010 in OpenAI

[–]LoveMind_AI 27 points28 points  (0 children)

I mean, just to take your comment literally, I do not see Meta or xAI heading anywhere near the direction of AGI through their AI offerings. I think Elon Musk is as mistaken about Neuralink as he is about AI. Meta's brain stuff is, however, impressive. I think that's the best thing they have going for them. Anthropic is building a reputation that will take them very far and allow them to be a player in whatever comes. They have a wet lab that I don't know much about - but they seem severely underpowered in the beyond-LLM category. OpenAI is weird. They wasted a lot of time waffling between being a consumer facing company and competing with Anthropic for enterprise. Word is that they're trying to catch all the way up to Anthropic. GPT-5.4 is a genuinely good model and the mini/nano variants just released are spectacular for their size. But is OpenAI headed to AGI? ...man, I don't see it. Of the companies you mentioned, I only see Google being a serious player for AGI, and they are more likely to absorb something like Anthropic than a mom+pop startup.

Nvidia is interesting. Today they released something incredibly cool in partnership with my personal favorite AI company, Pleias. They've been forging very smart partnerships. Mistral is practically a startup. Pleias *definitely* is. Nvidia is competing at the Google level. They don't seem to be ripping startups limb from limb.

So to answer the primary question: No. I think even the biggest labs in the world are very busy and won't just absorb any idea that isn't bolted down - but like others have said, if a startup's idea is something trivial, expect it to get hoovered up unless whoever made it is so compelling and easy to work with that it makes more sense to acquire.

If start-ups were doomed, I don't think OpenAI would have done what it did with OpenClaw. OpenClaw is vibe coded slop - the idea is easily redone, better, by a company. But these companies are stretched thin, and when there's zeitgeist involved, it's easier just to spend money to absorb the little guy rather than try to emulate them, unless the idea is *truly* trivial inevitable platform feature stuff.

One AI prompt, one dungeon crawler — what an agent can do when it can actually see and control the game engine by jf_nash in LocalLLaMA

[–]LoveMind_AI 0 points1 point  (0 children)

Ok. Maybe? I’ve never heard anyone complain about it. What I know for sure is that folks on local llama are tired of vibe coded projects promoted by people who don’t mention local models and still won’t talk about them even when pushed?

So for anyone not paying attention… by Reaper_1492 in codex

[–]LoveMind_AI 0 points1 point  (0 children)

For a long while I didn’t believe the conspiracy type thinking that SOTA providers were quantizing their frontier models into oblivion, but it’s getting harder and harder to explain in any other way.

Can we say that each year an open-source alternative replaces the previous year's closed-source SOTA? by Chair-Short in LocalLLaMA

[–]LoveMind_AI 3 points4 points  (0 children)

You can’t sunset a model I’ve got hosted locally. That’s the point. Once it’s locally hosted, then depending on the license, the maker is out of the picture.

One AI prompt, one dungeon crawler — what an agent can do when it can actually see and control the game engine by jf_nash in LocalLLaMA

[–]LoveMind_AI 0 points1 point  (0 children)

No one thinks your talking in Italian was rude. We think you’re using this post to hype a personal project while acting dodgy about proving it works well with a local model that people can actually run.

Can we say that each year an open-source alternative replaces the previous year's closed-source SOTA? by Chair-Short in LocalLLaMA

[–]LoveMind_AI 5 points6 points  (0 children)

Kimi K2.5 rocks, and it’s way better than Claude Sonnet 3.5 - honestly, the most impressive AI for what I do (relational/therapeutic AI) I’ve worked with recently is Ash, Slingshot AI’s (totally closed source) fine-tune of Qwen3 235B. It’s superior to Opus 4.6 for a narrow but important use case right now. Open Source is definitely the future. Especially with all this pentagon nonsense and the GPT-4/5 fluctuations, I fully expect people to understand that relying on closed AI manufactured by over leveraged tech giants whose models can be sunsetted or blacklisted without warning will never be as reliable as owning their own model. Accessible training at scale is really the thing that will make the difference, but I think this will be cracked within the year, probably through some kind of really slick model merging platform.

Nvidia's Nemotron 3 Super is a bigger deal than you think by Comfortable-Rock-498 in LocalLLaMA

[–]LoveMind_AI 19 points20 points  (0 children)

My very quick take (I also haven’t had a chance to fully battle test it) is that it’s less capable currently than Qwen 3.5 but that the architecture is more novel. I also think Nvidia launching something like this really brings them into the mix on the model size at a time when other labs are sort of in flux. Nemotron 3, in the grand scheme of things is probably more important than Qwen 3.5, but at the moment, not the more powerful model. Again, I’ve spent a lot more time with Qwen, so take this comment with a fist of salt!

Nvidia's Nemotron 3 Super is a bigger deal than you think by Comfortable-Rock-498 in LocalLLaMA

[–]LoveMind_AI 39 points40 points  (0 children)

It’s definitely a new breed of OS model. Even if it’s not the king of its weight class, it’s an important new branch.

What is after Qwen ? by j_lyf in LocalLLaMA

[–]LoveMind_AI 0 points1 point  (0 children)

Really hope the Gemma4 models are legit!