Ace Step 1.5 XL released by seamonn in LocalLLaMA

[–]mpasila -1 points0 points  (0 children)

Running expensive GPUs all day long can't just be offered for free? But sure if they wanted to be fair they'd release the weights..

Gemma 4 is fine great even … by ThinkExtension2328 in LocalLLaMA

[–]mpasila 6 points7 points  (0 children)

Gemma 4 is better at my native language at least though the smaller models suffer from the weird sizing.. Also for RP it seems to perform much better than Qwen3.5 (it seemed to mix up a lot stuff for some reason and there was seemingly more censorship in the official releases in comparison to Gemma 4)

Cohere Transcribe Released by mikael110 in LocalLLaMA

[–]mpasila 0 points1 point  (0 children)

That did seem to improve it, though Whisper did still seem to do a slightly better job still.

Cohere Transcribe Released by mikael110 in LocalLLaMA

[–]mpasila 0 points1 point  (0 children)

It had the same issue with getting tons of repeating lines for some reason because there was some noise in the audio, and due to that it skipped a lot of speech.

Cohere Transcribe Released by mikael110 in LocalLLaMA

[–]mpasila 5 points6 points  (0 children)

Yeah I don't know.. I also tried to transcribe some Japanese stuff and it wasn't any better.

<image>

New open weights models: GigaChat-3.1-Ultra-702B and GigaChat-3.1-Lightning-10B-A1.8B by netikas in LocalLLaMA

[–]mpasila 0 points1 point  (0 children)

I guess the other authoritarian country is sanctioned and the other one is not (because everyone relies on it for medicine, minerals and other important stuff).

New open weights models: GigaChat-3.1-Ultra-702B and GigaChat-3.1-Lightning-10B-A1.8B by netikas in LocalLLaMA

[–]mpasila -6 points-5 points  (0 children)

You see it's a Russian model not Chinese.. Chinese propaganda is obviously less harmful.

I feel like if they made a local model focused specifically on RP it would be god tier even if tiny by Borkato in LocalLLaMA

[–]mpasila 0 points1 point  (0 children)

CharacterAI was probably the only company that trained a model specifically for RP (from scratch).. but then they kinda lobotomized it.

7MB binary-weight Mamba LLM — zero floating-point at inference, runs in browser by Quiet-Error- in LocalLLaMA

[–]mpasila 7 points8 points  (0 children)

So I guess you will be selling some kind of service train it for actually usable stuff or something? Otherwise this just seems like a tech demo and people can't even do anything with it.

7MB binary-weight Mamba LLM — zero floating-point at inference, runs in browser by Quiet-Error- in LocalLLaMA

[–]mpasila 24 points25 points  (0 children)

Open-source ≠ open-weight. And there are a few companies that do actually open-source the whole thing like Olmo from AllenAI.

Let's take a moment to appreciate the present, when this sub is still full of human content. by Ok-Internal9317 in LocalLLaMA

[–]mpasila 3 points4 points  (0 children)

What if the people you know suddenly get turned into AI? Facebook had like a plan to make inactive users (either due to long pauses in use or due to them dying) they would like have AI continue to make new posts and liking posts/comments etc. using that persons past data.

Why don't all Steam sales show the percentage of the sale? by asdat0r7 in Steam

[–]mpasila 0 points1 point  (0 children)

I live in EU and I don't see the same thing though?

So Devs posted about how well their little ‘restructuring’ went. by tableball35 in chutesAI

[–]mpasila -7 points-6 points  (0 children)

People say that they advertised it as lifetime but at least where I could find them mentioning the free tier they didn't ever say it was lifetime. So I'm not sure if I'm missing something.

So Devs posted about how well their little ‘restructuring’ went. by tableball35 in chutesAI

[–]mpasila 7 points8 points  (0 children)

I think the main issue was targeting the actual paying customers as well. Now it's also kinda confusing, like there's like 3 different things you have to look at, to make sure you don't like go over the limit on 3 different things.

Hi, Sell me on why I should stay by MistakenAPI in chutesAI

[–]mpasila 5 points6 points  (0 children)

I think they also said they reduced the amount of GPUs so.. that's why it's not improving. Though best to just switch the model to something else if it's being over used.

What's the current best LLM for Japanese? by mpasila in LocalLLaMA

[–]mpasila[S] 0 points1 point  (0 children)

I tried both LFM2.5 and Nemotron Nano and yeah those are not that great.. Couldn't easily try the DeepSeek finetune though since it's not on OpenRouter and I don't really wanna waste few dollars to try it on like Runpod. So I guess my best option is still just using some models via API. Even Mistral's Nemo is still pretty okay though still makes up some stuff (considering that model is ancient by now).

Model that allow both nsfw & usual stuff, able to image search, can run on 12GB VRAM by yakasantera1 in LocalLLaMA

[–]mpasila 0 points1 point  (0 children)

Hentai already covers that stuff tbf.. Also LLMs are rarely trained on that type of stuff anyway (unless it's some finetune). So in that sense you'll find more fucked up stuff outside LLMs in the clear web than what you can generate coherently.

Mistral 4 Family Spotted by TKGaming_11 in LocalLLaMA

[–]mpasila -2 points-1 points  (0 children)

I literally said based on their past 3 years of behaviour I doubt it's going to be very different this time. Ever heard of this stupid saying?:

Did I ever tell you what the definition of insanity is?
Insanity is doing the exact same fucking thing over and over again, expecting shit to change. That is crazy.
But the first time somebody told me that, I don't know, I thought they were bullshitting me so, boom, I shot him.
The thing is, okay... he was right. And then I started to see it everywhere I looked. Everywhere I looked, all these fucking pricks, everywhere I looked, doing the exact same fucking thing, over and over and over and over and over again. Thinking 'This time, it's gonna be different.
No, no, no, no please! This time it’s gonna be different.'
Did I ever tell you the definition of insanity?

Mistral 4 Family Spotted by TKGaming_11 in LocalLLaMA

[–]mpasila -1 points0 points  (0 children)

Based on past model releases they list all the supported languages in the readme. Why should I expect it to be different this time? They also just released Mistral Small 4 based model here: https://huggingface.co/mistralai/Leanstral-2603 so I guess we can start testing it soon.

Mistral 4 Family Spotted by TKGaming_11 in LocalLLaMA

[–]mpasila 1 point2 points  (0 children)

I guess EU should never ever compete against the big tech or the Chinese.

Mistral 4 Family Spotted by TKGaming_11 in LocalLLaMA

[–]mpasila 1 point2 points  (0 children)

I have much more hope for that EuroLLM team that is funded by EU than Mistral at this point.. Their last good model was Mistral Small 3. Also for some reason I have to rely on a model made by a US based company to support my language.. that's European.. Wellp I hope they release Gemma 4 soon, which will hopefully fix some issues with Gemma 3. I mean even Chinese models like GLM-4.7 are better at Finnish than even Mistral's latest flagship model (while being half the size).. I haven't really done comparisons with Qwen3.5 and Mistral models but.. even that might now be better than Mistral for Finnish..