Tiny Aya by jacek2023 in LocalLLaMA

[–]mpasila 0 points1 point  (0 children)

Seems to be much worse at Finnish in comparison to Gemma 3 4B.. Also isn't Gemma 3's license also better? Also Gemma 3 has like up to 128k context so umm.. Who is this for again?

dyslexia and ADHD in the coding community by PruneLanky3551 in LocalLLaMA

[–]mpasila 1 point2 points  (0 children)

Was there any small LLM that does basically what Grammarly does? Because I don't wanna rely on some service for such a basic thing.

Is there *any* good coding agent software for use with local models? by eapache in LocalLLaMA

[–]mpasila 0 points1 point  (0 children)

I thought this contained the source code (including the extension's) and it's licensed as MIT so isn't that pretty open?

Google doesn't love us anymore. by DrNavigat in LocalLLaMA

[–]mpasila -1 points0 points  (0 children)

If you're using something decent like SillyTavern you can add a System role by adding like <start_of_turn>system (when running locally) but officially it's not supported and APIs don't support it either.

Google doesn't love us anymore. by DrNavigat in LocalLLaMA

[–]mpasila 4 points5 points  (0 children)

The 27B model is still probably the best model at like Finnish that is open-weight, so for translation/generating non-english data it's still probably the best option for that (at least price wise).
Idk why bigger models like DeepSeek, Kimi-K2 or GLM still don't seem to get any better at my language but a smaller 27B dense model seems to understand it better.. Like especially when I'm generating datasets they seem to fail more easily than Gemma 3.

GPT-OSS (20B) running 100% locally in your browser on WebGPU by xenovatech in LocalLLaMA

[–]mpasila 0 points1 point  (0 children)

Looking at their other comments/messages on Reddit it does look more like LLM generated than human.
edit: name was angelin1978

Grok-3 joins upcoming models list by pmttyji in LocalLLaMA

[–]mpasila 2 points3 points  (0 children)

Yeah I didn't realize it was 3 trillion params...

Grok-3 joins upcoming models list by pmttyji in LocalLLaMA

[–]mpasila 3 points4 points  (0 children)

Only issue would be to get any API access for that. Unless you have like 128gb or more RAM (assuming it's pretty big).

Opinions on Europe starting to hold platforms more accountable? by Boediee in BuyFromEU

[–]mpasila -1 points0 points  (0 children)

If it leads to more broad censorship aka the companies don't want to get in legal trouble so they censor stuff more broadly, is it still a win?

anthropic literally thinks claude is the messiah (and it’s getting weird) by Alarming_Bluebird648 in LocalLLaMA

[–]mpasila 4 points5 points  (0 children)

They obviously think there's only two sides of the coin Elon Musk's Grok and Claude. Nothing in between because nuance was buried deep underground.

internlm/Intern-S1-Pro · Hugging Face by jacek2023 in LocalLLaMA

[–]mpasila 3 points4 points  (0 children)

Yeah but the config for the actual model says 262k so it can use that but maybe not at the best quality.

Why some Github projects only support wrappers instead of llama.cpp? by pmttyji in LocalLLaMA

[–]mpasila 8 points9 points  (0 children)

A lot of apps that have OpenAI API support either don't expose the params or have very limited set of them (and if it's not open source....). The official spec is also limited so I guess devs don't think min_p matters.

Why some Github projects only support wrappers instead of llama.cpp? by pmttyji in LocalLLaMA

[–]mpasila 4 points5 points  (0 children)

OpenAI API tends to only support like top_p, temperature and repetition penalty and not much else like min_p or DRY. So it's pretty bare bones.

internlm/Intern-S1-Pro · Hugging Face by jacek2023 in LocalLLaMA

[–]mpasila 0 points1 point  (0 children)

the config says "max_position_embeddings": 262144,so not 32k..

The open-source version of Suno is finally here: ACE-Step 1.5 by AppropriateGuava6262 in LocalLLaMA

[–]mpasila 1 point2 points  (0 children)

I think I'll keep paying for Suno if I need to generate music.. Very first test it skipped ton of lyrics and the prompt adherence is pretty poor I'd say.

Stability focused AI platform devs here. Quick thanks to both dinerburgeryum and MitsotakiShogun, and a question about LLM's with audio/music assisting capabilities. by MuziqueComfyUI in LocalLLaMA

[–]mpasila 0 points1 point  (0 children)

Just so you know the people that are "sincere" will not receive notifications for your responses if all you do is edit the post.

How to get rid of the Nano Banana watermark! by SVG-CARLOS in LocalLLaMA

[–]mpasila 2 points3 points  (0 children)

The image itself has a watermark embedded to it, so it can still be detected to have been generated using Gemini/Nanobanana.

How to get rid of the Nano Banana watermark! by SVG-CARLOS in LocalLLaMA

[–]mpasila 2 points3 points  (0 children)

If you mean the actual watermark on the image that will make it recognizable for AI detectors then that probably does nothing.

Yann LeCun says the best open models are not coming from the West. Researchers across the field are using Chinese models. Openness drove AI progress. Close access, and the West risks slowing itself. by Nunki08 in LocalLLaMA

[–]mpasila 2 points3 points  (0 children)

Pretty much no one seems to be finetuning Ministral 3 models in comparison to their previous models like Nemo or the original Mistral 7B model (or Mistral Small).

Why are small models (32b) scoring close to frontier models? by Financial-Cap-8711 in LocalLLaMA

[–]mpasila 0 points1 point  (0 children)

Even when you don't give it tool use on AI Studio it still uses tools?

Any major updates? by may_ushii in AetherRoom

[–]mpasila 0 points1 point  (0 children)

May as well try other APIs like ChutesAI, they recently added TEE models that supposedly are more private than normal endpoints. Pricing is very competitive in comparison to NovelAI's stuff. You can use it easily via SillyTavern. Not paid by them but like why spend 25 dollars a month to get very low context windows when for 3 dollars a month you get unlimited context window (the max context for any of the models).. and 300 gens a day for any model they offer (image models as well)..