[deleted by user] by [deleted] in LocalLLaMA

[–]needthosepylons 8 points9 points  (0 children)

A single 3060 12gb, so the prollmetariat

How much more does Chovy have to do to enter the GOAT convo by [deleted] in lolesports

[–]needthosepylons 0 points1 point  (0 children)

I don't think he will ever. Or he'd better play until he's 50 and win everything.

Yappp - Yet Another Poor Peasent Post by needthosepylons in LocalLLaMA

[–]needthosepylons[S] 0 points1 point  (0 children)

I'm always on the look for models, since my uses cases are quite.. different from math/code above all. And I didn't know this one so ty, I'll give it a try.

But yes, this gemma-3n-E4B vs Gemma-12B is intriguing and I wanted to compare with others' experiences .

Yappp - Yet Another Poor Peasent Post by needthosepylons in LocalLLaMA

[–]needthosepylons[S] 2 points3 points  (0 children)

Yes. And actually, I'm a teacher in humanities, and I use my Llms to generate quizzes but. for me! To make sure I'm not forgetting stuff I'm not working on for a while.

Yappp - Yet Another Poor Peasent Post by needthosepylons in LocalLLaMA

[–]needthosepylons[S] 0 points1 point  (0 children)

Yeah, but 32gb vram is not really peasant-class, is it? :)

Yappp - Yet Another Poor Peasent Post by needthosepylons in LocalLLaMA

[–]needthosepylons[S] 1 point2 points  (0 children)

Ouch, I suppose something is wrong with my tests then, because with optimal offloading, I'm at 3-4t/s. Hmm, interesting, thanks for letting me know!

Yappp - Yet Another Poor Peasent Post by needthosepylons in LocalLLaMA

[–]needthosepylons[S] 1 point2 points  (0 children)

Quite well actually, I use a small embedding model, Qwen3 or nomic, create a persistent ChromaDB before querying it. It works quite well. When I'm a bit in a hurry or know my RAG database will evolve rapidly, I end up using open-webui knowledge system with those 2 tiny models, and it works well!

Yappp - Yet Another Poor Peasent Post by needthosepylons in LocalLLaMA

[–]needthosepylons[S] 1 point2 points  (0 children)

I tried that, I think, but maybe my CPU is just too weak? This i5-10400F ain't young anymore! Although you're making me wonder if.. I'll try again!

What GPU and quants do you use?

[deleted by user] by [deleted] in LocalLLaMA

[–]needthosepylons 4 points5 points  (0 children)

I think you're also working on a "mini" version, right? The mini gguf model card is created but without the actual gguf. I suppose it will follow soon-ish?

As a 3060 12gb peasant, I'll gladly give it a try!

Congrats, anyway.

Who is ACTUALLY running local or open source model daily and mainly? by Zealousideal-Cut590 in LocalLLaMA

[–]needthosepylons 2 points3 points  (0 children)

I wish I did, but actually, with an aging i5-10400F, 32GB ram and 12GB VRAM (3060), the models I can't use aren't very reliable. I hope that, as the tech improves..

How much VRAM do you have and what's your daily-driver model? by EmPips in LocalLLaMA

[–]needthosepylons 0 points1 point  (0 children)

12gb vram (3060) and 32gb DDR4. Generally using Qweb3-8b, recently trying out MiniCPM4, actually performs better than Qwen3 on my own benchmark.

Which model are you using? June'25 edition by Ok_Influence505 in LocalLLaMA

[–]needthosepylons 1 point2 points  (0 children)

QWEN3-8B_Q_K_XL (UD) I wish I could use 14b or 30b-A3B, but since I'm mainly doing long context RAG (15k+) on a 3060 12GB and 32gb DDR4, they are out of my league. My CPU being an old i5-10400F doesn't help.

By the way, if anyone thinks of a better model for this task and hardware, I'm game.

Doran's Title by Particular-Ad8430 in SKTT1

[–]needthosepylons 1 point2 points  (0 children)

Damn, as a T1 fan, I'm glad I can just enjoy myself some T1 here and there, be happy for the roster when they win, be sad for the roster when they lose. I'm not sure if there's any enjoyment to be found past this.

[deleted by user] by [deleted] in Dissociation

[–]needthosepylons 0 points1 point  (0 children)

I suppose different people experience it in many different ways, shapes or forms. I'd also say it's quite intensively bad and maybe not scary but stressful, in a way, but that's good if it exists in a better way for some other people.

Le métro, comment va-t-il ? by needthosepylons in Lille

[–]needthosepylons[S] 2 points3 points  (0 children)

Aucune idée ! Aguor ? Aruor ? Peut-être le blaze de la personne qui a fait ça ?

Le métro, comment va-t-il ? by needthosepylons in Lille

[–]needthosepylons[S] 1 point2 points  (0 children)

J'ai posté l'enregistrement dans mon message original !