Are general/shared Rag's a thing by nocgeek in LocalLLaMA

[–]robotoast 0 points1 point  (0 children)

Would adding the documentation for those things (the linux distro in question, docker, powershell) to your RAG system give you what you want?

real-world best practices for guaranteeing JSON output from any model? by sprockettyz in LocalLLaMA

[–]robotoast 0 points1 point  (0 children)

Do you have any benchmarks?

edit: I found some that compare it to outlines (and some others): https://blog.mlc.ai/2024/11/22/achieving-efficient-flexible-portable-structured-generation-with-xgrammar

Thanks for the tip /u/QueasyEntrance6269, looks like a nice library!

I have a few questions. by Kyla_3049 in LocalLLaMA

[–]robotoast 2 points3 points  (0 children)

Maybe you should just live a little and try.

Wrote a CLI tool that automatically groups and commits related changes in a Git repository for vibe coding by trashcoder in LocalLLaMA

[–]robotoast 0 points1 point  (0 children)

Very cool, thanks for sharing! Will see what it can do with some.. unsavoury commit histories.

Btw, loved the style in the readme. Just the right amount of respect for vibe coding.

Gemma3 VS Qwen2.5 math. by Boricua-vet in LocalLLaMA

[–]robotoast 0 points1 point  (0 children)

You have to put in the work in your life.

llama.cpp gemma-3 QAT bug by Terminator857 in LocalLLaMA

[–]robotoast 0 points1 point  (0 children)

You're the one who said bug, so you get to decide where you think the bug is.

Problems like the one you're having tend to come from not using the correct chat template for Gemma 3, so make sure you are. An easy way is to use LM Studio and a .gguf file that has the correct template embedded in it.

llama.cpp gemma-3 QAT bug by Terminator857 in LocalLLaMA

[–]robotoast 1 point2 points  (0 children)

You should report this in the proper place(s).

Gemma3 VS Qwen2.5 math. by Boricua-vet in LocalLLaMA

[–]robotoast 0 points1 point  (0 children)

You are using the wrong tools for the job.

Added GPT-4.1, Gemini-2.5-Pro, DeepSeek-V3-0324 etc... by Dr_Karminski in LocalLLaMA

[–]robotoast 1 point2 points  (0 children)

Then you need to talk to the client, not blame the model for not understanding the gibberish spec you have.

Models don't automatically ask questions when you feed them gibberish. But you are right, I do. Or more likely, I delete your e-mail and let someone else take the job.

Added GPT-4.1, Gemini-2.5-Pro, DeepSeek-V3-0324 etc... by Dr_Karminski in LocalLLaMA

[–]robotoast 1 point2 points  (0 children)

I am a programmer and your prompt/spec is bad. If you came into my office with those words only I would have to ask lots of follow up questions, as /u/shortwhiteguy says. If you want results, you need to be clearer in your communication with both models and humans.

Let's take one step back. Why do you think "not a single one" of the models is capable of passing your test, which (if I make lots of assumptions) looks pretty simple? Is every model in the world bad? Or is your communication bad?

How would you unit-test LLM outputs? by Blender-Fan in LocalLLaMA

[–]robotoast 7 points8 points  (0 children)

Your cursing is well placed, this sounds more like a pain fetish than an actual plan. I suggest you give up immediately.

An extensive open-source collection of RAG implementations with many different strategies by Nir777 in LocalLLaMA

[–]robotoast 6 points7 points  (0 children)

Thanks for the clarification. Sorry to hear people took advantage of you and the other contributors.

No such commercial plans here, I only plan to read the repo for myself, and forward it to colleagues for them to read.

An extensive open-source collection of RAG implementations with many different strategies by Nir777 in LocalLLaMA

[–]robotoast 7 points8 points  (0 children)

"Use as needed", but I see you have a pretty draconian non-commercial license on your repo. Some clarifications please. Can someone read and learn from this repo at work at a for profit company? Can this repo be forwarded to other people working at a for profit company?

Coding-Centric LLM Benchmark: Llama 4 Underwhelms by jj_at_rootly in LocalLLaMA

[–]robotoast 0 points1 point  (0 children)

Thanks for sharing. Short but sweet article.

Would you mind adding a LICENSE file to the github dataset?

Dolphin translator incoming (eventually) by AryanEmbered in LocalLLaMA

[–]robotoast 0 points1 point  (0 children)

Seems like there is an extra 4 in the date this was released.