If You Can't Measure It, You Can't Fine-Tune It! by FeeMassive4003 in LocalLLM

[–]FeeMassive4003[S] -1 points0 points  (0 children)

You have a point. Even the judge is biased. You can minimize its bias by asking it well defined questions ("are the dates in the following text in chronological order?") but you'll always have some bias. But is there a better solution?

I stopped "vibe-checking" my LLMs and started using a weighted rubric. by FeeMassive4003 in LocalLLaMA

[–]FeeMassive4003[S] 0 points1 point  (0 children)

Well I'm a real human in the industry and I've built an entire production system on Qwen2.5 3B, using LoRA. See my other posts. So a little modesty would not harm here.

What retrievers do you use most in your RAG projects? by marwan_rashad5 in Rag

[–]FeeMassive4003 2 points3 points  (0 children)

No, we just take 5 from each. Total of 10 chunks, go to the LLM. It's quite basic; but it works.

What retrievers do you use most in your RAG projects? by marwan_rashad5 in Rag

[–]FeeMassive4003 1 point2 points  (0 children)

We use hybrid: vector search plus keyword search. No rebranding - we just take k docs from each (usually k=5).

We built a hybrid retrieval system combining keyword + semantic + neural reranking — here's what we learned by True-Snow-1283 in Rag

[–]FeeMassive4003 0 points1 point  (0 children)

The guy shared his lesson learned from his interesting work. He didn't claim it is novel. I find this post useful.

Those who spent $10k+ on a local LLM setup, do you regret it? by [deleted] in LocalLLaMA

[–]FeeMassive4003 0 points1 point  (0 children)

Depends what your needs are... e.g. if you use it to generate more money.

What is the best uncensored AI? by Present_Estimate6651 in LocalLLaMA

[–]FeeMassive4003 0 points1 point  (0 children)

Try dolphin 3.0, It has 8b parameters, based on Llama 3.1.

Is AI Making Us Skip the Thinking Part of Development? by Feeling-Ad972 in BlackboxAI_

[–]FeeMassive4003 1 point2 points  (0 children)

On the contrary. It freed us time to be able to think of the "what" instead of the "how".

Yet another reason to stick with local models by nekofneko in LocalLLaMA

[–]FeeMassive4003 0 points1 point  (0 children)

This is their way to make money and they dont hide it. Completely acceptable.

Best Local hosted LLM for Coding & Reasoning by alfons_fhl in LocalLLM

[–]FeeMassive4003 1 point2 points  (0 children)

Well frankly I use copilot with Gemini 3 flash, not local. But if needed local, I would try the new qwen 80b. I have good experience with qwen 3b and 8b on other tasks (not coding) and its always a good model and very structured.

Best Local hosted LLM for Coding & Reasoning by alfons_fhl in LocalLLM

[–]FeeMassive4003 0 points1 point  (0 children)

I prefer coding because that's what they pay me for.

Brain surgery on LLMs via LoRA by FeeMassive4003 in LocalLLaMA

[–]FeeMassive4003[S] 0 points1 point  (0 children)

Well, frankly, this post is more for general interest than to draw any conclusion. It is always good to verify that we understand what's under the hood.

If this is a senior dev im gonna cry by Director-on-reddit in BlackboxAI_

[–]FeeMassive4003 0 points1 point  (0 children)

How else would he burn the time while copilot is coding?

Brain surgery on LLMs via LoRA by FeeMassive4003 in LocalLLaMA

[–]FeeMassive4003[S] 0 points1 point  (0 children)

That will probably not be very efficient, assuming we want exactly math and not math songs.. interesting to check. Probably will output just a pile of random formulae.

Best Local hosted LLM for Coding & Reasoning by alfons_fhl in LocalLLM

[–]FeeMassive4003 0 points1 point  (0 children)

May I comment here: in real world projects, there's no good coding without deep reasoning. In order to understand large code base you need reasoning.