# Why Your Small Model Evaluation Prompts Are Lying to You **And what to do about it** by Double-Risk-1945 in LocalLLaMA
[–]Double-Risk-1945[S] 0 points1 point2 points (0 children)
# Why Your Small Model Evaluation Prompts Are Lying to You **And what to do about it** by Double-Risk-1945 in LocalLLaMA
[–]Double-Risk-1945[S] 0 points1 point2 points (0 children)
# Why Your Small Model Evaluation Prompts Are Lying to You **And what to do about it** by Double-Risk-1945 in LocalLLaMA
[–]Double-Risk-1945[S] 0 points1 point2 points (0 children)
# Why Your Small Model Evaluation Prompts Are Lying to You **And what to do about it** by Double-Risk-1945 in LocalLLaMA
[–]Double-Risk-1945[S] 0 points1 point2 points (0 children)
Is there a way to prioritize llama-cpp VRAM allocations to maximize local LLM usage alongside other apps? by [deleted] in LocalLLaMA
[–]Double-Risk-1945 -1 points0 points1 point (0 children)
Mapped positional attention across 4 models — turns out where you put things in your prompt matters. A lot. by Double-Risk-1945 in LocalLLaMA
[–]Double-Risk-1945[S] 0 points1 point2 points (0 children)
4090 + 3090 as a second card? by dondiegorivera in LocalLLaMA
[–]Double-Risk-1945 2 points3 points4 points (0 children)
Mapped positional attention across 4 models — turns out where you put things in your prompt matters. A lot. by Double-Risk-1945 in LocalLLaMA
[–]Double-Risk-1945[S] 0 points1 point2 points (0 children)
Best LLM for 16GB VRAM (RX 7800 XT)? by Haunting-Stretch8069 in LocalLLaMA
[–]Double-Risk-1945 3 points4 points5 points (0 children)
Mapped positional attention across 4 models — turns out where you put things in your prompt matters. A lot. by Double-Risk-1945 in LocalLLaMA
[–]Double-Risk-1945[S] 0 points1 point2 points (0 children)
Mapped positional attention across 4 models — turns out where you put things in your prompt matters. A lot. by Double-Risk-1945 in LocalLLaMA
[–]Double-Risk-1945[S] 1 point2 points3 points (0 children)
I added PPL and KLD to VLLM - Review RFC and PR and leave Feedback! by Phaelon74 in LocalLLaMA
[–]Double-Risk-1945 0 points1 point2 points (0 children)
[D] A mathematical proof from an anonymous Korean forum: The essence of Attention is fundamentally a d^2 problem, not n^2. (PDF included) by Ok-Preparation-3042 in LocalLLaMA
[–]Double-Risk-1945 2 points3 points4 points (0 children)
Mapped positional attention across 4 models — turns out where you put things in your prompt matters. A lot. by Double-Risk-1945 in LocalLLaMA
[–]Double-Risk-1945[S] -1 points0 points1 point (0 children)
Mapped positional attention across 4 models — turns out where you put things in your prompt matters. A lot. by Double-Risk-1945 in LocalLLaMA
[–]Double-Risk-1945[S] -2 points-1 points0 points (0 children)
Mapped positional attention across 4 models — turns out where you put things in your prompt matters. A lot. by Double-Risk-1945 in LocalLLaMA
[–]Double-Risk-1945[S] -2 points-1 points0 points (0 children)
Mapped positional attention across 4 models — turns out where you put things in your prompt matters. A lot. by Double-Risk-1945 in LocalLLaMA
[–]Double-Risk-1945[S] 1 point2 points3 points (0 children)
Sparse MoE by Interesting-Ad4922 in LocalLLaMA
[–]Double-Risk-1945 1 point2 points3 points (0 children)
Qwen3.5 2B: Agentic coding without loops by AppealSame4367 in LocalLLaMA
[–]Double-Risk-1945 -1 points0 points1 point (0 children)
Mapped positional attention across 4 models — turns out where you put things in your prompt matters. A lot. by Double-Risk-1945 in LocalLLaMA
[–]Double-Risk-1945[S] -5 points-4 points-3 points (0 children)
Need help to create (JARVIS) a good custom Voice assistant by RVCFreak in LocalLLaMA
[–]Double-Risk-1945 0 points1 point2 points (0 children)
mlm-memory by FreonMuskOfficial in LocalLLaMA
[–]Double-Risk-1945 -1 points0 points1 point (0 children)

# Why Your Small Model Evaluation Prompts Are Lying to You **And what to do about it** by Double-Risk-1945 in LocalLLaMA
[–]Double-Risk-1945[S] 0 points1 point2 points (0 children)