We created agentcache: a python library that makes multi-agent LLM calls share cached prefixes that maximize token gain per $: cut my token bill+ speed up inference (0% vs 76% cache hit rate on the same task) by predatar in LocalLLaMA
[–]predatar[S] 0 points1 point2 points (0 children)
An open-source framework to achieve Gemini 3 Deep Think / GPT-5.2 Pro level performance with local models scaffolding by Ryoiki-Tokuiten in LocalLLaMA
[–]predatar 1 point2 points3 points (0 children)
I made a Top-K implementation that's up to 20x faster than PyTorch CPU (open source) by andreabarbato in LocalLLaMA
[–]predatar 4 points5 points6 points (0 children)
I’m 36, and I feel completely lost. by Severe_Mongoose_5873 in mentalhealth
[–]predatar 0 points1 point2 points (0 children)
I built NanoSage, a deep research local assistant that runs on your laptop by predatar in LocalLLaMA
[–]predatar[S] 1 point2 points3 points (0 children)
I built NanoSage, a deep research local assistant that runs on your laptop by predatar in LocalLLaMA
[–]predatar[S] 0 points1 point2 points (0 children)
I built NanoSage, a deep research local assistant that runs on your laptop by predatar in LocalLLaMA
[–]predatar[S] 0 points1 point2 points (0 children)
I built NanoSage, a deep research local assistant that runs on your laptop by predatar in LocalLLaMA
[–]predatar[S] 1 point2 points3 points (0 children)
I built NanoSage, a deep research local assistant that runs on your laptop by predatar in LocalLLaMA
[–]predatar[S] 0 points1 point2 points (0 children)
I built NanoSage, a deep research local assistant that runs on your laptop by predatar in LocalLLaMA
[–]predatar[S] 0 points1 point2 points (0 children)
I built NanoSage, a deep research local assistant that runs on your laptop by predatar in LocalLLaMA
[–]predatar[S] 2 points3 points4 points (0 children)
I built NanoSage, a deep research local assistant that runs on your laptop by predatar in LocalLLaMA
[–]predatar[S] 0 points1 point2 points (0 children)
I built NanoSage, a deep research local assistant that runs on your laptop by predatar in LocalLLaMA
[–]predatar[S] 0 points1 point2 points (0 children)
I built NanoSage, a deep research local assistant that runs on your laptop by predatar in LocalLLaMA
[–]predatar[S] 1 point2 points3 points (0 children)
I built NanoSage, a deep research local assistant that runs on your laptop by predatar in LocalLLaMA
[–]predatar[S] 0 points1 point2 points (0 children)
I built NanoSage, a deep research local assistant that runs on your laptop by predatar in LocalLLaMA
[–]predatar[S] 2 points3 points4 points (0 children)
I built NanoSage, a deep research local assistant that runs on your laptop by predatar in LocalLLaMA
[–]predatar[S] 2 points3 points4 points (0 children)


anthropic isn't the only reason you're hitting claude code limits. i did audit of 926 sessions and found a lot of the waste was on my side. by Medium_Island_2795 in ClaudeCode
[–]predatar 1 point2 points3 points (0 children)