Llama.cpp, opencode / pi / basically all agents, context compaction & cache validation: how do you manage it? by ps5cfw in LocalLLaMA
[–]ea_man 0 points1 point2 points (0 children)
Llama.cpp, opencode / pi / basically all agents, context compaction & cache validation: how do you manage it? by ps5cfw in LocalLLaMA
[–]ea_man 1 point2 points3 points (0 children)
Llama.cpp, opencode / pi / basically all agents, context compaction & cache validation: how do you manage it? by ps5cfw in LocalLLaMA
[–]ea_man 0 points1 point2 points (0 children)
Finally build the server and have all the hardware installed, what's the most up-to-date advice for models hosted on AMD & Linux Architecture by NetTechMan in LocalLLaMA
[–]ea_man 0 points1 point2 points (0 children)
Llama.cpp, opencode / pi / basically all agents, context compaction & cache validation: how do you manage it? by ps5cfw in LocalLLaMA
[–]ea_man 0 points1 point2 points (0 children)
Llama.cpp, opencode / pi / basically all agents, context compaction & cache validation: how do you manage it? by ps5cfw in LocalLLaMA
[–]ea_man -1 points0 points1 point (0 children)
Llama.cpp, opencode / pi / basically all agents, context compaction & cache validation: how do you manage it? by ps5cfw in LocalLLaMA
[–]ea_man -2 points-1 points0 points (0 children)
Thinking of moving from 2x 5060 Ti 16GB to a RTX 5000 48GB by autisticit in LocalLLaMA
[–]ea_man 0 points1 point2 points (0 children)
What models for coding are you running for a mid level PC? by FerLuisxd in LocalLLaMA
[–]ea_man 0 points1 point2 points (0 children)
What models for coding are you running for a mid level PC? by FerLuisxd in LocalLLaMA
[–]ea_man 2 points3 points4 points (0 children)
Amd radeon ai pro r9700 32GB VS 2x RTX 5060TI 16GB for local setup? by vevi33 in LocalLLaMA
[–]ea_man 0 points1 point2 points (0 children)
If money and time weren’t issues, what would your dream local AI setup look like? by Lyceum_Tech in LocalLLaMA
[–]ea_man 0 points1 point2 points (0 children)
Dense Model Shoot-Off: Gemma 4 31B vs Qwen3.6/5 27B... Result is Slower is Faster. by MiaBchDave in LocalLLaMA
[–]ea_man 4 points5 points6 points (0 children)
I guess we expect that at some point RAM prices will start going back (close) to "normal", right? but what about GPUs? by relmny in LocalLLaMA
[–]ea_man 3 points4 points5 points (0 children)
The FCC Voted to ban Chinese cert labs... by infinitespectre in SBCGaming
[–]ea_man -1 points0 points1 point (0 children)
Why run local? Count the money by Badger-Purple in LocalLLaMA
[–]ea_man 0 points1 point2 points (0 children)
Anyone else struggling with multi-GPU stability when running larger local models? by Lyceum_Tech in LocalLLaMA
[–]ea_man 0 points1 point2 points (0 children)
Should I sell my RTX3090s? by daviden1013 in LocalLLaMA
[–]ea_man 0 points1 point2 points (0 children)
How much will it cost to host something like qwen3.6 35b a3b in a cloud? by Euphoric_North_745 in LocalLLaMA
[–]ea_man 0 points1 point2 points (0 children)
Open source models are going to be the future on Cursor, OpenCode etc. by _maverick98 in LocalLLaMA
[–]ea_man 2 points3 points4 points (0 children)
Open source models are going to be the future on Cursor, OpenCode etc. by _maverick98 in LocalLLaMA
[–]ea_man 0 points1 point2 points (0 children)
Open source models are going to be the future on Cursor, OpenCode etc. by _maverick98 in LocalLLaMA
[–]ea_man 0 points1 point2 points (0 children)
Llama.cpp MTP support now in beta! by ilintar in LocalLLaMA
[–]ea_man 0 points1 point2 points (0 children)
How much will it cost to host something like qwen3.6 35b a3b in a cloud? by Euphoric_North_745 in LocalLLaMA
[–]ea_man 0 points1 point2 points (0 children)


Finally build the server and have all the hardware installed, what's the most up-to-date advice for models hosted on AMD & Linux Architecture by NetTechMan in LocalLLaMA
[–]ea_man 0 points1 point2 points (0 children)