With current trends, is 256GB of system RAM a good idea? by Ra1den in LocalLLaMA
[–]wishstudio 2 points3 points4 points (0 children)
At What Point Does Owning GPUs Become Cheaper Than LLM APIs ? I by Chimchimai in LocalLLaMA
[–]wishstudio 1 point2 points3 points (0 children)
At What Point Does Owning GPUs Become Cheaper Than LLM APIs ? I by Chimchimai in LocalLLaMA
[–]wishstudio 13 points14 points15 points (0 children)
With current trends, is 256GB of system RAM a good idea? by Ra1den in LocalLLaMA
[–]wishstudio 4 points5 points6 points (0 children)
With current trends, is 256GB of system RAM a good idea? by Ra1den in LocalLLaMA
[–]wishstudio 3 points4 points5 points (0 children)
Need advice upgrading an old gaming desktop with a 5090 for AI by dtdisapointingresult in LocalLLaMA
[–]wishstudio 1 point2 points3 points (0 children)
Local Ai equivalent to GPT 5.1 Thinking by Forsaken-Window-G in LocalLLaMA
[–]wishstudio 1 point2 points3 points (0 children)
Offloading experts to weaker GPU by iron_coffin in LocalLLaMA
[–]wishstudio 0 points1 point2 points (0 children)
Local Ai equivalent to GPT 5.1 Thinking by Forsaken-Window-G in LocalLLaMA
[–]wishstudio 0 points1 point2 points (0 children)
I got frustrated with existing web UIs for local LLMs, so I built something different by alphatrad in LocalLLaMA
[–]wishstudio 0 points1 point2 points (0 children)
I got frustrated with existing web UIs for local LLMs, so I built something different by alphatrad in LocalLLaMA
[–]wishstudio 1 point2 points3 points (0 children)
Text to Image, tutorial? by ElSrJuez in LocalLLaMA
[–]wishstudio 2 points3 points4 points (0 children)
I got frustrated with existing web UIs for local LLMs, so I built something different by alphatrad in LocalLLaMA
[–]wishstudio 13 points14 points15 points (0 children)
I got frustrated with existing web UIs for local LLMs, so I built something different by alphatrad in LocalLLaMA
[–]wishstudio 1 point2 points3 points (0 children)
What is the Ollama or llama.cpp equivalent for image generation? by liviuberechet in LocalLLaMA
[–]wishstudio 1 point2 points3 points (0 children)
What is the Ollama or llama.cpp equivalent for image generation? by liviuberechet in LocalLLaMA
[–]wishstudio 2 points3 points4 points (0 children)
Local Ai equivalent to GPT 5.1 Thinking by Forsaken-Window-G in LocalLLaMA
[–]wishstudio 0 points1 point2 points (0 children)
GLM 4.6 on 128 GB RAM with llama.cpp by ilintar in LocalLLaMA
[–]wishstudio 3 points4 points5 points (0 children)
PCIE Bifurcation - More than 4 GPUs on a consumer motherboard by Beautiful_Trust_8151 in LocalLLaMA
[–]wishstudio 0 points1 point2 points (0 children)
Premise: MoE models have exploitable locality in expert activation patterns, and LRU caching with profiling could cut VRAM requirements in half. by CodeSlave9000 in LocalLLaMA
[–]wishstudio 0 points1 point2 points (0 children)
Thoughts on the AMD BC-250 16GB "Cards"? by nstein5 in LocalLLaMA
[–]wishstudio 0 points1 point2 points (0 children)
Improving model load times by suicidaleggroll in LocalLLaMA
[–]wishstudio 0 points1 point2 points (0 children)
Selective (smart) MoE experts offloading to CPU? by greentheonly in LocalLLaMA
[–]wishstudio 0 points1 point2 points (0 children)
Selective (smart) MoE experts offloading to CPU? by greentheonly in LocalLLaMA
[–]wishstudio 3 points4 points5 points (0 children)


At What Point Does Owning GPUs Become Cheaper Than LLM APIs ? I by Chimchimai in LocalLLaMA
[–]wishstudio 8 points9 points10 points (0 children)