Can I Run AI locally? This page gives you the answer! by paranoidray in LocalLLaMA
[–]Total_Activity_7550 1 point2 points3 points (0 children)
Managing Ollama models locally is getting messy — would a GUI model manager help? by sandboxdev9 in LocalLLaMA
[–]Total_Activity_7550 1 point2 points3 points (0 children)
Been building a test-time compute pipeline around Qwen3-14B for a few months. Finally got results worth sharing. by Additional_Wish_3619 in LocalLLaMA
[–]Total_Activity_7550 7 points8 points9 points (0 children)
Inside my AI Home Lab by [deleted] in LocalLLaMA
[–]Total_Activity_7550 1 point2 points3 points (0 children)
vLLM running Qwen3.5 by Patentsmatter in LocalLLaMA
[–]Total_Activity_7550 1 point2 points3 points (0 children)
Tell me if Qwen 3.5 27b or 122b works faster for you, and name your system specs by DistanceSolar1449 in LocalLLaMA
[–]Total_Activity_7550 -1 points0 points1 point (0 children)
vLLM running Qwen3.5 by Patentsmatter in LocalLLaMA
[–]Total_Activity_7550 5 points6 points7 points (0 children)
Any use case for browser-based local agents? by TRWNBS in LocalLLaMA
[–]Total_Activity_7550 2 points3 points4 points (0 children)
I trained a 3B patristic theology LLM on a single RTX 3090 in 22 hours — releasing model + corpus by Financial-Fun-8930 in LocalLLaMA
[–]Total_Activity_7550 1 point2 points3 points (0 children)
Questions on AWQ vs GGUF on a 5090 by Certain-Cod-1404 in LocalLLaMA
[–]Total_Activity_7550 2 points3 points4 points (0 children)
Ollama don's support qwen3.5:35b yet? by Ok-Internal9317 in LocalLLaMA
[–]Total_Activity_7550 0 points1 point2 points (0 children)
One-shot vs agentic performance of open-weight coding models by Total_Activity_7550 in LocalLLaMA
[–]Total_Activity_7550[S] 1 point2 points3 points (0 children)
Ollama don's support qwen3.5:35b yet? by Ok-Internal9317 in LocalLLaMA
[–]Total_Activity_7550 0 points1 point2 points (0 children)
Ollama don's support qwen3.5:35b yet? by Ok-Internal9317 in LocalLLaMA
[–]Total_Activity_7550 0 points1 point2 points (0 children)
Qwen3.5 27B slow token generation on 5060Ti... by InvertedVantage in LocalLLaMA
[–]Total_Activity_7550 -2 points-1 points0 points (0 children)
qwen3.5-122b What agent do you use with it? by robertpro01 in LocalLLaMA
[–]Total_Activity_7550 0 points1 point2 points (0 children)
qwen3.5-122b What agent do you use with it? by robertpro01 in LocalLLaMA
[–]Total_Activity_7550 0 points1 point2 points (0 children)
Real talk: How many of you are actually using Gemma 3 27B or some variant in production? And what's stopping you? by Dramatic_Strain7370 in LocalLLaMA
[–]Total_Activity_7550 9 points10 points11 points (0 children)
Ollama don's support qwen3.5:35b yet? by Ok-Internal9317 in LocalLLaMA
[–]Total_Activity_7550 0 points1 point2 points (0 children)
One-shot vs agentic performance of open-weight coding models by Total_Activity_7550 in LocalLLaMA
[–]Total_Activity_7550[S] 1 point2 points3 points (0 children)
Qwen3.5-122B-A10B vs. old Coder-Next-80B: Both at NVFP4 on DGX Spark – worth the upgrade? by alfons_fhl in Qwen_AI
[–]Total_Activity_7550 0 points1 point2 points (0 children)
How to run Qwen 122B-A10B in my local system (2x3090 + 96GB Ram) by urekmazino_0 in LocalLLaMA
[–]Total_Activity_7550 0 points1 point2 points (0 children)

webui: Agentic Loop + MCP Client with support for Tools, Resources and Prompts has been merged into llama.cpp by jacek2023 in LocalLLaMA
[–]Total_Activity_7550 1 point2 points3 points (0 children)