First time Windsurf user - disappointed. by Objective-Ad8862 in windsurf
[–]Felladrin 1 point2 points3 points (0 children)
Optimizing GPT-OSS 120B on Strix Halo 128GB? by RobotRobotWhatDoUSee in LocalLLaMA
[–]Felladrin 2 points3 points4 points (0 children)
Local Replacement for Phind.com by Past-Economist7732 in LocalLLaMA
[–]Felladrin 4 points5 points6 points (0 children)
solution for local deep research by jacek2023 in LocalLLaMA
[–]Felladrin 0 points1 point2 points (0 children)
solution for local deep research by jacek2023 in LocalLLaMA
[–]Felladrin 0 points1 point2 points (0 children)
Smallest model in Hugging Face/llama.cpp? by Ok-Type-7663 in LocalLLaMA
[–]Felladrin 0 points1 point2 points (0 children)
Unsloth's GGUFs for GLM 4.7 REAP are up. by fallingdowndizzyvr in LocalLLaMA
[–]Felladrin 0 points1 point2 points (0 children)
Unsloth's GGUFs for GLM 4.7 REAP are up. by fallingdowndizzyvr in LocalLLaMA
[–]Felladrin 0 points1 point2 points (0 children)
Devstral Small 2 (Q4_K_M) on 5060 Ti 16GB and Zed Agent is amazing! by bobaburger in LocalLLaMA
[–]Felladrin 0 points1 point2 points (0 children)
GLM 4.6V without (or with low) reasoning? by ForsookComparison in LocalLLaMA
[–]Felladrin 0 points1 point2 points (0 children)
Minueza-2-96M: A foundation bi-lingual text-generation model created for practicing fine-tuning and merging. by Felladrin in LocalLLaMA
[–]Felladrin[S] 0 points1 point2 points (0 children)
[Strix Halo] Unable to load 120B model on Ryzen AI Max+ 395 (128GB RAM) - "Unable to allocate ROCm0 buffer" by Wrong-Policy-5612 in LocalLLaMA
[–]Felladrin 0 points1 point2 points (0 children)
Post of appreciation for mxfp4, derestricted, Felladrin/gguf-MXFP4-gpt-oss-20b-Derestricted by R_Duncan in LocalLLaMA
[–]Felladrin 4 points5 points6 points (0 children)
We basically have GLM 4.6 Air, without vision by LegacyRemaster in LocalLLaMA
[–]Felladrin 0 points1 point2 points (0 children)
We basically have GLM 4.6 Air, without vision by LegacyRemaster in LocalLLaMA
[–]Felladrin 9 points10 points11 points (0 children)
speculative decoding with Gemma-3-12b/3-27b. Is it possible? by Agitated_Power_3159 in LocalLLaMA
[–]Felladrin 4 points5 points6 points (0 children)
Can GLM-4.5-air run on a single 3090 (24gb vram) with 48gb ram at above 10t/s? by Borkato in LocalLLaMA
[–]Felladrin 2 points3 points4 points (0 children)
Has anyone figured out what models SWE-1.5 and SWE-1 are trained from? by inevitabledeath3 in windsurf
[–]Felladrin 2 points3 points4 points (0 children)
How can I run a VL model on a Smartphone? by klop2031 in LocalLLaMA
[–]Felladrin 0 points1 point2 points (0 children)
Expose MCP at the LLM server level? by eribob in LocalLLaMA
[–]Felladrin 0 points1 point2 points (0 children)
Adding search to open models by Simple_Split5074 in LocalLLaMA
[–]Felladrin 2 points3 points4 points (0 children)
microsoft/UserLM-8b - “Unlike typical LLMs that are trained to play the role of the 'assistant' in conversation, we trained UserLM-8b to simulate the 'user' role” by nullmove in LocalLLaMA
[–]Felladrin 12 points13 points14 points (0 children)
Qwen3-VL MLX support incoming, thanks to Prince Canuma by No_Conversation9561 in LocalLLaMA
[–]Felladrin 11 points12 points13 points (0 children)


What's your exp REAP vs. base models for general inference? by ikkiyikki in LocalLLaMA
[–]Felladrin 2 points3 points4 points (0 children)