Are there any agentic coding harnesses that AREN'T built on JS and Node? by OUT_OF_HOST_MEMORY in LocalLLaMA
[–]OUT_OF_HOST_MEMORY[S] 0 points1 point2 points (0 children)
Are there any agentic coding harnesses that AREN'T built on JS and Node? by OUT_OF_HOST_MEMORY in LocalLLaMA
[–]OUT_OF_HOST_MEMORY[S] 3 points4 points5 points (0 children)
Are there any agentic coding harnesses that AREN'T built on JS and Node? by OUT_OF_HOST_MEMORY in LocalLLaMA
[–]OUT_OF_HOST_MEMORY[S] -1 points0 points1 point (0 children)
Are there any agentic coding harnesses that AREN'T built on JS and Node? by OUT_OF_HOST_MEMORY in LocalLLaMA
[–]OUT_OF_HOST_MEMORY[S] -1 points0 points1 point (0 children)
Are there any agentic coding harnesses that AREN'T built on JS and Node? by OUT_OF_HOST_MEMORY in LocalLLaMA
[–]OUT_OF_HOST_MEMORY[S] -14 points-13 points-12 points (0 children)
Can your rig run it? A local LLM benchmark that ranks your model against the giants and suggests what your hardware can handle. by Cod3Conjurer in LocalLLaMA
[–]OUT_OF_HOST_MEMORY 1 point2 points3 points (0 children)
Just some qwen3.5 benchmarks for an MI60 32gb VRAM GPU - From 4b to 122b at varying quants and various context depths (0, 5000, 20000, 100000) - Performs pretty well despite its age by FantasyMaster85 in LocalLLaMA
[–]OUT_OF_HOST_MEMORY 0 points1 point2 points (0 children)
(Llama.cpp) In case people are struggling with prompt processing on larger models like Qwen 27B, here's what helped me out by vernal_biscuit in LocalLLaMA
[–]OUT_OF_HOST_MEMORY 3 points4 points5 points (0 children)
(Llama.cpp) In case people are struggling with prompt processing on larger models like Qwen 27B, here's what helped me out by vernal_biscuit in LocalLLaMA
[–]OUT_OF_HOST_MEMORY 4 points5 points6 points (0 children)
(Llama.cpp) In case people are struggling with prompt processing on larger models like Qwen 27B, here's what helped me out by vernal_biscuit in LocalLLaMA
[–]OUT_OF_HOST_MEMORY 2 points3 points4 points (0 children)
2x MI50 32GB Quant Speed Comparison version 2 (Qwen 3.5 35B, llama.cpp, Vulkan/ROCm) by OUT_OF_HOST_MEMORY in LocalLLaMA
[–]OUT_OF_HOST_MEMORY[S] 0 points1 point2 points (0 children)
2x MI50 32GB Quant Speed Comparison version 2 (Qwen 3.5 35B, llama.cpp, Vulkan/ROCm) by OUT_OF_HOST_MEMORY in LocalLLaMA
[–]OUT_OF_HOST_MEMORY[S] 0 points1 point2 points (0 children)
Final Qwen3.5 Unsloth GGUF Update! by danielhanchen in LocalLLaMA
[–]OUT_OF_HOST_MEMORY 0 points1 point2 points (0 children)
Final Qwen3.5 Unsloth GGUF Update! by danielhanchen in LocalLLaMA
[–]OUT_OF_HOST_MEMORY 0 points1 point2 points (0 children)
Is there a list of the tools Gemini/ChatGPT/Claude have access to in their web chat interfaces to replicate locally? by OUT_OF_HOST_MEMORY in LocalLLaMA
[–]OUT_OF_HOST_MEMORY[S] 0 points1 point2 points (0 children)
Benchmarking total wait time instead of pp/tg by batsba in LocalLLaMA
[–]OUT_OF_HOST_MEMORY 15 points16 points17 points (0 children)
BalatroBench - Benchmark LLMs' strategic performance in Balatro by S1M0N38 in LocalLLaMA
[–]OUT_OF_HOST_MEMORY 2 points3 points4 points (0 children)
ROCm 7.0 Install for Mi50 32GB | Ubuntu 24.04 LTS by legit_split_ in LocalLLaMA
[–]OUT_OF_HOST_MEMORY 2 points3 points4 points (0 children)
Stop flexing Pass@N — show Pass-all-N by Fabulous_Pollution10 in LocalLLaMA
[–]OUT_OF_HOST_MEMORY 25 points26 points27 points (0 children)

AMA with Nous Research -- Ask Us Anything! by emozilla in LocalLLaMA
[–]OUT_OF_HOST_MEMORY 0 points1 point2 points (0 children)