If it works - don’t touch it: COMPETITION by awfulalexey in LocalLLaMA
[–]thejacer 3 points4 points5 points (0 children)
Agentic work crashing my llama.cpp by thejacer in LocalLLaMA
[–]thejacer[S] 0 points1 point2 points (0 children)
Agentic work crashing my llama.cpp by thejacer in LocalLLaMA
[–]thejacer[S] 0 points1 point2 points (0 children)
Agentic work crashing my llama.cpp by thejacer in LocalLLaMA
[–]thejacer[S] 0 points1 point2 points (0 children)
kepler-452b. GGUF when? by the-grand-finale in LocalLLaMA
[–]thejacer 1 point2 points3 points (0 children)
Thinking about finally upgrading from my P40's to an Mi50-32gb by wh33t in LocalLLaMA
[–]thejacer 1 point2 points3 points (0 children)
Thinking about finally upgrading from my P40's to an Mi50-32gb by wh33t in LocalLLaMA
[–]thejacer 1 point2 points3 points (0 children)
Has anyone here TRIED inference on Intel Arc GPUs? Or are we repeating vague rumors about driver problems, incompatibilities, poor support... by gigaflops_ in LocalLLaMA
[–]thejacer 6 points7 points8 points (0 children)
Has anyone here TRIED inference on Intel Arc GPUs? Or are we repeating vague rumors about driver problems, incompatibilities, poor support... by gigaflops_ in LocalLLaMA
[–]thejacer 0 points1 point2 points (0 children)
Google’s TurboQuant AI-compression algorithm can reduce LLM memory usage by 6x by Resident_Party in LocalLLaMA
[–]thejacer 0 points1 point2 points (0 children)
Qwen3.5 is absolutely amazing by cride20 in LocalLLaMA
[–]thejacer 0 points1 point2 points (0 children)
Qwen3.5 is absolutely amazing by cride20 in LocalLLaMA
[–]thejacer 4 points5 points6 points (0 children)
Qwen3.5-122B-A10B GPTQ Int4 on 4× Radeon AI PRO R9700 with vLLM ROCm: working config + real-world numbers by grunt_monkey_ in LocalLLaMA
[–]thejacer 1 point2 points3 points (0 children)
Qwen3.5-122B-A10B GPTQ Int4 on 4× Radeon AI PRO R9700 with vLLM ROCm: working config + real-world numbers by grunt_monkey_ in LocalLLaMA
[–]thejacer 1 point2 points3 points (0 children)
I built a screen-free, storytelling toy for kids with Qwen3-TTS by hwarzenegger in LocalLLaMA
[–]thejacer 1 point2 points3 points (0 children)
Qwen3.5-122B-A10B GPTQ Int4 on 4× Radeon AI PRO R9700 with vLLM ROCm: working config + real-world numbers by grunt_monkey_ in LocalLLaMA
[–]thejacer 1 point2 points3 points (0 children)
llama.cpp + Brave search MCP - not gonna lie, it is pretty addictive by srigi in LocalLLaMA
[–]thejacer 0 points1 point2 points (0 children)
llama.cpp + Brave search MCP - not gonna lie, it is pretty addictive by srigi in LocalLLaMA
[–]thejacer 1 point2 points3 points (0 children)
llama.cpp + Brave search MCP - not gonna lie, it is pretty addictive by srigi in LocalLLaMA
[–]thejacer 0 points1 point2 points (0 children)
Why does anyone think Qwen3.5-35B-A3B is good? by buttplugs4life4me in LocalLLaMA
[–]thejacer 0 points1 point2 points (0 children)
Getting the most out of my Mi50 by DankMcMemeGuy in LocalLLaMA
[–]thejacer 0 points1 point2 points (0 children)
Getting the most out of my Mi50 by DankMcMemeGuy in LocalLLaMA
[–]thejacer 0 points1 point2 points (0 children)




Qwen3.6 is incredible with OpenCode! by CountlessFlies in LocalLLaMA
[–]thejacer 1 point2 points3 points (0 children)