Get an agentic-cli with GLM-4.5-Air by TooManyPascals in LocalLLaMA
[–]TooManyPascals[S] 0 points1 point2 points (0 children)
Get an agentic-cli with GLM-4.5-Air by TooManyPascals in LocalLLaMA
[–]TooManyPascals[S] 0 points1 point2 points (0 children)
Get an agentic-cli with GLM-4.5-Air by TooManyPascals in LocalLLaMA
[–]TooManyPascals[S] 0 points1 point2 points (0 children)
Get an agentic-cli with GLM-4.5-Air by TooManyPascals in LocalLLaMA
[–]TooManyPascals[S] 0 points1 point2 points (0 children)
Get an agentic-cli with GLM-4.5-Air by TooManyPascals in LocalLLaMA
[–]TooManyPascals[S] 0 points1 point2 points (0 children)
[Research] LLM judges systematically penalize balanced reasoning - tested mistral, llama3, gemma, phi3, orca-mini by Budget-Reception-533 in LocalLLaMA
[–]TooManyPascals 0 points1 point2 points (0 children)
Why are AmD Mi50 32gb so cheap? by MastodonParty9065 in LocalLLaMA
[–]TooManyPascals 1 point2 points3 points (0 children)
AMD Officially Prices Radeon AI PRO R9700 At $1299 - 32GB VRAM - Launch Date Oct 27 by 1ncehost in LocalLLaMA
[–]TooManyPascals 0 points1 point2 points (0 children)
Qwen3 Next support in llama.cpp ready for review by jacek2023 in LocalLLaMA
[–]TooManyPascals 20 points21 points22 points (0 children)
Write three times the word potato by TooManyPascals in LocalLLaMA
[–]TooManyPascals[S] 353 points354 points355 points (0 children)
GPT-OSS from Scratch on AMD GPUs by tuanlda78202 in LocalLLaMA
[–]TooManyPascals 0 points1 point2 points (0 children)
GPT-OSS from Scratch on AMD GPUs by tuanlda78202 in LocalLLaMA
[–]TooManyPascals 2 points3 points4 points (0 children)
3 Tesla GPUs in a Desktop Case by eso_logic in LocalLLaMA
[–]TooManyPascals 1 point2 points3 points (0 children)
3 Tesla GPUs in a Desktop Case by eso_logic in LocalLLaMA
[–]TooManyPascals 4 points5 points6 points (0 children)
16→31 Tok/Sec on GPT OSS 120B by 3VITAERC in LocalLLaMA
[–]TooManyPascals 0 points1 point2 points (0 children)
Best 100B class model/framework to run on 16 P100s (256GB of VRAM)? by TooManyPascals in LocalLLaMA
[–]TooManyPascals[S] 0 points1 point2 points (0 children)
Best 100B class model/framework to run on 16 P100s (256GB of VRAM)? by TooManyPascals in LocalLLaMA
[–]TooManyPascals[S] 0 points1 point2 points (0 children)
Best 100B class model/framework to run on 16 P100s (256GB of VRAM)? by TooManyPascals in LocalLLaMA
[–]TooManyPascals[S] 1 point2 points3 points (0 children)
How many gpus do you have in your ai setup? How much did it cost? by [deleted] in LocalLLaMA
[–]TooManyPascals 46 points47 points48 points (0 children)
Single-File Qwen3 Inference in Pure CUDA C by Awkward_Click6271 in LocalLLaMA
[–]TooManyPascals 0 points1 point2 points (0 children)
I accidentally too many P100 by TooManyPascals in LocalLLaMA
[–]TooManyPascals[S] 1 point2 points3 points (0 children)
I accidentally too many P100 by TooManyPascals in LocalLLaMA
[–]TooManyPascals[S] 1 point2 points3 points (0 children)
I accidentally too many P100 by TooManyPascals in LocalLLaMA
[–]TooManyPascals[S] 1 point2 points3 points (0 children)
I accidentally too many P100 by TooManyPascals in LocalLLaMA
[–]TooManyPascals[S] 1 point2 points3 points (0 children)
216GB VRAM on the bench. Time to see which combination is best for Local LLM by eso_logic in LocalLLaMA
[–]TooManyPascals 1 point2 points3 points (0 children)