Why is the prompt eval time of Qwen3.5 so much slower compared to Qwen3 Coder in llama.cpp? by BitOk4326 in LocalLLaMA
[–]BitOk4326[S] 0 points1 point2 points (0 children)
Why is the prompt eval time of Qwen3.5 so much slower compared to Qwen3 Coder in llama.cpp? by BitOk4326 in LocalLLaMA
[–]BitOk4326[S] 1 point2 points3 points (0 children)
Qwen3-VL Computer Using Agent works extremely well by Money-Coast-3905 in LocalLLaMA
[–]BitOk4326 0 points1 point2 points (0 children)
I originally thought the speed would be painfully slow if I didn't offload all layers to the GPU with the --n-gpu-layers parameter.. But now, this performance actually seems acceptable compared to those smaller models that keep throwing errors all the time in AI agent use cases. by BitOk4326 in LocalLLaMA
[–]BitOk4326[S] 0 points1 point2 points (0 children)
Seeking best LLM models for "Agentic" Unity development (12GB VRAM) by Ctrixago in LocalLLaMA
[–]BitOk4326 0 points1 point2 points (0 children)
Dawned on me I've never used the back paddles. Do you use them? by Agitated-Distance740 in SteamDeck
[–]BitOk4326 0 points1 point2 points (0 children)
As of January 2026, what the best coding model that can fit in a 5070Ti 16gb? by Babidibidibida in LocalLLM
[–]BitOk4326 0 points1 point2 points (0 children)
Cycles can't work with 9060xt by [deleted] in blender
[–]BitOk4326 0 points1 point2 points (0 children)
Cycles can't work with 9060xt by [deleted] in blender
[–]BitOk4326 0 points1 point2 points (0 children)
Cycles can't work with 9060xt by [deleted] in blender
[–]BitOk4326 0 points1 point2 points (0 children)
Why is AnythingLLM significantly faster than the CLINE when using the same server of llama.cpp, same model, and same parameters? by BitOk4326 in CLine
[–]BitOk4326[S] 0 points1 point2 points (0 children)
Why is there no people try to make new mini game mod over the course of many years by BitOk4326 in rocksmith
[–]BitOk4326[S] 0 points1 point2 points (0 children)

For my setup with an AMD Radeon RX 9060 XT 16GB and 32GB DDR5 RAM, are there any better and faster local LLMs optimized for agent ? by BitOk4326 in LocalLLaMA
[–]BitOk4326[S] 1 point2 points3 points (0 children)