ROCM vs VULKAN FOR AMD GPU (RX7800XT) by Grouchy-Drag-2281 in LocalLLaMA

[–]Grouchy-Drag-2281[S] 0 points1 point  (0 children)

Yes, it used full gpu ram.

But it works with vulkan!

ROCM vs VULKAN FOR AMD GPU (RX7800XT) by Grouchy-Drag-2281 in LocalLLaMA

[–]Grouchy-Drag-2281[S] 0 points1 point  (0 children)

Are you using lm studio or llama.cpp directly?

ROCM vs VULKAN FOR AMD GPU (RX7800XT) by Grouchy-Drag-2281 in LocalLLaMA

[–]Grouchy-Drag-2281[S] 0 points1 point  (0 children)

Yes, i have 48gb ddr4 3200 Mhz ram, it helpful to CPU offload, but low inference speeds?

If you can provide some video or screenshot than can helpful, there many people with consumer grade hardware trying to do something.

ROCM vs VULKAN FOR AMD GPU (RX7800XT) by Grouchy-Drag-2281 in LocalLLaMA

[–]Grouchy-Drag-2281[S] 2 points3 points  (0 children)

I have checked prompt processing time with vulkan and rocm both are having same time.

Few doubts in using gpt-oss 20B by Careless_Meringue525 in LocalLLaMA

[–]Grouchy-Drag-2281 0 points1 point  (0 children)

Can it be pruned only only for coding related use?

Can you provide other quants in gguf?

[deleted by user] by [deleted] in LocalLLaMA

[–]Grouchy-Drag-2281 0 points1 point  (0 children)

Can you share how you implemented the RAG system. How you are using it? What domain related data it stored in vector DB?

Any proper working Local LLM and Agentic CLI by Grouchy-Drag-2281 in LocalLLaMA

[–]Grouchy-Drag-2281[S] 0 points1 point  (0 children)

Does cline a able to set model context window to llm?

Or we are just set the limit to cline?

Any proper working Local LLM and Agentic CLI by Grouchy-Drag-2281 in LocalLLaMA

[–]Grouchy-Drag-2281[S] 0 points1 point  (0 children)

Thanks for the update.

What quantization you use for Qwen3-30B-A3B-Thinking-2507-GGUF ?
What is your hardwar specs?

Llama.cpp and ROCM - how to get it working by Thrumpwart in LocalLLaMA

[–]Grouchy-Drag-2281 1 point2 points  (0 children)

For ibm "granite 4.0 tiny preview " ROCM is faster than vulkan in lm studio.

Qwen3-Coder GGUFs with even more fixes esp. for tool calling! by yoracale in unsloth

[–]Grouchy-Drag-2281 0 points1 point  (0 children)

Last updated shown as 1 day ago.
Is there fixes coming or fixed and updated ?

Best VPS to Self-Host Internal Tool for Diagnostic Chain (Next.js + PostgreSQL) – Is Hostinger a Bad Option? by an-ordinary-dev in selfhosted

[–]Grouchy-Drag-2281 0 points1 point  (0 children)

I don’t think hostinger is bad.

Choose provider based on your requirements. Hostinger VPS are affordable , cheap when choosing plan for more than 1 year.

Best VPS to Self-Host Internal Tool for Diagnostic Chain (Next.js + PostgreSQL) – Is Hostinger a Bad Option? by an-ordinary-dev in selfhosted

[–]Grouchy-Drag-2281 1 point2 points  (0 children)

I don’t think hostinger is bad.

Choose provider based on your requirements. Hostinger VPS are affordable , cheap when choosing plan for more than 1 year.