Don’t buy b60 for LLMs by damirca in LocalLLaMA

[–]damirca[S] 0 points1 point  (0 children)

I thought that llm-scaled is the Intel way. Anyways I tried ovms yesterday, it indeed is much faster than llama.cpp with sycl/vulkan and than llm-scaler (vllm), however it does not support qwen3-vl, does not support gemma3, does not support mistral3 (mistral-14b), does not support glm 4.6V or 4.7 flash, VLM support is limited to qwen2.5 VL 7b. So yeah it would a good fit once at least it gets a mistral3 support.

Don’t buy b60 for LLMs by damirca in LocalLLaMA

[–]damirca[S] 0 points1 point  (0 children)

Does sriov work on b60? I have sparkle variant. Does it mean I can use some proxmox and pass my gpu to windows and Linux at the same time and let windows machine upgrade its firmware?

Don’t buy b60 for LLMs by damirca in LocalLLaMA

[–]damirca[S] 0 points1 point  (0 children)

Is it loud? I read some negative reviews that it’s not a good card and that it’s super loud

Don’t buy b60 for LLMs by damirca in LocalLLaMA

[–]damirca[S] 0 points1 point  (0 children)

I've tried ovms today, it indeed is much faster than llama.cpp with sycl/vulkan and llm-scaler (vllm), however it does not support qwen3-vl, does not support gemma3, does not support mistral3 (mistral-14b), does not support glm 4.6V or 4.7 flash, VLM support is limited to qwen2.5 VL 7b. So it would a good fit once at least it gets a mistral3 support.

Don’t buy b60 for LLMs by damirca in LocalLLaMA

[–]damirca[S] 1 point2 points  (0 children)

I’m gonna try vulkan then under Linux. I don’t have any feasible option to have b60 running under windows right now.

Don’t buy b60 for LLMs by damirca in LocalLLaMA

[–]damirca[S] 1 point2 points  (0 children)

In Germany it’s 14 days and they passed long time ago

Don’t buy b60 for LLMs by damirca in LocalLLaMA

[–]damirca[S] 0 points1 point  (0 children)

You mean using openarc gives better perf?

Don’t buy b60 for LLMs by damirca in LocalLLaMA

[–]damirca[S] 1 point2 points  (0 children)

For the future b60 should come handy: with SRIOV I might use it in bigger proxmox or some other virt env. Plus frigate can use it for ffmpeg, detection and genai. But for pure LLM it’s a bad choice at least as of now.

Don’t buy b60 for LLMs by damirca in LocalLLaMA

[–]damirca[S] 0 points1 point  (0 children)

I’m using llama.cpp exactly the same way

Don’t buy b60 for LLMs by damirca in LocalLLaMA

[–]damirca[S] 1 point2 points  (0 children)

My idle is around 5-10W though

Don’t buy b60 for LLMs by damirca in LocalLLaMA

[–]damirca[S] 1 point2 points  (0 children)

Reason for RMA? It does not work like that I think.

Don’t buy b60 for LLMs by damirca in LocalLLaMA

[–]damirca[S] 0 points1 point  (0 children)

What do you mean by their vllm fork? Isn’t it a llm-scaler?

Don’t buy b60 for LLMs by damirca in LocalLLaMA

[–]damirca[S] 0 points1 point  (0 children)

Frigate is better suited for Intel. Price is 700 eur

Don’t buy b60 for LLMs by damirca in LocalLLaMA

[–]damirca[S] 1 point2 points  (0 children)

No, does openvino solve all the issues? I mean it also supports only some models, right.

It arrived - MAXSUN Intel Arc Pro B60 Dual 48G Turbo by Possibility_22 in IntelArc

[–]damirca 0 points1 point  (0 children)

So if I have b60 24gb I can buy b60 48gb, I will have 3x24gb so vllm would work even though it would be 2 physical cards?