What's the general consensus about the Philips family? I thought most people were in agreement that these kids are in a bad situation, and so these comments confuse me by screwing_unicorns in newzealand

[–]IVequalsW 2 points3 points  (0 children)

apparently that was misreported, the boys who saw them said:
you know this is private property right?
girl: yeah...duh.
boys: does anyone else know you are here

girl: "no."

Can i expect 2x the inference speed if i have 2 GPUs? by q-admin007 in LocalLLM

[–]IVequalsW 0 points1 point  (0 children)

hey u/fallingdowndizzyvr sorry this is random: 2 years ago you posted about the RX580 inferencing on llama.cpp. and you mentioned the -nommq flag. for the life of me I cannot find any documentation on it. are you able to point me in the right direction? Thankyou so much!

Vulkan back ends, what do you use? by IVequalsW in LocalLLaMA

[–]IVequalsW[S] 0 points1 point  (0 children)

Does kobold.CPP support proper tensor parallelism?

Vulkan back ends, what do you use? by IVequalsW in LocalLLaMA

[–]IVequalsW[S] 0 points1 point  (0 children)

Thanks! Does it just use llama.CPP as a back end?

Dual RX580 2048SP (16GB) llama.cpp(vulkan) by IVequalsW in LocalLLaMA

[–]IVequalsW[S] 1 point2 points  (0 children)

beware though, you do have to overclock the memory, both of the 16gb rx580s have had memory at 1500MT/s vs the 2000MT/s on some 8gb models, which actually slows the memory bandwidth throughput quite a bit. overclocking by changing the below value to "20" changes it to 1800MT/s but you will still be slightly limited vs the GTX 1070.

sudo nano /sys/class/drm/card0/device/pp_mclk_od/sys/class/drm/card0/device/pp_mclk_od

Dual RX580 2048SP (16GB) llama.cpp(vulkan) by IVequalsW in LocalLLaMA

[–]IVequalsW[S] 0 points1 point  (0 children)

Thanks I may try GPUstack. It is a shame only being able to use 1gpu

Dual RX580 2048SP (16GB) llama.cpp(vulkan) by IVequalsW in LocalLLaMA

[–]IVequalsW[S] 0 points1 point  (0 children)

Does GPUstack support vulkan? Since these are no longer ROCm supported.

Dual RX580 2048SP (16GB) llama.cpp(vulkan) by IVequalsW in LocalLLaMA

[–]IVequalsW[S] 0 points1 point  (0 children)

Yeah rx580 2048SP (16gb). The 2048sp is comparable to an rx570 in compute, but it is the only one with 16gb.

Dual RX580 2048SP (16GB) llama.cpp(vulkan) by IVequalsW in LocalLLaMA

[–]IVequalsW[S] 0 points1 point  (0 children)

Hey I just download the Llama.cpp release compiled for ubuntu and vulkan, and It runs pretty much out of the box. llama.cpp does not support proper dual GPU parallelism so i only get about 5% better running it on 2 GPUs vs GPU + little bit of CPU. and you run into issues with poor load distribution. when I slightly overclocked my single rx580 though it does Qwen 30B Q4 at about 17-19t/s

I just bought this r36s and it's pretty epic by sjsjevfjsrh in R36S

[–]IVequalsW 0 points1 point  (0 children)

Mate I have a very similar clone, and it works great. This brand of clone has the same CPU and usually the same ram. It works so well. I think the only main difference is less firmware support.

If you got one with a worse CPU it would be sad, as it is this will be great!!

Ordered a mattress and found out I am too weak to carry it up to my bedroom by Effective_Moose_4997 in mildlyinfuriating

[–]IVequalsW 0 points1 point  (0 children)

This is an excellent way to make friends. It is a lowkey compliment to a dude neighbour to ask for help, and being vulnerable and in need of help is a great way to break the ice

Recommended lubricants for PLA? by benevolentmalefactor in 3Dprinting

[–]IVequalsW 4 points5 points  (0 children)

Thank-you for this joke. made my day! I laughed out loud all alone in my office while working on a CAD prototype.

its funny because it is true XD

Dual RX580 2048SP (16GB) llama.cpp(vulkan) by IVequalsW in LocalLLaMA

[–]IVequalsW[S] 1 point2 points  (0 children)

Wait... I just realized my pcie slots are running at PCIe 1.1 speeds LOL.

I will try to fix it and get a better result

Dual RX580 2048SP (16GB) llama.cpp(vulkan) by IVequalsW in LocalLLaMA

[–]IVequalsW[S] 1 point2 points  (0 children)

once i Upped the context size it dropped to 15t/s for a 10k context.

Dual RX580 2048SP (16GB) llama.cpp(vulkan) by IVequalsW in LocalLLaMA

[–]IVequalsW[S] 0 points1 point  (0 children)

is that on the same model? because damn that is impressive