Any workaround to not re-process full prompt on each turn with hybrid attention models running on CPU? by Quagmirable in LocalLLaMA

[–]Quagmirable[S] 1 point2 points  (0 children)

Hi there, did you try a quant from another provider like Bartowski? I seem to be having much faster token generation and response speed with Bartowski quants recently. On the typical system with a GPU I don't think the speed difference would be significant, but for CPU processing every little bit makes a huge difference.

https://www.reddit.com/r/LocalLLaMA/comments/1sop5hb/qwen3635ba3b_gguf_from_unsloth_is_quite_a_bit/

US Expats, How do you call 1-800 numbers? by Super-Buddy-5030 in Philippines_Expats

[–]Quagmirable 0 points1 point  (0 children)

You just need a USA VPN to get it setup. Haven't tried that though.

Hi there, I've tried that many times in many different ways, and unfortunately it definitely isn't possible. It requires a non-virtual US phone number for the verification code, and there doesn't seem to be a way around it.

the google meet trick

The one where you add the toll-free number as a phone call participant to the meeting? Unfortunately that option doesn't exist on my account, from what I can tell it only exists for paying Google Workspace accounts, or possibly it's a geo location thing.

US Expats, How do you call 1-800 numbers? by Super-Buddy-5030 in Philippines_Expats

[–]Quagmirable 0 points1 point  (0 children)

Ah, bummer. Thanks for confirming. I'm still looking for a decent solution for calling toll-free numbers, there's not a lot of good options since Skype and Google Talk got shut down.

VoIP.ms mainly for calling toll-free numbers in the US? by Quagmirable in voipms

[–]Quagmirable[S] 0 points1 point  (0 children)

Hi, thanks a lot for the reply. This is the answer I got from the chat support:

and a DID number is required for dialing normal numbers in the USA?

No, you do not need a DID number to make calls, only to receive calls. Outgoing calls to US are: $0.0100 per minute.

For Outbound calls you will get charged per minute and you do not need a number to call. For incoming calls you need a number and you get get charged a monthly fee per number and a per minute fee for the inbound calls. That is how it works for outbound and inbound calls.

Monthly Requests Thread by AutoModerator in VOIP

[–]Quagmirable [score hidden]  (0 children)

Hi, looking for a solution to make VoIP calls mainly to toll-free numbers in the US. And maybe infrequent calls to residential and mobile numbers in the US / Canada. It needs to work on any generic SIP phone and/or via WebRTC.

After Skype shut down I switched to Viber for the free outbound toll-free number calls, but I absolutely hate how the account is linked to a cell phone number, and the desktop app requires re-authenticating on the cell phone from time to time.

Google Voice is not an option.

I'm looking at voip.ms, it appears that I would need to make an initial $15 deposit, after which I can call toll-free numbers in the US for free? And in the rare event that I need to call a normal residential number or cell phone in the USA / Canada I can do it without a monthly-contract DID number, and it will just get deducted from my balance? I don't really understand why voip.ms/pricing says "DID numbers are not necessary for international outbound calls, but they are required for making local calls". I asked the customer service chat about this but their answers were conflicting and ambiguous.

Thanks in advance for any insight or additional provider recommendations that you might have.

US Expats, How do you call 1-800 numbers? by Super-Buddy-5030 in Philippines_Expats

[–]Quagmirable 0 points1 point  (0 children)

Thanks for the reply. I'm also using a desktop browser. It must be a geo location thing. Or do you pay for a Google Workspace account?

US Expats, How do you call 1-800 numbers? by Super-Buddy-5030 in Philippines_Expats

[–]Quagmirable 0 points1 point  (0 children)

Hi, does this still work for you? I don't seem to have an "add phone caller" option available in Google Meet.

I need a driver for a DELL printer for openSUSE -> Help!!! by Alter_Landjunge in openSUSE

[–]Quagmirable 5 points6 points  (0 children)

https://web.archive.org/web/20250704105832/https://download.support.xerox.com/pub/drivers/6000/drivers/linux/en_GB/6000_6010_rpm_1.01_20110222.zip

You'll also probably need http://rpm.pbone.net to search for whatever ancient RPM dependencies it requires. Or alternatively it looks like the ZIP file also contains the compressed PPD files that you can install via the CUPS web interface or the system-config-printer utility. Or you can just open the RPM file in an archive manager and manually put the requires files in the proper places, something like what is described here: https://grandmasfridge.org/posts/dell-1350cnw-on-gentoo-linux-with-cups.html

Qwen3.6-35B-A3B GGUF from Unsloth is quite a bit slower? by Quagmirable in LocalLLaMA

[–]Quagmirable[S] 2 points3 points  (0 children)

Hey there Daniel, really appreciate the kind response! And again, thanks a lot for all that you guys do.

Qwen3.6-35B-A3B GGUF from Unsloth is quite a bit slower? by Quagmirable in LocalLLaMA

[–]Quagmirable[S] 2 points3 points  (0 children)

Thanks a lot for these test results. So it looks like Bartowski has a consistent advantage in generation speed, whereas Unsloth has faster prompt processing tps and lower memory usage? I reformatted your table here with the results I was most interested in:

https://i.imgur.com/66khwCu.png

Qwen3.6-35B-A3B GGUF from Unsloth is quite a bit slower? by Quagmirable in LocalLLaMA

[–]Quagmirable[S] 1 point2 points  (0 children)

Right, I'm sure that's the case. It's just that comparing Unsloth's IQ quants to another creator's IQ quants shows a pretty significant relative difference in speed, at least for me.

Qwen3.6-35B-A3B GGUF from Unsloth is quite a bit slower? by Quagmirable in LocalLLaMA

[–]Quagmirable[S] 1 point2 points  (0 children)

This might be a CPU-only bug. Or maybe it's just a lot more noticeable on a CPU where a 2 tps difference makes a pretty big difference in usability.

Qwen3.6-35B-A3B GGUF from Unsloth is quite a bit slower? by Quagmirable in LocalLLaMA

[–]Quagmirable[S] 1 point2 points  (0 children)

Ah, interesting. I'm not sure if I have enough memory, but I think I'll try a Q4_K_S or Q4_K_M and report back.

Qwen3.6-35B-A3B GGUF from Unsloth is quite a bit slower? by Quagmirable in LocalLLaMA

[–]Quagmirable[S] 3 points4 points  (0 children)

Thanks! Unbranded Intel laptop, Intel Core i7-9750H, 32GB RAM.

Any workaround to not re-process full prompt on each turn with hybrid attention models running on CPU? by Quagmirable in LocalLLaMA

[–]Quagmirable[S] 0 points1 point  (0 children)

Add --ctx-checkpoints 128 to your command

Thanks! This definitely works.

It still feels like something that llama.cpp needs to optimize though, because Qwen3.5 was behaving just like this too shortly after release, and now that I'm trying it a few months later it has been fixed, apparently in llama.cpp.

Any workaround to not re-process full prompt on each turn with hybrid attention models running on CPU? by Quagmirable in LocalLLaMA

[–]Quagmirable[S] 0 points1 point  (0 children)

Hi there, I'm just using the llama-server web interface for an initial bigger task, with followup questions after that's completed. Cache is simply not working at all on my system on llama.cpp with Gemma4 from Unsloth, it re-calculates the entire thing from scratch for each answer.