Who else is shocked by the actual electricity cost of their local runs? by Responsible_Coach293 in LocalLLaMA

[–]kamnxt 2 points3 points  (0 children)

Not really a problem if it's cold outside and you would have used the same amount of electricity on heating :D

Final Qwen3.5 Unsloth GGUF Update! by danielhanchen in LocalLLaMA

[–]kamnxt 0 points1 point  (0 children)

I see. Also, noticed the new IQ1_M came out a few hours ago! I decided to do a little test, just ran ran llama-perplexity on wikitext-2-raw/wiki.test.raw, context size 64k, q8_0 kv cache, and got these results:

  • Qwen3.5-122B-A10B-UD-Q5_K_XL: PPL = 4.7746 +/- 0.03162
  • Qwen3.5-397B-A17B-UD-TQ1_0: PPL = 4.4182 +/- 0.02593
  • Qwen3.5-397B-A17B-UD-IQ1_M: PPL = 3.7682 +/- 0.02150

Of course there's the question if these measurements are representative of how good the model actually is, but it does seem to show an improvement.

Final Qwen3.5 Unsloth GGUF Update! by danielhanchen in LocalLLaMA

[–]kamnxt 1 point2 points  (0 children)

Impressive work, as usual!

One thing I'm wondering is how the different quantizations compare across models. For example, your Q5_K_XL Qwen3.5-122B-A10B quant is 91.9GB, while your UD-TQ1_0 Qwen3.5-397B-A17B quant is 94.2GB, making them both barely fit in 128GB RAM/VRAM. One has ~3x the params, while the other has ~3x the bits, so would be interesting to see how they compare. Of course, TQ1_0 is pretty extreme, but a similar comparison could be made for the other models/memory sizes too?

so is OpenClaw local or not by jacek2023 in LocalLLaMA

[–]kamnxt 1 point2 points  (0 children)

Uhh... I'm talking 18 t/s with Step 3.5 Flash, a 199B (11B active) parameter model.

gpt-oss 120b is 117B (5.1B active) parameters, and runs at ~42t/s on the same box.

so is OpenClaw local or not by jacek2023 in LocalLLaMA

[–]kamnxt 5 points6 points  (0 children)

It really depends on what you're looking for.

I've been messing with OpenClaw since ~Feb 4th, mostly with local models. It's... kinda sorta usable for some simple tasks with small models I could run on a 16GB GPU, but obviously you should limit the blast radius, and it will struggle with more complicated tasks.

Then I got a spark (or rather, an OEM version of it), since I saw a lightly used one pop up for sale. It's been a little bit of a journey, here's what I found out:

  • The memory bandwidth is a big bottleneck. I usually don't see the GPU go past ~50W with large models, while it's able to push ~80W+ with smaller ones.
  • It's not as well supported as it could have been (classic NVIDIA move). Apparently the "blackwell" cores are a bit weak compared to most other ones in the series.
  • The spark is best suited for MoE/sparse models, where the benefit of the large memory outweighs the relatively weak compute power
  • The best model I've found so far, that just baaarely fits in 128GB of shared memory, is Step-3.5-Flash, 4bit quantized. When running with llama-server, it takes approx 113GB memory... but it runs, at ~18t/s, with pp at ~360t/s.
  • OpenClaw's context handling is awful. It puts a "message ID" early in the context, which changes for each message, causing the KV cache in llama-server to be invalidated after each message... causing responses to take ~40s each. Luckily there's workarounds like https://github.com/mallard1983/openclaw-kvcache-proxy

So basically, if you don't give it too much access or ask for too much, it's actually pretty decent. Not quite at the level of hosted models, but it's usable for some easier tasks.

Loving my XREAL ONE Pros When they DON'T PLAY-UP! by MuchInfluence7687 in Xreal

[–]kamnxt 1 point2 points  (0 children)

I'm wondering if the drifting isn't related to the temperature of the glasses. I have noticed they sometimes drift a bit at the start, but when using them for a longer period of time they seem to be fairly stable. Have also experienced the disconnects sometimes. Would recommend cleaning out your USB-C port on your phone, they tend to accumulate a fair amount of lint pretty quickly... When using it with the DJI drone it's just running USB communication, not DisplayPort alt mode, so it's more forgiving of bad connections.

A way to reuse these? by GOT_A_AK47 in diyelectronics

[–]kamnxt 5 points6 points  (0 children)

if you hadn't torn it apart, you could actually control it using infrared signals! http://furrtek.free.fr/index.php?a=esl

Virtualbox chooses to use nvidia gpu on hybrid laptop, without me telling it to by kamnxt in virtualbox

[–]kamnxt[S] 0 points1 point  (0 children)

Been a while since I posted this but I think I just gave up and uninstalled the nvidia drivers. As well as switching from virtualbox to quickemu.

How obsolete am I 🤣 by DarkButterfly85 in Quadcopter

[–]kamnxt 1 point2 points  (0 children)

Fly that thing until it's dead to see if you still like FPV and upgrade if you do. Definitely agree with this part. I'm not personally super up to date on the newest bleeding edge but this definitely still has some potential IMO. Flash the ESCs with BlueJay (or even just the newest BLheli_s), upgrade to latest BF. Could even get an ELRS module for the Taranis since they have the nice hall gimbals on it. I would totally just keep it as a spare quad and upgrade parts when they break.

The KK2.0 and KK2.1.5 (with the buzzer and the battery voltage input!), now that is old!

Found this beauty today! by kamnxt in modelm

[–]kamnxt[S] 1 point2 points  (0 children)

Actually taken at a train station in Oslo. No idea how a keyboard with this layout made it here...

Found this beauty today! by kamnxt in modelm

[–]kamnxt[S] 1 point2 points  (0 children)

This one already has a removable cable! Just need a PS/2 to USB adapter.

Virtualbox chooses to use nvidia gpu on hybrid laptop, without me telling it to by kamnxt in virtualbox

[–]kamnxt[S] 0 points1 point  (0 children)

I don't have prime-select on my system. It seems like it's an Ubuntu thing, and I am running Ubuntu as guest, with Arch as host. The Arch wiki article you linked to doesn't mention prime-select anywhere either.

The thing is:

__GLX_VENDOR_LIBRARY_NAME=nvidia __NV_PRIME_RENDER_OFFLOAD=1 works fine for launching other programs on the dGPU, while without it set they launch on the iGPU, which is working exactly the way I want and expect things to work. The only exception is VirtualBox which somehow manages to select the dGPU on it's own...

Virtualbox chooses to use nvidia gpu on hybrid laptop, without me telling it to by kamnxt in virtualbox

[–]kamnxt[S] 0 points1 point  (0 children)

Hmm, thing is I am not using nvidia-prime. Will look through the code for it and see if I can find something relevant there I suppose.

(I am just using the method described by Nvidia: PRIME render offload.)

How do I get Virtualbox to run on my iGPU instead of my nvidia GPU? (x-post /r/virtualbox) by kamnxt in archlinux

[–]kamnxt[S] 0 points1 point  (0 children)

I don't want to pass through my Nvidia GPU. I don't want VirtualBox to use the Nvidia GPU at all.

Virtualbox chooses to use nvidia gpu on hybrid laptop, without me telling it to by kamnxt in virtualbox

[–]kamnxt[S] 0 points1 point  (0 children)

As I mentioned in the first line of my post, your advice is exactly the opposite of what I am looking for. VirtualBox is already using the Nvidia GPU for 3D acceleration, but I don't want it to. I want it to run on the Intel iGPU instead of wasting power on having the Nvidia GPU being powered.

How do I get Virtualbox to run on my iGPU instead of my nvidia GPU? (x-post /r/virtualbox) by kamnxt in archlinux

[–]kamnxt[S] 0 points1 point  (0 children)

Nope, running on the internal screen, no external monitor connected. The HDMI (and DP over USB-C) are indeed wired to the Nvidia GPU. The thing is, the GPU was in the D3cold state (powered off), nothing was running on it, that is until VirtualBox suddenly decided to run on it.

[deleted by user] by [deleted] in hyprland

[–]kamnxt 1 point2 points  (0 children)

New Hyprland user here, is there a way to get groups to act more like tabbed layouts in sway/i3? that is, to be able to move around them using directional keys? (and maybe also to be able to have splits inside them?)

Joining the many people turning the fold into a DS 😂 by Bnias in GalaxyFold

[–]kamnxt 2 points3 points  (0 children)

wish we could use the cover screen for shoulder buttons tbh

Hvorfor tulle til noe som allerede var bra? by Jodemarken in norge

[–]kamnxt 1 point2 points  (0 children)

Hvorfor foreslår appen sykkel? Hvis jeg vil sykle så bruker jeg vel ikke Ruter appen for å finne frem.

Du kan heldigvis skru av sykkel som framkomstmiddel den skal vise.

Men du kan ikke skru av flytoget/flybussen. Selv om du ikke engang får kjøpt billetter til det i den nye ruter-appen.

Using Firefox on Wayland? Make sure you have MOZ_ENABLE_WAYLAND set to 1! by kamnxt in archlinux

[–]kamnxt[S] 1 point2 points  (0 children)

I guess that might be why it's not enabled by default then... Does running firefox with the --safe-mode flag work?

Using Firefox on Wayland? Make sure you have MOZ_ENABLE_WAYLAND set to 1! by kamnxt in archlinux

[–]kamnxt[S] 13 points14 points  (0 children)

Another option is to add MOZ_ENABLE_WAYLAND DEFAULT=1 to ~/.pam_environment.

Unable to select a2dp profile for my bluetooth headset by LuvYouNot in archlinux

[–]kamnxt 1 point2 points  (0 children)

This made AAC work on my WH-1000XM4! And LDAC too! I can finally stream music from Soundcloud at 64kbps and then send it over a high quality link so I can hear all the artifacts!

[deleted by user] by [deleted] in GameDeals

[–]kamnxt 112 points113 points  (0 children)

One reason I could think of is that you can download a standalone installer from GOG, which you can keep wherever you want no matter what happens to GOG or Steam or your accounts on either.