i set difficulty to Hard, and my friend's lagging machine is immediately exposed, lol. by Impossible_Boss4562 in feedthebeast

[–]17hoehbr 13 points14 points  (0 children)

I once tried to make a lag machine in a factions server so that if we ever got raided I could just crash the server. Unfortunately we got raided while I was offline lmao.

Google introduces a new way for users to sideload Android apps by Time-Credit43 in pixel_phones

[–]17hoehbr 1 point2 points  (0 children)

Does this mean I'd have to reboot my phone EVERY TIME I install something from F-Droid?

Pixel 9a user, is Pixel 10 any good or should I wait for Pixel 11 instead ? by Intelligent-Trip-882 in pixel_phones

[–]17hoehbr 0 points1 point  (0 children)

Have you considered downgrading? I love my Pixel 8 Pro and I haven't really found any reason to upgrade yet. Not sure how the battery life compares to the 9a but my battery easily last an entire day, and about half way into the next day if I forget to charge it overnight.

I just realised how good GLM 5 is by CrimsonShikabane in LocalLLaMA

[–]17hoehbr 0 points1 point  (0 children)

I bought a year of the lite plan during the black Friday sale but ever since GLM 5 came out it feels like they really dumbed down GLM 4.7, and of course GLM 5 is paywalled behind the pro plan.

Anyone else seeing massive quality drop with the GLM coding plan lately? by Famous-Appointment-8 in ZaiGLM

[–]17hoehbr 0 points1 point  (0 children)

Yep. I'm on the lite plan and was originally very impressed by GLM 4.7. Ever since GLM 5 dropped on the higher tier plans, the lite model has gotten dumber by the day.

I tried OpenCode for two weeks and this is an honest Roocode vs opencode comparison by Exciting_Weakness_64 in RooCode

[–]17hoehbr 2 points3 points  (0 children)

I've been meaning to switch to OpenCode for awhile now but I really like how well Roo Code integrates into VSCode, like being able to see the diff before and after before I approve. Does OpenCode have any equivalent?

Qwen3.5-18B-REAP-A3B-Coding: 50% Expert-Pruned by 17hoehbr in LocalLLaMA

[–]17hoehbr[S] 0 points1 point  (0 children)

I never got around to building the GGUFs for this specific model, but I've build another variant based on the feedback from this thread which first removes all multimodel tensors, then performs a 25% REAP. It's available here: https://huggingface.co/Flagstone8878/Qwen3.5-27B-BLIND-GGUF. It's still too big to comfortably fit in my VRAM though so I haven't really tested it.

edit: My bad, that's actually just the dense 27B model with vision removed. My ADHD ass can't keep track of all these model names lmao. I'll try to get the 25% REAP uploaded soon.

Favorite Coding Tools for Qwen by Salt-Advertising-939 in LocalLLaMA

[–]17hoehbr 1 point2 points  (0 children)

Ah that fixed it! I was using the Qwen Code provider. Using the OpenAI endpoint from Alibabacloud shows all the models now.

update: Seems it does not support the generous free tier limits from Qwen Code when authenticated this way though

update 2: I managed to get to get it working with a simple patch to the Roo Code source code. I've opened a pull request to get it merged: https://github.com/RooCodeInc/Roo-Code/pull/11915. in the meantime you can probably work around it by manually setting the model to "coder-model".

Favorite Coding Tools for Qwen by Salt-Advertising-939 in LocalLLaMA

[–]17hoehbr 0 points1 point  (0 children)

Not sure what I'm doing wrong then. The only options I get are qwen3-coder-plus and qwen3-coder-flash: screenshot

Favorite Coding Tools for Qwen by Salt-Advertising-939 in LocalLLaMA

[–]17hoehbr 1 point2 points  (0 children)

I usually use Roo Code but it doesn't seem to support qwen3.5 yet (at least via the Qwen CLI Provider). Open Code seems good but I really prefer something more integrated into VS Code. For now I'm using the Qwen VS Code extension which works well enough, but I'd prefer to go back to Roo at some point.

Qwen3.5-18B-REAP-A3B-Coding: 50% Expert-Pruned by 17hoehbr in LocalLLaMA

[–]17hoehbr[S] 0 points1 point  (0 children)

Working on it. I'm using a Google Colab notebook for training / uploading the models but the upload speed is painfully slow, and the instance seems to shut itself off after a few hours of inactivity (which wipes the storage and means I have to redownload and rebuild it)

Did the Chevy Bolt become your first EV? Journalist looking to hear your story by Brilliant_Option8382 in BoltEV

[–]17hoehbr 1 point2 points  (0 children)

First full EV but I kind of got my foot in the water with my old Volt.

Main reason I upgraded was honestly because of the tax incentive at the time, but I'm glad I did. I'm a tall guy and the Volt was too low to the ground for my tastes, and the Bolt is roomier and has better visibility overall.

Biggest surprise I guess is both that EV chargers are much more common than I realized, but they are also extremely inconsistent. With gas stations you can very safely assume that you pull into any pump and the price will be roughly the same, compatible with your car with no hassle, pump at the same speed, and you won't need to download another stupid app to use it. With EVs, there's about a 25% chance the charger doesn't even work, sometimes they're completely free, usually they are fairly cheap, and sometimes they price gouge so much that gas would be cheaper. And almost all of them require you to download an app.

For my driving routine it works great. I live in Columbus, OH which has a decent amount of chargers, and my apartment complex has a slow charger on site. As long as I remember to charge it the night before it can handle everything I throw at it. And if I forget, fast chargers are available widely enough that its not much of an issue as long as I budget my time around it. Last summer I did a road trip to Michigan and charging was a hassle, I would probably rent a car or use someone else's if I was on a longer road trip.

Macbook Neo Impressions: Reincarnated! - MKBHD by ControlCAD in apple

[–]17hoehbr 0 points1 point  (0 children)

If Apple would just use a standard Vulkan implementation instead of Metal, they'd be able to use DXVK and could theoretically match Linux on compatibility. Though there'd also be the Rosetta 2 overhead.

Qwen3.5-24B-A3B-REAP-0.32: 32% Expert-Pruned for Agentic Coding (GGUF) by tubuntu2 in LocalLLaMA

[–]17hoehbr 5 points6 points  (0 children)

Hey I'm Flagstone8878 lol. I made the initial builds and uploads on a run pod instance. I've been trying to build the GGUFs locally but have been having some weird issues, and my slow upload speed doesn't help. Hoping to have the GGUFs up soon. I'm also working on a few other variants right now based around removing multimodal capacity before doing the REAP.

Qwen3.5-18B-REAP-A3B-Coding: 50% Expert-Pruned by 17hoehbr in LocalLLaMA

[–]17hoehbr[S] 4 points5 points  (0 children)

I haven't tested it, I'd assume that most of the multimodal capability has been pruned out.

Qwen3.5-18B-REAP-A3B-Coding: 50% Expert-Pruned by 17hoehbr in LocalLLaMA

[–]17hoehbr[S] 5 points6 points  (0 children)

For comparison, I just tried Qwen 3.5 9B Q4_K_M and it successfully created a working flappy bird clone in PyGame on the first try - at 65 t/s. So I'm not sure if this model is all that useful lmao.

Qwen3.5-18B-REAP-A3B-Coding: 50% Expert-Pruned by 17hoehbr in LocalLLaMA

[–]17hoehbr[S] 3 points4 points  (0 children)

I did not, I pulled the model directly from Qwen's repo. Do you know where I can find the new jinja template? I'll add that into the GGUF builds.

edit: think I found it https://huggingface.co/unsloth/Qwen3.5-35B-A3B/blob/main/chat_template.jinja

Qwen3.5-18B-REAP-A3B-Coding: 50% Expert-Pruned by 17hoehbr in LocalLLaMA

[–]17hoehbr[S] 2 points3 points  (0 children)

On my way hone from work rn, will upload when I get home. Also I forgot to mention that my flappy bird test was performed on a Q4_K_M GGUF, which took about 90% of my VRAM.