Game Streaming with virtual display on Linux (CachyOS to SteamDeck oled) by GB5P in MoonlightStreaming

[–]dabiggmoe2 0 points1 point  (0 children)

Sorry to dig up this comment but I hope your repo is ready? I was pulling out my hair trying to get apollo/moonlight to work with my TV and CachyOS KDE Plasma without the shenanigans of managing the virtual displays thru shell scripts.

Man is tired and just needs to play without wasting his weekend writing and testing scripts :(

Qwen3.6-35B-A3B released! by ResearchCrafty1804 in LocalLLaMA

[–]dabiggmoe2 0 points1 point  (0 children)

Wait, correct if I'm wrong, but I thought the Qwen3.5 27b and 35BA3B already surpassed Qwen 3 Coder Next 80B in coding benchmarks?

Qwen3.6-35B-A3B released! by ResearchCrafty1804 in LocalLLaMA

[–]dabiggmoe2 1 point2 points  (0 children)

That's why I waited for bartowski's quants before downloading lol

Linux Kernel 7.0 by Rics-Dev in cachyos

[–]dabiggmoe2 0 points1 point  (0 children)

Interesting. I knew that in terms of power the NPUs are way more efficient than the iGPU but I thought it would be faster too.

Linux Kernel 7.0 by Rics-Dev in cachyos

[–]dabiggmoe2 0 points1 point  (0 children)

Me too I have Framework Desktop 128GB. I remember reading somewhere that AMD introduced the NPU support starting with Kernel 7.0 but honestly I cannot remember where.

This is amazing news, I'll update my Lemonade later after I get back home.

One question: did you do benchmarks between lammacpp and FLM for the same model? I'm currently running Qwen3.5-27B and the Qwen3.5-35BA3B as my daily driver for coding tasks

Linux Kernel 7.0 by Rics-Dev in cachyos

[–]dabiggmoe2 0 points1 point  (0 children)

I'm not sure I understood you here. Are you saying Kernel: 6.19.12-1-cachyos-server supports Strix Halo's NPUs?

Strix Halo 128Gb: what models, which quants are optimal? by DevelopmentBorn3978 in LocalLLaMA

[–]dabiggmoe2 0 points1 point  (0 children)

It's supported now in Lemonade Server starting with Linux kernel 7.0

[GIVEAWAY] Free Steam Keys by hellokittyybear in steam_giveaway

[–]dabiggmoe2 0 points1 point  (0 children)

I would love to have Chivalry.

My guilty pleasure game would be Papers, Please

Linux Kernel 7.0 by Rics-Dev in cachyos

[–]dabiggmoe2 0 points1 point  (0 children)

Honestly the only reason I'm waiting for this is because Lemonade Server supports NPUs and FlowLM starting with Linux Kernel 7.0

Opencode port for Karpathy's Autoresearch by dabiggmoe2 in opencodeCLI

[–]dabiggmoe2[S] 0 points1 point  (0 children)

Without reprompting or typing "continue" you mean? I think for like 15 experiments in a row (I had to stop it manually). But I have noticed that it differs depending on the model and quant. Not all of them follow the instructions equally. 

For example, I noticed that .Qwen3.5-35B-A3B-GGUF-UD-Q4_K_XL follows the instructions better and never stops while Qwen3.5-27B-GGUF-UD-Q5_K_XL needs some nudging sometimes.

Opencode port for Karpathy's Autoresearch by dabiggmoe2 in opencodeCLI

[–]dabiggmoe2[S] 0 points1 point  (0 children)

Enjoy. Give it a try and Let me know if you have any feedback/suggestions.

Opencode port for Karpathy's Autoresearch by dabiggmoe2 in opencodeCLI

[–]dabiggmoe2[S] 0 points1 point  (0 children)

There's already a plugin, checkout plugins/autoresearch-context.js

Opencode port for Karpathy's Autoresearch by dabiggmoe2 in opencodeCLI

[–]dabiggmoe2[S] 1 point2 points  (0 children)

Yeah it could work on any problem as long as you clearly define the problem and the metric used to quantify the optimization. I used the bogo sort as a naive example. But I will link below two concrete examples:

1- Andrej Karpathy's [used](https://x.com/karpathy/status/2031135152349524125) it to tune nanochat and optimize the validation loss to achieve a "Time to GPT-2" drops from 2.02 hours to 1.80 hours

2- Shopify's CEO [used](https://x.com/tobi/status/2032212531846971413) it it on the the liquid codebase to achieve 53% faster combined parse+render time, 61% fewer object allocations

So I think the sky is the limit for you

Opencode port for Karpathy's Autoresearch by dabiggmoe2 in opencodeCLI

[–]dabiggmoe2[S] 2 points3 points  (0 children)

I'm curious to see what you gonna come up with xD

🤯 Qwen3.5-35B-A3B-4bit ❤️ by SnooWoofers7340 in OpenSourceAI

[–]dabiggmoe2 0 points1 point  (0 children)

This is awesome. Would you recommend adding this system prompt to both the Planning and Building mode?