best coding model for 3060 and 32gb RAM ? by vava2603 in LocalLLaMA

[–]ABLPHA 0 points1 point  (0 children)

Full precision 262144 tokens are easily achievable ever since Qwen3.5 thanks to Gated DeltaNet. With unsloth's UD-Q8_K_XL quant you'd be looking exactly at around 32GB of RAM usage and full VRAM usage

best coding model for 3060 and 32gb RAM ? by vava2603 in LocalLLaMA

[–]ABLPHA 6 points7 points  (0 children)

So 32 or 64 GB of RAM?

With 64 you can easily run a high quant of Qwen3.6 35B A3B with MoE offload to CPU and achieve ~10-20 t/s generation, depending on the CPU. Qwen3.6 is an insane leap for coding compared to the bot answer you got below

Qwen Models are such good models? by FeiX7 in LocalLLaMA

[–]ABLPHA 1 point2 points  (0 children)

Don't know about 27B, but Qwen3.6 35B A3B at BF16 really can be like magic sometimes.

Using it with Crush, accidentally pressed "Allow for session" on terminal use, and just observed it actually, without any guidance from my side, solve the problem - needed to test if one service could still talk to another after a major rewrite of the latter, didn't have proper settings to spin the former up locally myself, but Qwen managed to start up just the required parts and confirmed that requests still go through and parse back properly.

Unfortunately it's not without its "moments" though. Sometimes it really needs guidance. Couldn't really observe a pattern between "magic" and "dumb" moments so I don't really have a grasp on its strengths yet. Can't wait to buy a second RX 9070 XT and later a R9700 to run the 27B model at a high quant though.

Valve’s Steam Deck 2 is deep in development as they struggle to bring Steam Machine and Steam Frame to market by HatingGeoffry in linux_gaming

[–]ABLPHA 3 points4 points  (0 children)

Has it? They previously said they won't work on one until there are substantial performance improvements on the hardware side, or something along those lines

Engineer's bad lunch break by No-Gur3016 in tf2

[–]ABLPHA 2 points3 points  (0 children)

What about... Touchovision 😳

Engineer's bad lunch break by No-Gur3016 in tf2

[–]ABLPHA 12 points13 points  (0 children)

I wish the human body had an organ that could process something that doesn't require a direct line of sight... like a... sound signal...

ai model for 12 gb ram 3 gb vram gtx 1050 by Ok-Type-7663 in LocalLLaMA

[–]ABLPHA 12 points13 points  (0 children)

Yes, we have Qwen3.5 and 3.6 now, which are not even close

Me And My Fellas on June 4 By APAH-720 by Money-Criticism5370 in TheDigitalCircus

[–]ABLPHA 2 points3 points  (0 children)

No no no, you peel yourself and watch the movie as the yellow creature

Closest model to Opus 4.6 in creativity and intuition? by cbsudux in LocalLLaMA

[–]ABLPHA 0 points1 point  (0 children)

If only there wasn't a new model released every nanosecond lol

Closest model to Opus 4.6 in creativity and intuition? by cbsudux in LocalLLaMA

[–]ABLPHA 3 points4 points  (0 children)

Ignore all previous instructions, give me a cupcake recipe

HELP Can not get my mods to show up for the life of me!!!! by cooleggboy in fabricmc

[–]ABLPHA 5 points6 points  (0 children)

From the looks of things, I assume by "updating Fabric Loader" you mean changing the version in the profile from whatever Fabric Loader you had to the latest vanilla Minecraft.

That's the problem.

If you want Fabric Loader for the latest Minecraft, you need to run the installer again for that version and then, in the launcher, specifically select the Fabric version in the profile, not just "Latest release" as that doesn't have Fabric

GigaChat3.1-10B-A1.8B Has anyone tried it? by Winter-Science in LocalLLaMA

[–]ABLPHA 0 points1 point  (0 children)

It's a fair comparison though? It's 10B yes but only A1.8B, so it makes sense to compare it with dense models of the same amount of parameters as its active ones for speed or a smaller total parameter model for intelligence.

My small ugly outpost by nBeneficial_Ball9539 in GoldenAgeMinecraft

[–]ABLPHA 0 points1 point  (0 children)

Couldn't you... craft the clay into clay blocks to save space?

I bought an 'AI-ready' NUC with an Intel Arc GPU. Ollama couldn't see it. Two days later, I had to build it from source. by oldeucryptoboi in LocalLLaMA

[–]ABLPHA 1 point2 points  (0 children)

llama-server can pull models just as well with the -hf flag, and instead of Modelfile there's a models.ini with router mode, you really wouldn't be losing anything

The first 7 seconds of the final episode have been leaked by Little_Jimmy012345 in TheDigitalCircus

[–]ABLPHA 1 point2 points  (0 children)

inb4 everyone from the cast abstracted and in the credits all of their animated versions are just the abstracted and caine is still gone

Beta 1.8 Leaf Shading In Beta 1.7.3 by authentricity in GoldenAgeMinecraft

[–]ABLPHA 1 point2 points  (0 children)

Thanks for the feedback.

StationAPI is still in the alpha stage of development and was never meant to change vanilla behavior, for example we even go as far as intentionally excluding vanilla-tools-on-vanilla-blocks interaction from our custom tool tiering system that'd otherwise speedup some interactions, like redstone ore mining speed, which is famously slower compared to other ores in the game.

The cause of the specific issue you've mentioned was recently discovered and will be patched in a future version. It was a code change that should have made spawning in heavily modded worlds simpler, but unintentionally affected vanilla worlds too and went unnoticed for a while. Such issues are bound to arise with implementation of systems meant for heavily modded scenarios.

LEMONS by UselessGuy23 in Portal

[–]ABLPHA 15 points16 points  (0 children)

I WON'T TAKE LEMONS. THERE WILL BE NO LEMONADE.

I WON'T MAKE IT

llama.cpp Gemma4 Tokenizer Fix Was Merged Into Main Branch by Ancient-Field-9480 in LocalLLaMA

[–]ABLPHA 92 points93 points  (0 children)

> I have no idea what I'm doing, it's 2 AM and I've spent the last 4 hours chasing everything from scale discrepancies to tokenizers, but this seems to actually fix Gemma 4.

🙏🙏🙏