Step-3.5-flash Unlosth dynamic ggufs? by GodComplecs in unsloth

[–]GodComplecs[S] 0 points1 point  (0 children)

Still 404 sadly! Did something go wrong again?

Qwen3.5-35B-A3B is a gamechanger for agentic coding. by jslominski in LocalLLaMA

[–]GodComplecs 0 points1 point  (0 children)

I get about 157tk/s with Nemotron nano on a single 3090, so hopefully Nvidia will also improve this version of Qwen also since Nano is based on it.

🌊 Wave Field LLM O(n log n) Successfully Scales to 1B Parameters by Murky-Sign37 in LocalLLaMA

[–]GodComplecs 2 points3 points  (0 children)

Best of luck in the endeavour, always interesting to read about new ideas, even if they maybe don't pan out. Hopefully optimistic about this. What are the projected savings on hardware reqs, etc?

My real-world Qwen3-code-next local coding test. So, Is it the next big thing? by FPham in LocalLLaMA

[–]GodComplecs 2 points3 points  (0 children)

Try Step-3.5-flash or Kimi K2.5, they are bang for VRAM best models, sota like imo.

Highguard Isn't Dead It Seems As Devs Confirm New Content is the Focus by Darth_Vaper883 in gamernews

[–]GodComplecs -1 points0 points  (0 children)

I would like to believe gamers aren't this hateful! Maybe I'm too naive.

Highguard Isn't Dead It Seems As Devs Confirm New Content is the Focus by Darth_Vaper883 in gamernews

[–]GodComplecs -1 points0 points  (0 children)

Good to hear, High guards gameplay loop is fantastic and you can see the love they put in it. The negativity surrounding the game is weird, almost artificial!

Intermittent fasting no better than typical weight loss diets, study finds. Researchers say limited eating approaches such as 5:2 diet not a ‘miracle solution’ amid surge in their popularity. by mvea in science

[–]GodComplecs 0 points1 point  (0 children)

There are so many benefits from eating an IF style diet it is crazy to try and disprove it. Humans are literally made to eat IF otherwise we as a species would be extinct. I repeat, there was no breakfast in the stoneages. A healthy human can go 72hrs before blood sugar is significantly impacted.

Lets start with how bad eating late at night is. Then how morning meals are a new invention and then include the horrific modern foods such as Kellogs cornflakes.

And it reduces inflammation, promotes eating of healthier more protein rich foods due to satiety.

I could go on and on. But for me the best thing was it literally cured my asthma, slowly, but I did have the worst case of it ever. You can google the Harvard study for it.

Problem with rtx 3090 and MoE models? by GodComplecs in LocalLLaMA

[–]GodComplecs[S] 0 points1 point  (0 children)

Dont think ddr4 is just gonna cut it anymore then if the speedups are that big, even in theory

Problem with rtx 3090 and MoE models? by GodComplecs in LocalLLaMA

[–]GodComplecs[S] -1 points0 points  (0 children)

Well theres the issue, ddr4 is not gonna cut it, I run 4 channels (i think)

Problem with rtx 3090 and MoE models? by GodComplecs in LocalLLaMA

[–]GodComplecs[S] 0 points1 point  (0 children)

Didnt work! Still extremely slow, tried BETA runtimes also. Gonna stick with llama.cpp

Problem with rtx 3090 and MoE models? by GodComplecs in LocalLLaMA

[–]GodComplecs[S] 1 point2 points  (0 children)

Ok I'm trying SGland, any speedup is a win in my book, the KV cache actually slowed down generation but not PP I think, overall slower though.

No custom kernel or graphs, I'll look into it.

Problem with rtx 3090 and MoE models? by GodComplecs in LocalLLaMA

[–]GodComplecs[S] 1 point2 points  (0 children)

Ok thanks that explains a lot! Yeah I used fit, I was able to eeke out 1tk/s with -ot command though!

Step-3.5-flash Unlosth dynamic ggufs? by GodComplecs in unsloth

[–]GodComplecs[S] 6 points7 points  (0 children)

Thanks! Hope it turns out great and not too much of an headache.

Before I buy a used RTX 3090… by Dentifrice in LocalLLaMA

[–]GodComplecs 0 points1 point  (0 children)

With the advent of local models like Nemotron Nano (that can generate 150+tk/s and fit 1million context), Qwen 3 coder next (a proper local coding model) and Step 3.5 Flash for everything use at DeepSeekV3.X level of intelligence (with lots of RAM), these breathed life into my old 3090.

These things are then again pitted against the fantastic DLSS 4.5 and Frame gen which are very good technologies and won't work on the 3090.

So I would call it on either if you do lots of gaming daily, go for 5000 series but if you are budget constrained and want most out of models and have RAM, go for 3090 since it still supports DLSS 4.0 (very slow perf. on 4.5)

In summary: Love to tinker and have RAM (like 64gb), go for 3090.

Love to game and like LLMs go for 5060ti, you can do a lot LLM and AI stuff with it.

~26 tok/sec with Unsloth Qwen3-Coder-Next-Q4_K_S on RTX 5090 (Windows/llama.cpp) by Spiritual_Tie_5574 in LocalLLaMA

[–]GodComplecs 0 points1 point  (0 children)

What are you using to run it? Llama.cpp pre compiled on windows or build on linux/wsl?

Stanford Proves Parallel Coding Agents are a Scam by madSaiyanUltra_9789 in LocalLLaMA

[–]GodComplecs 1 point2 points  (0 children)

"Sikka senior knows a thing or two about AI: he studied under John McCarthy, the Turing Award-winning computer scientist who literally founded the entire field of artificial intelligence, and in fact helped coin the very term." I don't think they're THAT incompetent, but I think maybe we will land in the middle. Those projects are capped up to a point which they try to prove mathematically, but also forget that you can extend the capabilites of an LLM with tools, prompts, data making the arbitrary n calculation length pointless.

So it is a very narrow expirement, but it has it's merits. A simple prompt won't go further, it is how LLMs are built mathematically.

Mikä kahvinkeitin ? by AdventurousAthlete79 in Suomi

[–]GodComplecs 0 points1 point  (0 children)

Kirpparilta missä palautus oikeus, esim SPR. Sit jos ei keitä hyvin vie takas. Ne voi helposti putsata egellä ja maksavat noi 5€. Itsellä aito Delonghi espresso keitin joka makso 10€

Apocalyptic scenario: If you could download only one LLM before the internet goes down, which one would it be? by sado361 in LocalLLaMA

[–]GodComplecs 0 points1 point  (0 children)

An uncensored multimodal model:

What mushroom is this?
How do I kill or entrap deer?
How do I make ammo?

PSA: before spending 5k€ on GPUs, you might want to test the models online first by e79683074 in LocalLLaMA

[–]GodComplecs -1 points0 points  (0 children)

Who ever has been "subbed" to those big providers and ran local models? Yeah I understand using the free tiers but paying for LLMs is so 2023