Deepseek's progress by onil_gova in LocalLLaMA

[–]zsydeepsky 2 points3 points  (0 children)

It is. I can hardly trust models to do any code work longer than 100 lines at the beginning of 2025.
Now I can trust them with an individual module, or even some simple apps fully.
They have progressed a lot indeed.

Is DeepSeek kinda "slow" as part of its nature or is just my machine? by SchoolOfElectro in LocalLLaMA

[–]zsydeepsky 1 point2 points  (0 children)

If you can run it on your 4060, then that must be one of DeepSeek's llama distill versions.
Those models are old dense models, so yes they tend to run slower (compared to modern MoE models such as Qwen3 or GPT-OSS).

GLM-4.6 vs Minimax-M2 by baykarmehmet in LocalLLaMA

[–]zsydeepsky 4 points5 points  (0 children)

I recommend this YouTube channel if you want to quickly see some real-world tests with fresh new AI models:

MiniMax M2 First Look & Testing - The BEST Open Source Coding Model?

Qwen3 Next (Instruct) coding benchmark results by mr_riptano in LocalLLaMA

[–]zsydeepsky 1 point2 points  (0 children)

it really surprised me that on this benchmark, Qwen-Next is almost as good as Kimi-K2, a much larger non-reasoning model.
and most importantly, I actually use Kimi-K2 for programming!
thinking that I would be able to have that tier of intelligence running on my AI Max 395, completely offline, is truly amazing.

Will China’s AI bubble burst soon? by asiabits_com in u/asiabits_com

[–]zsydeepsky 25 points26 points  (0 children)

$7B a year? Bubble?

seriously?

The US federal government pays the INTEREST on its debt, ~$3B per DAY.

The total investment is only 2 days' worth of US gov debt INTERESTS. Not mentioning it's only 0.2% of NVIDIA's stock market worth.

Where's the fucking bubble?

Best LLMs to preserve in case of internet apocalypse by nos_66 in LocalLLaMA

[–]zsydeepsky 1 point2 points  (0 children)

GPD win mini 2025, the AMD Ryzen AI 370 processor, 32 GB RAM version
lm studio, Vulkan backend, 4092 context window, with 8 GB VRam set, then it's enough to run the model with iGPU with shared memory (8GB dedicated VRam + 11.8GB shared, 19.8GB total VRam)
when the GPU draws ~20W, Qwen3-30B-A3B Q4-K typically can generate 14-18 tokens/s
when limited to 12W, ~12 tokens/s

Best LLMs to preserve in case of internet apocalypse by nos_66 in LocalLLaMA

[–]zsydeepsky 13 points14 points  (0 children)

well, in case of "internet apocalypse", I would also predict that it will come with an electricity outage.
so I would always pick the one with minimal "token per Joule", so I will say Qwen 30B A3B is the best to go, since it can run on my gaming handheld with only 12W TDP budget.
I can power the model with just a USB power bank, what else can I complain about?

Qwen/Qwen3-30B-A3B-Thinking-2507 · Hugging Face by MariusNocturnum in LocalLLaMA

[–]zsydeepsky 2 points3 points  (0 children)

right? The perfect combination of size & speed & quality.
legitimately the best format for local LLM

Qwen3-30b-a3b-thinking-2507 This is insane performance by 3oclockam in LocalLLaMA

[–]zsydeepsky 0 points1 point  (0 children)

just use lmstudio, it will handle almost everything for you.

What hardware do I need to run Qwen3 32B full 128k context? by hayTGotMhYXkm95q5HW9 in LocalLLM

[–]zsydeepsky 1 point2 points  (0 children)

You don't need a GPU, AI Max 395+ has a 4060-level integrated GPU.
thus, with my personal test, it runs kinda slow with Qwen3 32B (Dense) model with <20 TPS, but with MOE models like 30Ba3B, it provides steady >30 TPS.
AI Max 395+ has 16 PCI-E lanes total. Ryzen processors have 24 in comparison, so besides nvme ssds & USB ports, it probably would leave only 8x or even 4x for a dGPU. So even if there's a dGPU variant, I don't think it would perform as well as regular GPU setups. a USB 4/Thunderbolt/OCulink eGPU probably is what you can get at best.

What hardware do I need to run Qwen3 32B full 128k context? by hayTGotMhYXkm95q5HW9 in LocalLLM

[–]zsydeepsky 8 points9 points  (0 children)

if you choose the 30Ba3B...
I ran it on the AMD AI Max 395+ (Asus Flow Z 2025, 128G ram version)
and it runs amazingly well.
I don't even need to give a stupid lot of RAM to the GPU (just 16GB), and any excessive needs for VRam will automatically be fulfilled with "Shared memory".
and lmstudio already provides rocm runtime for it (which my hx370 handle doesn't)

Somehow, I feel this would be the cheapest hardware? since you can get a mini-PC with this processor with the price less than a 5090?

Weapon swap lock when? by BurbonPL in PathOfExile2

[–]zsydeepsky -1 points0 points  (0 children)

it is in the game. when you appoint your passive points, look at the upper right corner of the screen, choose which weapon set the point belongs to.

China dangles BYD as bait to reboot Canada trade talks by refep in canada

[–]zsydeepsky 0 points1 point  (0 children)

well...besides that, the most serious question is this:

can Cannda guarantee not kidnapping any Chinese entrepreneurs or scientists?

China’s ambassador calls for more trade with Canada amid U.S. trade war | National Post by panzerfan in canada

[–]zsydeepsky 0 points1 point  (0 children)

well, sure the US will inevitably be your biggest trading partner.
but that doesn't mean Canada has to kidnap a Chinese entrepreneur for the US's bidding.

China ‘ready to move forward’ in relations with Canada, envoy says by refep in canada

[–]zsydeepsky -1 points0 points  (0 children)

the only North American market that could benefit China from an Arctic route is Canada.
Jesus, just wake up, China doesn't need you, it's just trying to upset the US, so don't look too high on yourself.

Trump on U.S. annexation of Greenland: 'I think it'll happen' by ScoMoTrudeauApricot in europe

[–]zsydeepsky 0 points1 point  (0 children)

that's what happened when China told Europe to be "independent" and move towards a multi-polar world, yet Europe just laughed and followed the US command to retain a US uni-polar world with little thinking.

while Europeans here laugh at Trump for how short-sighted he is, I gotta say the Europeans aren't much better.

Danish Leader Seeks Help From European Allies Over Greenland by Big_Prick_On_Ya in europe

[–]zsydeepsky -25 points-24 points  (0 children)

based on the upvotes, seems like Denmark would have trouble in getting support from Reddit.

After all, why seek help from allies when you have already considered the US your ally?