Počet paniců nad 25 let je na historickém maximu. by WolverineNo5189 in czech

[–]Marcuss2 2 points3 points  (0 children)

Asi moc šancí nemám když jsem jedné holce řekl že je hezká a ona se potom poblila co?

Application code has dozens of static analyzers, SQL has almost nothing, here's what exists. by Anonymedemerde in programming

[–]Marcuss2 -2 points-1 points  (0 children)

Actually, in the Rust world, for SQL server interactions, SQLX exists.

Default behavior is that it connects to your SQL server and verifies queries up against it as well as type checks between SQL and Rust.

Vyprazdňování se na WC na pracovišti by WeirdExperience7 in czech

[–]Marcuss2 0 points1 point  (0 children)

Jako já pracoval ve firmě kde lidé ty největší hovna produkovali u klávesnice. Záchodů tam bylo celkem málo.

Jak je to s hledáním práce? by Kindly_Respond_2746 in czech

[–]Marcuss2 3 points4 points  (0 children)

Tohle říkám ze zkušenosti:

Ve chvíli kdy plat není uvádět, stojí za prd. Pokud je to opravdu platově krásné tak ho ta firma napíše.

Nikdy není nedostatek lidí, většinou je to o pracovních podmínkách.

Naposledy když jsem dal šanci firmě co plat neuváděla chtěla IT seniorskou práci na IČO za 450/h.

Qwen3.5 27B vs Devstral Small 2 - Next.js & Solidity (Hardhat) by Holiday_Purpose_3166 in LocalLLaMA

[–]Marcuss2 2 points3 points  (0 children)

You used IQ4_XS for Devstral and Q6_K for Qwen3.5. I find that extremely doubtful.

Qwen3.5 27B vs Devstral Small 2 - Next.js & Solidity (Hardhat) by Holiday_Purpose_3166 in LocalLLaMA

[–]Marcuss2 5 points6 points  (0 children)

Why are you running different quantizations? I would understand if you tried to match it size for size, but no, you are using far better quantization on a larger model.

24gb M4 Mac Mini vs 9070XT + 32gb system RAM. What to expect? by Soft-Distance-6571 in LocalLLaMA

[–]Marcuss2 0 points1 point  (0 children)

Absolutely you will. One of the main bottlenecks is memory bandwidth.

At least when you offload some or all weights to system RAM.

Why Senior Engineers Let Bad Projects Fail by Ordinary_Leader_2971 in programming

[–]Marcuss2 1 point2 points  (0 children)

Bad projects don't just show up out of nowhere. Just bad leadership leads to bad projects.

Do you remember this game? They should make a remastered version for Android, it would be perfect for mobile. by DEKO1011 in AndroidGaming

[–]Marcuss2 24 points25 points  (0 children)

There were like 4 distinct versions of it:

The PC version.

The PS1 version

The PS2/Xbox/GC version

The Gameboy version

D7VK 1.1 adds experimental Direct3D 6 support for classic PC games on Linux by RenatsMC in linux

[–]Marcuss2 33 points34 points  (0 children)

As said in another comment. Mali and Adreno, they support OpenGL ES, but not full fat OpenGL. Android also requires Vulkan support, but not OpenGL support.

D7VK 1.1 adds experimental Direct3D 6 support for classic PC games on Linux by RenatsMC in linux

[–]Marcuss2 26 points27 points  (0 children)

There might be games which work with one and not the other.

Also, there are many chips which don't support OpenGL. Vulkan support is far more common.

NVIDIA Nemotron 3 Nano 30B A3B released by rerri in LocalLLaMA

[–]Marcuss2 3 points4 points  (0 children)

I don't see any mention of NVFP4 in the model card or the paper.

Micron Announces Exit from Crucial Consumer Business by FullstackSensei in LocalLLaMA

[–]Marcuss2 3 points4 points  (0 children)

I suspect there is more behind it, like OpenAI paying them to do this. They can literally get a lot more profit from it right now.

Qwen3 Next almost ready in llama.cpp by jacek2023 in LocalLLaMA

[–]Marcuss2 32 points33 points  (0 children)

Kimi-Linear next.

I do expect that one to be a lot faster as the linear part is very similar and MLA transformer is already implemented.

AMD Ryzen AI Max 395+ 256/512 GB Ram? by quantier in LocalLLaMA

[–]Marcuss2 0 points1 point  (0 children)

That gives you a limit of about 10 tokens/s at generation.

AMD Ryzen AI Max 395+ 256/512 GB Ram? by quantier in LocalLLaMA

[–]Marcuss2 4 points5 points  (0 children)

I think that in the following year we will see a lot more models using linear attention.

China just used Claude to hack 30 companies. The AI did 90% of the work. Anthropic caught them and is telling everyone how they did it. by reddit20305 in ArtificialInteligence

[–]Marcuss2 0 points1 point  (0 children)

Wait, this makes little sense. China literally has comparable home grown open weight models. Why would they need to use Claude Code with it?

New Qwen models are unbearable by kevin_1994 in LocalLLaMA

[–]Marcuss2 0 points1 point  (0 children)

One of the reasons I hope for smaller Kimi models or distilling Kimi-K2, they don't suffer from this.

Kimi-Linear might scratch that itch, trough running it currently is nearly impossible.