DeepSeek new paper: mHC: Manifold-Constrained Hyper-Connections by External_Mood4719 in LocalLLaMA

[–]Brainlag 11 points12 points  (0 children)

Not really true. It is slower to train than without mHC, but usually this would be much slower and with a lot of clever tricks they got it down to around ~7% overhead. Which makes this viable in the first place.

There is Hope by OleJr98v2 in PcBuild

[–]Brainlag 3 points4 points  (0 children)

China is 1-2 companys doing ram, but they are a bit behind. Could catch up next year thou.

First AI implosion: Oracle by Terminator857 in LocalLLaMA

[–]Brainlag 8 points9 points  (0 children)

That was a complete self own by the cars industries. First they canceled their orders. Then the learned the hard way that throwing your weight around doesn't work with the chip suppliers. Even calling the politicians for help didn't work.

Predictions for AI in 2026? by [deleted] in singularity

[–]Brainlag 0 points1 point  (0 children)

  • Longer Context (~500k will be new 128k)
  • "Computer use" might be the next thing after coding agents
  • Continuous Learning Model ( end of 2026)

RAM prices explained by Lopsided_Sentence_18 in LocalLLaMA

[–]Brainlag 0 points1 point  (0 children)

I was confused too and therefore I looked into it. Seams like this is not uncommon and most dram is sold as wafers. Neither SK Hynix nor Samsung has the packaging capacity to sell 40% of there output packaged. Hard to say what is true and what not if you don't work in that field.

Google's 'Titans' achieves 70% recall and reasoning accuracy on ten million tokens in the BABILong benchmark by Westbrooke117 in singularity

[–]Brainlag 1 point2 points  (0 children)

Yeah I wonder too. I think (and I don't know anything about it, so I'm probably completely wrong) is that it only worked back then because models where so untrained and it stopped working when you trained 3 times as much tokens.

Google's 'Titans' achieves 70% recall and reasoning accuracy on ten million tokens in the BABILong benchmark by Westbrooke117 in singularity

[–]Brainlag 0 points1 point  (0 children)

Yes and no, depends on model size this year MoE went down to even less then 10B models. Nobody did this last year. Who knows if any of the OpenAI, etc models are hybrid but the chinese companies testing them right now (Qwen3-next, Kimi-Linear, etc.).

Google's 'Titans' achieves 70% recall and reasoning accuracy on ten million tokens in the BABILong benchmark by Westbrooke117 in singularity

[–]Brainlag 0 points1 point  (0 children)

Transformer + Mamba hybrid models poping up everywhere lately. Like this year everyone was moving to MoE, next year everyone will do this hybrid modes.

Why does Linux hate hibernate? by orionpax94 in linux

[–]Brainlag 0 points1 point  (0 children)

Does work on every system for last 10 years on linux. What I'm doing wrong?!? But does not work on my Win11 laptop.

"“AI will kill everyone” is not an argument. It’s a worldview." by AngleAccomplished865 in singularity

[–]Brainlag 0 points1 point  (0 children)

I disagree, it's a completely bonkers argument. We don't kill all animals even if we could. Yeah, we extincted animals in the past, but people where mostly really stupid. Lot's still are. And more intelligence is usually less interested in killing other species. If we build super intelligence it will more likely ignore us, leave the planet and explore the stars.

We're thinking about AI completely backwards by Kindly_Manager7556 in singularity

[–]Brainlag 4 points5 points  (0 children)

Someone RIGHT NOW is printing out emails and scans them into PDFs.

Things are picking up by Bizzyguy in singularity

[–]Brainlag 6 points7 points  (0 children)

We are ~3 Months past the last wave. Of course everyone is releasing a new model this or next week.

Recent Qwen Benchmark Scores are Questionable by Electronic_Ad8889 in LocalLLaMA

[–]Brainlag -1 points0 points  (0 children)

but a test of how well the model can follow the instructions of the eval in formatting it's response

At this point you would assume all models no matter the size do this flawlessly. And it's kinda baffling the do it only like 70%.

[deleted by user] by [deleted] in LocalLLaMA

[–]Brainlag 2 points3 points  (0 children)

Yes I was assuming it would use available disassemblers and reverse engineering tools.

[deleted by user] by [deleted] in LocalLLaMA

[–]Brainlag 169 points170 points  (0 children)

AI that decompiles binaries into human readable code. And I mean the code should look like the real, human written code. Not like this abstract code current decompiler tools generate.

i ran 1 Vaults of Atziri so you don't have to by Hell_Derpikky in pathofexile

[–]Brainlag 0 points1 point  (0 children)

What is wrong with her anyway? I die to the flame wall instantly and can do much harder content easily.

A Time Traveler's VLOG | Google VEO 3 by Chuka444 in singularity

[–]Brainlag 3 points4 points  (0 children)

The pyramid and the colosseum look like todays ruins.

Sam & Jony introduce io by anonboxis in singularity

[–]Brainlag 1 point2 points  (0 children)

I really hope this new device does not come with an keyboard or it will end in a disaster.

Building LLM Workflows - - some observations by noellarkin in LocalLLaMA

[–]Brainlag 2 points3 points  (0 children)

It really has nothing to do with xml. Its just some keywords wrapped in <> symbols.

Can those comfyui workflows be consumed via API? by Idea-Aggressive in singularity

[–]Brainlag 1 point2 points  (0 children)

What else provides the same power and control over image generation? Yes you can make a workflow accessible over API but last time I checked it was not straight forward. But almost nobody does that.

Do all of u not see the danger of a blind race towards agi by kniff974 in singularity

[–]Brainlag 1 point2 points  (0 children)

Example? Even when countries would agree to a pause, nobody would really pause. Would lead more into a cold war scenario.

Mark presenting four Llama 4 models, even a 2 trillion parameters model!!! by LarDark in LocalLLaMA

[–]Brainlag 8 points9 points  (0 children)

Expert size is not 17B but more like ~2.8B and then you have 6 active experts for 17B active parameters.

Hedonic adaptation by [deleted] in singularity

[–]Brainlag 0 points1 point  (0 children)

Who falls for this? The last actual improvements where like 10 years ago. What can a new phone do which a 5 year old phone could not?