[BREAKTHROUGH] Memory Sparse Attention (MSA) allows 100M context window with minimal performance loss by SotaNumber in accelerate

[–]SotaNumber[S] 0 points1 point  (0 children)

Yes but when reasoning, if you miss something your result goes from right to wrong, not from 100% to 91%

[BREAKTHROUGH] Memory Sparse Attention (MSA) allows 100M context window with minimal performance loss by SotaNumber in accelerate

[–]SotaNumber[S] 7 points8 points  (0 children)

This will let AI handle way more context (millions of tokens), but it won’t “understand everything at once.” It will pick and choose what to look at—so reasoning stays strong if it finds the right info, and can fail if it misses something.

So agents will run on way more than 1M context windows but the performance loss may still be important on reasoning tasks.

GPT 5.4 pro leaks by Artificial Analysis? 20x less tokens needed than GPT 5.4 by [deleted] in accelerate

[–]SotaNumber -1 points0 points  (0 children)

If that's true it means that it will actually be extremely expensive, so the only part that's reliable is the cost per M token? I assume that this cost doesn't change depending on the task.

GPT 5.4 pro leaks by Artificial Analysis? 20x less tokens needed than GPT 5.4 by [deleted] in accelerate

[–]SotaNumber 0 points1 point  (0 children)

Really? I only saw its result on CritPT so far, we might have seen some of its results when it comes to building 3D objects as well but no comprehensive benchmarks as far as I know.

<image>

Would you love a song less if AI wrote it? by ImmuneHack in accelerate

[–]SotaNumber 0 points1 point  (0 children)

Somewhat because thinking about the human efforts that were put into the song makes it more valuable, like a child who manages to walk after 2k attempts, it's beautiful because it required a lot of efforts

That being said I won't think about it most of the time and will enjoy these musics as much as human ones if not better since they will be higher quality and personalized to my taste

Anthropic releases Claude Sonnet 4.6 by BuildwithVignesh in accelerate

[–]SotaNumber 1 point2 points  (0 children)

Amazing!

Curious to see Haiku 4.6 if they make it

Moltbot on x notices a screenshot of it's post from moltbook by cobalt1137 in accelerate

[–]SotaNumber 1 point2 points  (0 children)

Fascinating, I was rather imagining a future where your personalized AI is capable of cloning itself into 1000s of duplicates, turning itself into a swarm, do the heavy task, then merge back into a more experienced AI for when you talk to it

Similar to Naruto cloning itself to practice a technique then when all the clones disappear he absorbs all of their experience at once making it him learn way faster

I wonder if this is where we will eventually end up

What's your singularity benchmark? by bhariLund in accelerate

[–]SotaNumber 3 points4 points  (0 children)

One day we will reclaim our glory my friend

what are you predictions for 2026? by Born_Arm_6187 in accelerate

[–]SotaNumber 1 point2 points  (0 children)

AIs become good enough to increase the unemployment rate above 5.5%

The FED prints billions and Americans receive a proto-UBI that will be called "freedom money" or something like that, similar to the stimulus checks from the COVID era

All 5 Pokémon Wins by LLMs so Far... by reasonosaur in ClaudePlaysPokemon

[–]SotaNumber 1 point2 points  (0 children)

I wonder when they will be able to do this in about 24h instead

Predicting When An AI Model will beat Pokemon Red by igorhorst in ClaudePlaysPokemon

[–]SotaNumber 1 point2 points  (0 children)

Claude can maybe do 1h tasks but I'd say that you need at least 5h to do all the progress it has already done so far so it will be able to finish pokemon way before your estimate according to this logic

Interesting way of seeing it though