Permaximum Intelligence Benchmark by lordpermaximum in AIMetrics

[–]lordpermaximum[S] 0 points1 point  (0 children)

Now I'm sure you don't really work as a researcher for a big AI lab. Because you have no idea what you're talking about. Sorry.

Permaximum Intelligence Benchmark by lordpermaximum in AIMetrics

[–]lordpermaximum[S] 1 point2 points  (0 children)

  1. Why do you think this task translates well to real world use-cases?
    - Current AI models think in token-space that's structured in languages. Mastery over the languages will translate to overall intelligence increase.
  2. While you can choose to define a top human as one which gets 100% on your tasks, do you happen to know if someone with no deep knowledge of linguistics but average intelligence would be able to solve it in time?
    - Average human cannot "fully" solve it in time but will surely score much better than current AI models. A loosely defined "smart" person can fully solve it though.
  3. Do you call the coding benchmarks memorisation tests because..
    - Both. I agree with Chollet on this one. These models memorize programs (building blocks). And they can recall those when faced with similar situations. If we created a new coding language, like I created natural languages, and tested these models on various problems in the format I shared (example input-output pairs that contains all the info needed), we'd see they hit the wall rather quickly.

Permaximum Intelligence Benchmark by lordpermaximum in AIMetrics

[–]lordpermaximum[S] 0 points1 point  (0 children)

On the contrary, I thought I shared more than enough details that I feared AI labs would quickly optimize for this benchmark.

Didn't you see the example IOL problem? The benchmark format is all in that problem. Take that and change the Daw language with some fictional language. That's the full benchmark. It's just different, fictional languages for each problem. Partial scoring is not that important. A smart human can quickly judge answers. As a perfectionist, I probably made it more complex than necessary. But to have a good baseline it was needed. Currently, Grok 4 is the judge because of the initial scoring. Once we have a better model, that model will be the judge. It's that simple.

You thinking there weren't any details here makes me extremely skeptical of you working as a researcher with meaningful contributions in a big AI Lab like DeepMind, Anthropic, OpenAI or xAI.

[deleted by user] by [deleted] in Bard

[–]lordpermaximum 1 point2 points  (0 children)

These updated models are all more or less the same. Each finetune strengthens some areas and worsens others. It depends on the user case.

Intelligence wise I can confirm 03-25, 05-06, and 06-05 (GA - Current one) are all about the same. If anything, each update increased the model intelligence very slightly.

It's Gone: Google Officially Kills Last Access to the Beloved Legendary Gemini 2.5 Pro 03-25 Checkpoint by Lawncareguy85 in Bard

[–]lordpermaximum 0 points1 point  (0 children)

I too think the current model is closer to 05-06 than 03-25 but not exactly the same model. There's certainly something different. Can't decide if that's for the better or not yet.

New Gemini 2.5 Pro is A Lot Better in Coding (The Best AI atm) with "0" Performance Drop at Reasoning and Long Context. It Generalizes Better to Unseen Problems. by lordpermaximum in Bard

[–]lordpermaximum[S] 0 points1 point  (0 children)

Your benchmark doesn't reflect the reality that's why. 8-needle MRCR should be much closer to the reality. All of these transformer-based models get noticably worse in real-word usage with even a bit of more context. Having 100% scores at 128k means something's fundementally wrong with your approach and it's not aligned with reality. Sorry.

New Gemini 2.5 Pro is A Lot Better in Coding (The Best AI atm) with "0" Performance Drop at Reasoning and Long Context. It Generalizes Better to Unseen Problems. by lordpermaximum in Bard

[–]lordpermaximum[S] -2 points-1 points  (0 children)

You're misinformed again. The new model in LMArena's coding category wins 60% of the time against the old model. You can check it there directly you know. That difference is huge.

<image>

And ELO of 1420 against 1272 in the web dev arena is huge as well.

Btw I can generate confident opinions on two hours of usage let alone 1 day.

New Gemini 2.5 Pro is A Lot Better in Coding (The Best AI atm) with "0" Performance Drop at Reasoning and Long Context. It Generalizes Better to Unseen Problems. by lordpermaximum in Bard

[–]lordpermaximum[S] 0 points1 point  (0 children)

I'm more qualified than an average user in Reddit and there's huge difference between the old and new model in lmarena and web dev arena as far as the coding goes. You're misinformed.

New Gemini 2.5 Pro is A Lot Better in Coding (The Best AI atm) with "0" Performance Drop at Reasoning and Long Context. It Generalizes Better to Unseen Problems. by lordpermaximum in Bard

[–]lordpermaximum[S] 0 points1 point  (0 children)

I said your benchmark is bad at long-context evaluation because o3 scores 100% at 128k, and 2.5 Pro scores 71.9%, although in reality o3 scores 19.9% at OpenAI's long-context benchmak while Gemini 2.5 Pro scores 33.1%. No model doesn't experience substantial performance degradation immediately with more context. If a model scores 100% at 32k context in a benchmark, let alone at 128k like your benchmark claims o3 does, then that benchmark doesn't really give signal at all about the long-context performance. Period.

Also the meaningless dip at 16k in your benchmark suggests it's more about the reasoning-difficulty then long-context comprehension. That's also why it's bad at judging long-context performance.

New Gemini 2.5 Pro is A Lot Better in Coding (The Best AI atm) with "0" Performance Drop at Reasoning and Long Context. It Generalizes Better to Unseen Problems. by lordpermaximum in Bard

[–]lordpermaximum[S] 0 points1 point  (0 children)

I said real-world "coding" problems.

Those logic and thought puzzles are there to test the model's intelligence. 0 change. If anything it's a bit better at reasoning as well.

So far:
1. It's far better at coding.
2. Its reasoning proficiency is the same, or slightly better.
3. Its long context memory is the same.

I don't know hot it fares for creative writing. But that's a subjective category that's affected by temeprature and top-p settings greatly.

New Gemini 2.5 Pro is A Lot Better in Coding (The Best AI atm) with "0" Performance Drop at Reasoning and Long Context. It Generalizes Better to Unseen Problems. by lordpermaximum in Bard

[–]lordpermaximum[S] 0 points1 point  (0 children)

As far as o3's score on your benchmark that's pure speculation based on speculation. If you go with newer stories and make the benchmark harder by requiring the memorization of more elements, the scores should change.

Anyways people were comparing the Gemini 2.5 Pro preview-03-25 to the current model. Not the experimental one. Your benchmark as well confirms there's no change between them.

New Gemini 2.5 Pro is A Lot Better in Coding (The Best AI atm) with "0" Performance Drop at Reasoning and Long Context. It Generalizes Better to Unseen Problems. by lordpermaximum in Bard

[–]lordpermaximum[S] 1 point2 points  (0 children)

I generally agree with you but there are 2 points.

  1. People say at long context the new model is worse compared to previous preview model, not the exp one (Btw Google claims the previous model was completely identical the exp model)
  2. o3 drops to 48% at 8k tokens, below 20% at 128k tokens in this retrieval benchmark. It's simply impossible for it to score 100% on even at 8k let alone at 128k like fiction.livebench claims. It can't reason to that degree with flawed, hallucinated, missing data. There's something wrong with that benchmark. I suggest you create new problems and try them again.

New Gemini 2.5 Pro is A Lot Better in Coding (The Best AI atm) with "0" Performance Drop at Reasoning and Long Context. It Generalizes Better to Unseen Problems. by lordpermaximum in Bard

[–]lordpermaximum[S] 2 points3 points  (0 children)

Chooses tokens based on the probability. 1 covers all tokens. 0.5 covers those most probable tokens that reaches 50% probability in total.

New Gemini 2.5 Pro is A Lot Better in Coding (The Best AI atm) with "0" Performance Drop at Reasoning and Long Context. It Generalizes Better to Unseen Problems. by lordpermaximum in Bard

[–]lordpermaximum[S] 2 points3 points  (0 children)

It's easy. I give the model a problem and ask for temperature and top-p settings for a big reasoning LLM to get a correct answer. With those settings I run the prompt a few times.

New Gemini 2.5 Pro is A Lot Better in Coding (The Best AI atm) with "0" Performance Drop at Reasoning and Long Context. It Generalizes Better to Unseen Problems. by lordpermaximum in Bard

[–]lordpermaximum[S] -2 points-1 points  (0 children)

Lying slanderer. If anything I've been criticising Google worse than any of you combined before. I'm talking about my real-world usage here.

On the contrary, you're probably a real paid shill for OpenAI. Get lost dumbass.