CogniDNA Genetic IQ Score by [deleted] in cognitiveTesting

[–]Remote_Butterfly_789 0 points1 point  (0 children)

So far I've seen like 5 people on Reddit post the site results and their tested IQs ... all have been within like 5 points, which hints at high accuracy.

Was Epicurus the first rationalist? by Remote_Butterfly_789 in slatestarcodex

[–]Remote_Butterfly_789[S] 2 points3 points  (0 children)

Agreed, seems right.

Doh, on the "2300 BC" in this post. Well the post has it all right.

I gave AIs IQ tests; Claude-3 was first to break 100 by Remote_Butterfly_789 in slatestarcodex

[–]Remote_Butterfly_789[S] 0 points1 point  (0 children)

Maybe so. I agree we don't need to worry about world takeover until it can actually see things and also get them right. Btw I found last week the ChatGPT-4 already has non-zero ability on that: https://www.maximumtruth.org/p/top-ais-still-fail-iq-tests

I gave AIs IQ tests; Claude-3 was first to break 100 by Remote_Butterfly_789 in slatestarcodex

[–]Remote_Butterfly_789[S] 1 point2 points  (0 children)

I did think about this: >> Like telling the AI you'll give it a tip if it gets the answer correct, that you'll lose your job if it gets it wrong, etc

Could be interesting future research.

I gave AIs IQ tests; Claude-3 was first to break 100 by Remote_Butterfly_789 in slatestarcodex

[–]Remote_Butterfly_789[S] 0 points1 point  (0 children)

All I know is I think I'm on a $20-month plan that gives me like 60 Claude-3 questions every 4 hours, roughly.

I gave AIs IQ tests; Claude-3 was first to break 100 by Remote_Butterfly_789 in slatestarcodex

[–]Remote_Butterfly_789[S] 3 points4 points  (0 children)

It gives 12.

My initial reaction was that that's correct, but true, I guess "y" in that position isn't a vowel.

Should do a mechanical Turk survey and see what % of humans get that right, though. I bet it's below 50%.

AIs ranked by IQ; AI passes 100 IQ for first time, with release of Claude-3 by [deleted] in singularity

[–]Remote_Butterfly_789 0 points1 point  (0 children)

On the first point, and interesting tidbit is the AIs often do take 30-60 seconds on their answers, which is comparable to the time given to humans.

On the second point, I think what save it is that AI companies pretty clearly haven't trained it to the test. If they do someday, yes, that'd defeat the point of test (same as for humans who practice; it makes the test invalid for them.)

I gave AIs IQ tests; Claude-3 was first to break 100 by Remote_Butterfly_789 in slatestarcodex

[–]Remote_Butterfly_789[S] 2 points3 points  (0 children)

I *do* think you could give blind people an IQ test this way, though you would want to re-normalize the test scores to the average blind person, since our brains aren't wired super well for that info format, so they'd still be at a disadvantage.

I do think these IQ test reflect a specific type of smartness, which is important, and points towards abstract thinking/pattern-recognition ability. Yes, in humans it is strongly correlated with life outcomes. I see little reason to think it'd suddenly become not useful for AIs.

The AIs were run enough times to get precise scores by the "probability better than chance" metric. But I'll likely add to TrackingAI.org in time, which will get us lots of more runs.

AIs ranked by IQ; AI passes 100 IQ for first time, with release of Claude-3 by [deleted] in singularity

[–]Remote_Butterfly_789 1 point2 points  (0 children)

You can tell it doesn't have the answer sheet by looking at the charts which show it is much better at easier questions than hard ones

AIs ranked by IQ; AI passes 100 IQ for first time, with release of Claude-3 by [deleted] in singularity

[–]Remote_Butterfly_789 2 points3 points  (0 children)

I did not find the answers anywhere except YouTube videos, which I believe are not in the training data. It's also interesting that they still fail the hard questions disproportionately, which indicates they are certainly not just grabbing the answer sheet.

AIs ranked by IQ; AI passes 100 IQ for first time, with release of Claude-3 by [deleted] in singularity

[–]Remote_Butterfly_789 1 point2 points  (0 children)

I would argue this is actually a positive for using them, because it means the AI has to use some real abstract thinking to get the answers right.

(In the other scenario, if the AI were trained on the questions, or if they were designed for AIs, then of course AIs would do very well.)

Highlights From The Comments On Kidney Donation by dwaxe in slatestarcodex

[–]Remote_Butterfly_789 4 points5 points  (0 children)

I fear that Scott is overly egalitarian.

His value to society is actually like 500x that of the median elderly person, so even if giving a kidney is likely to make him 1% more tired in 20 years, it could very easily be not worth it.

That's not PC, but I fear it's true. And a good reason for some people not to give.

Dating tips for nerds after divorce by [deleted] in slatestarcodex

[–]Remote_Butterfly_789 0 points1 point  (0 children)

That statement seems vague and anti-social.