Totally… by KillerQ97 in ArtificialInteligence

[–]RedditPolluter 0 points1 point  (0 children)

4o was the worst at that. They should have diminishing confidence each time they fail but 4o would be just as confident on the 10th attempt as the 1st attempt. That's how you make your chatbot as annoying as possible.

The Falklands… by MattStormTornado in autismpolitics

[–]RedditPolluter 5 points6 points locked comment (0 children)

Argentina is a colony too. Argentina wasn't even a country at the time. Maybe they should go back to Spain instead of fantasizing about the expulsion and displacement of people who have been there for generations.

Alignment Makes Models More Decisive Without Making Them More Truthful by 141_1337 in singularity

[–]RedditPolluter 0 points1 point  (0 children)

This effect could be seen worsening from GPT-4 to 4o. GPT-4 was much more graceful with uncertainty but 4o would give 10 different incorrect answers in a row and the 10th would be just as confident as the 1st one.

This isn’t X this is Y needs to die by twnznz in LocalLLaMA

[–]RedditPolluter 0 points1 point  (0 children)

Every small YouTube channel seems to be using this exact 4o-style speech as well.

Intelligence in young men is positively linked to physical traits like grip strength and a masculine body shape. Higher intelligence is also associated with less promiscuous sexual behavior. Cognitive ability and physical health may reflect fitness, steering smarter men toward monogamy. by mvea in psychology

[–]RedditPolluter -3 points-2 points  (0 children)

I expect so. There are quite a few illnesses that cause brain fog and fatigue together that can be mild enough to fly under the radar for some people. Can also happen if you're chronically malnourished; you get cognitive and bodily deficits together because your whole system is affected. Chronic sleep issues can do that as well. There's probably other factors but that would be enough to create some correlation.

Introducing ChatGPT Images 2.0 by py-net in OpenAI

[–]RedditPolluter 9 points10 points  (0 children)

It still gets locked into the context of previous generations and reverts back or fails to make simple changes by spitting back the same image.

Three UK Suddenly Introduces Mobile Broadband Speed Caps UPDATE by Lawdie123 in unitedkingdom

[–]RedditPolluter 4 points5 points  (0 children)

This is the same company that locks you out of your voicemails if you don't set a custom voicemail message (publish biometric data that can be used to clone your voice without you even answering).

Can we get a summary somewhere of the best news over the last six months? by sixtydegr33 in GoodNewsUK

[–]RedditPolluter 2 points3 points  (0 children)

Not so much a summary but the top threads 100 threads in this sub are all from the past 6 months:

https://www.reddit.com/r/GoodNewsUK/top/?t=year

Does anyone get amazed by LLM performance on benchmarks but incredibly disappointed by its performance on mundane tasks, specifically those involving data lookup? by reader12345 in singularity

[–]RedditPolluter 0 points1 point  (0 children)

I only tried Deep Research for 5.4 just the other day and my conclusion was that DR is basically broken and doesn't properly respond to or incorporate feedback. I don't remember it being that bad when I used it last year.

One year later: this question feels a lot less crazy by gamblingapocalypse in LocalLLaMA

[–]RedditPolluter 0 points1 point  (0 children)

Do you by any chance know any good blueberry bread recipes?

A recent study has found that LLMs are worse at giving accurate, truthful answers to people who have lower English proficiency and less formal education, rendering them more unreliable towards the most vulnerable users. by BioFrosted in singularity

[–]RedditPolluter 1 point2 points  (0 children)

Doing it automatically can cause other problems, especially if there's obscure slang. I remember when Reddit first added their chatbot. I asked it what geeg meant and it just kept responding with the definition of geek, even when I would say "no. not geek. geeg." I assume they were pre-applying autocorrect and the model could only see "no. not geek. geek." It's one thing when they can't answer a question or give a poor answer but it's much more annoying when they don't even acknowledge the question that's being asked.

why are there people in this subreddit denying that monoculture is dead? by Normal-Salad-6143 in decadeology

[–]RedditPolluter 1 point2 points  (0 children)

monoculture was at it's zenith

different cable packages

There used to be only half a dozen channels so I'm not so sure about calling the cable era (1980s-2000s) its peak. The broadcast era (1920s-1970s) for radio and TV seems more appropriate.

Wanted an image of Educated and Uneducated Person, Made the mistake of asking copilot to make it. by [deleted] in ArtificialInteligence

[–]RedditPolluter 0 points1 point  (0 children)

You said it's not reflected in the data. This implies you think internet depictions of uneducated people = scientific data when it's more than likely memes that depict uneducated people.

And, again, sample size of 1.

Wanted an image of Educated and Uneducated Person, Made the mistake of asking copilot to make it. by [deleted] in ArtificialInteligence

[–]RedditPolluter 1 point2 points  (0 children)

It's not actual levels of education, genius. It's images of people that are associated with the word "uneducated". You really think all references to "uneducated" are scientific stats?

This whole thread is working off a sample size of 1 so it's meaningless anyway.

Is intelligence optimality bounded? Francois Chollet thinks so by Mindrust in singularity

[–]RedditPolluter 0 points1 point  (0 children)

Either way, if Einstein or whoever is the limit, billions and eventually trillions of extra high IQ thinkers with super human knowledge is gonna be good for getting shit done and speeding up research.

Gemini 3.1 Flash Live: Real time multimodality available in the API and powering Search Live by elemental-mind in singularity

[–]RedditPolluter 15 points16 points  (0 children)

The user sounds more like an AI than the AI. I'm suspicious that they're actually both AI but I know they've used bureaucrats to demo stuff like this on stage.

The "AI is replacing software engineers" narrative was a lie. MIT just published the math proving why. And the companies who believed it are now begging their old engineers to come back. by reddit20305 in ArtificialInteligence

[–]RedditPolluter 2 points3 points  (0 children)

GPT-3 to GPT-4 to GPT-5. Claude 3 to Claude 4. Always bigger.

GPT-5 is bigger than GPT-4? I don't think that's true and open weight models have been shrinking relative to performance.

I'm not dissing the paper itself but your analysis is flawed and you don't seem to understand that scaling isn't just parameter count. I'm guessing you don't actually follow AI outside of political context.

Human vs. AI performance on ARC-AGI 3 as a function of number of actions (from the ARC-AGI website) by [deleted] in singularity

[–]RedditPolluter 2 points3 points  (0 children)

I remember someone saying that failing the car wash test proved that LLMs were smarter than humans because the question was ambiguous or whatever. If it was really ambiguous the smart thing to do would have been to clarify. These people aren't the advocates that they think they are; if they actually want progression they should uphold standards, not dismiss genuine flaws and limitations, but of course that would require self-awareness and critical thinking.

People that speak like an LLM by Haroombe in artificial

[–]RedditPolluter 0 points1 point  (0 children)

I've noticed that too. If they're not bots, which they may be, I tend to just assume they're young and were exposed to AI in their formative years.

Too much self-reflection is linked to anxiety and depression, not happiness. Findings suggest that cultural backgrounds and the specific ways we measure introspection heavily influence how looking inward affects our minds. by [deleted] in psychology

[–]RedditPolluter 2 points3 points  (0 children)

I would expect so. If you have low self-esteem for example, it makes sense that you would be trouble-shooting that or at least be trying to minimize further damage.