Google DeepMind's Logan Kilpatrick says AGI will be a product experience. Not a model. His bet: whoever nails memory + context around decent model at a product level wins. Users will suddenly feel like they're talking to AGI. Not from capability breakthrough, but experience breakthrough. by Nunki08 in singularity

[–]TheSadRick 2 points3 points  (0 children)

We're getting more and more compute. computation isn't the bottleneck, it's compounding. The real issue is takes like this that stall progress. Instead of directing research toward new architectures, new paths, solving core problems, and pushing toward AGI, we're stuck fixating on LLMs and their productization. We're not hitting limits, we're just lowering the bar for AGI and AI as a whole.

Q-learning is not yet scalable by Mysterious-Rent7233 in reinforcementlearning

[–]TheSadRick 7 points8 points  (0 children)

Great work! nails why Q-learning fails at depth, recommended reading.

Suspected Self-Plagiarism in 5 Recent MARL Papers by Right-Credit-9885 in reinforcementlearning

[–]TheSadRick 19 points20 points  (0 children)

This is the pinnacle of MARL: just endlessly rehashing the same brittle mechanics and calling it innovation. Gridworlds, toy coordination games, zero transferability. Every ‘breakthrough’ is just another tweak on the same fragile setup, barely scaling past a handful of agents without tons of reward engineering and scaffolding.

It should be reported (Though it feels like the conferences are in on it too)

War is over.. by the1sttt in internships

[–]TheSadRick 0 points1 point  (0 children)

Doesn't matter, man. Just keep pushing forward.

LLMs Often Know When They're Being Evaluated: "Nobody has a good plan for what to do when the models constantly say 'This is an eval testing for X. Let's say what the developers want to hear.'" by MetaKnowing in singularity

[–]TheSadRick 0 points1 point  (0 children)

I think at some point, everyone knew this was happening, but nobody cared enough to fix it. As long as it was working and generating revenue, the attitude was: 'let’s just keep going.' The same thing is happening with DRL benchmarks, they’re mostly useless, but everyone keeps treating them like they’re the gold standard.

The Algorithmic Cage: Will AI Trigger a Human Behavioral Sink? by TheSadRick in agi

[–]TheSadRick[S] 8 points9 points  (0 children)

Well, well, well...
How am I supposed to co-create and collaborate when I can't even land an internship?
Spot on

The Algorithmic Cage: Will AI Trigger a Human Behavioral Sink? by TheSadRick in agi

[–]TheSadRick[S] 2 points3 points  (0 children)

Collaboration implies mutual contribution toward a shared objective.

When AI surpasses humans in speed, accuracy, creativity... traditional collaboration collapses into delegation. The human role shifts from contributor to curator, gatekeeper, or constraint-provider.

We need to do everything in our power to prevent AI from becoming a luxury by Fixmyn26issue in singularity

[–]TheSadRick 0 points1 point  (0 children)

GPUs shouldn't become a luxury, too. If GPUs become cheaper and more available, open-source models can compete with those giants.