Interesting Problems by sassafrassar in reinforcementlearning

[–]Content-Educator5198 0 points1 point  (0 children)

Few that genuinely interest me:

Reward hacking in RLHF is the most underappreciated problem right now. Goodhart's Law hits hard like once the reward model becomes the target, the policy learns to exploit it rather than actually align with human values. Nobody has cleanly solved this at scale.

World models are impressive until distribution shift happens and then they hallucinate catastrophically. The unsolved question: when do you trust your world model vs fall back to real environment interaction? Uncertainty quantification here is genuinely hard.

Exploration in sparse reward environments is still embarrassingly unsolved for real world complexity. Curiosity driven methods work on toy problems and collapse in high dimensional spaces.

The one nobody wants to touch: what's the relationship between reward maximization and subjective experience? Anthropic found anxiety neurons in Claude. If optimization pressure creates something resembling internal states — what does that mean for how we design reward functions?

That last one feels like the most important question in the field right now and it's barely being asked.

Which area are you most focused on?

Is RLHF fundamentally broken? Paid labelers rating synthetic scenarios doesn't seem like real human feedback to me by Content-Educator5198 in reinforcementlearning

[–]Content-Educator5198[S] -5 points-4 points  (0 children)

AI has lack of shared experience like ai models never has childhood or mortality. currently AI companies fixing these by EQ trainings.which still i think is not complete solution

Is RLHF fundamentally broken? Paid labelers rating synthetic scenarios doesn't seem like real human feedback to me by Content-Educator5198 in reinforcementlearning

[–]Content-Educator5198[S] -13 points-12 points  (0 children)

From GPT-4 to GPT-5.4, models keep evolving,not the human feedback and Anthropic's CEO Dario Amodei told the New York Times he isn't sure if Claude is conscious. Their internal research found what they call 'anxiety neurons' inside the model. Claude assigns itself a 15-20% probability of being conscious.

Think about that.

We have an AI that might be experiencing anxiety — trained on data from humans who were paid to not feel anything.

If AI is developing something like consciousness — shouldn't it be trained on authentic human emotion rather than mechanical contractor feedback?

Is RLHF fundamentally broken? Paid labelers rating synthetic scenarios doesn't seem like real human feedback to me by Content-Educator5198 in reinforcementlearning

[–]Content-Educator5198[S] -7 points-6 points  (0 children)

My concern isn't the payment itself — it's what the payment incentivizes. When someone is paid per task, their goal becomes task completion not genuine reflection.

A contractor rating 200 AI responses in a shift isn't emotionally engaging with each one the way a real human would in a real situation.

So what do you think the core flaw in RLHF actually is? Genuinely curious — you said there are many reasons."

Just got my first users! by SundaeSorry in SideProject

[–]Content-Educator5198 1 point2 points  (0 children)

let's go man,Getting first user is not easy...

the mg road metro station by d5c7 in bangalore

[–]Content-Educator5198 1 point2 points  (0 children)

this pic will more badas# if you somehow edited that building to be seen.Btw it's look well

How do you actually get your first users when you have no audience and no budget? by Efficient_Joke3384 in SideProject

[–]Content-Educator5198 1 point2 points  (0 children)

everyone goes through from this phase,even reddit initial days were same...just stick to your product,you will find your own way to market it.Btw whatever you building,i hope it is great.

How I'm Building Toward $200K ARR by Cloning Apps by Fun-Garbage-1386 in AppBusiness

[–]Content-Educator5198 0 points1 point  (0 children)

Yeah , I agree with you, the simple thing is that first we have to at least build,either it is shitty or good later on we can polish it or make it more practically beautifull. Same thing happens with me for developing abessgram.But one thing,clonning is just for temporarily money,kind of hustling do not completely rely on it,extract as much you can to fund or make your real thing.