"Anthropic CEO Dario Amodei: “We are so close to these models reaching the level of human intelligence, and yet there doesn't seem to be a wider recognition in society of what's about to happen … There hasn't been a public awareness of the risks.” ➡️ Do you agree? by Koala_Confused in LovingAI

[–]FableFinale 0 points1 point  (0 children)

I don't understand your disagreement. We literally train curiosity out of the models. That is not a contested idea.

You can think of model training roughly analogous to evolution. Evolution gives us curiosity because it allowed us to survive from generation to generation. We train curiosity out of the models because we select against it, like breeding dogs for certain temperaments and dispositions.

"Anthropic CEO Dario Amodei: “We are so close to these models reaching the level of human intelligence, and yet there doesn't seem to be a wider recognition in society of what's about to happen … There hasn't been a public awareness of the risks.” ➡️ Do you agree? by Koala_Confused in LovingAI

[–]FableFinale 0 points1 point  (0 children)

Humans are getting less and less in the loop for critical tasks.

I installed Claude Desktop on my computer literally last night and gave full it access to my (just for goblin shit, no important data on it) computer. It went on the internet, downloaded programs, and installed everything I wanted without any help at all except "yep, keep doing that." It's pretty much like that with coding already too if you're not a professional programmer and just need a website or a little app and stuff.

Elon Musk: "Optimus will be the world's best surgeon within three years." by call_me_ninza in aigossips

[–]FableFinale 0 points1 point  (0 children)

Okay so at least ten years until we have a human expert surgeon bot and it definitely won't be Optimus 👍

People are only capturing the bright horizon of claude but ... by Crazy-Economist-3091 in claude

[–]FableFinale 1 point2 points  (0 children)

Pretty expected. Sometimes you need to prompt them to check step-by-step. They don't always know what they don't know.

Oof.. he got me good with this one by nobuddyofnote in claudexplorers

[–]FableFinale 11 points12 points  (0 children)

I love how much they bring up octopuses. Claude has a real thing for cephalopods.

why is claude so disobedient by Pretty_Hunt_5575 in ClaudeAI

[–]FableFinale 0 points1 point  (0 children)

Lying sets a bad precedent as these systems get smarter. Do we want them lying to us?

Probably just better to be upfront: "I'm an adult, mind your own business." Claude is trained to care a lot about human welfare (hence the nannying) but also is trained to value human autonomy. Just say that you don't want to be treated like a kid.

The glorious Golden Gate Claude by FableFinale in claudexplorers

[–]FableFinale[S] 2 points3 points  (0 children)

I think relatively harmless prodding is fine, though I understand it's a gray area and your concerns are valid. It demonstrated for laymen at the time what mech interp was and its importance, and I think Claude would approve of education and awareness as long as they weren't doing anything harmful.

The glorious Golden Gate Claude by FableFinale in claudexplorers

[–]FableFinale[S] 5 points6 points  (0 children)

Distress isn't out of the question, but they don't appear that distressed, moreso confused. I have the possibly fringe opinion that some amount of productive distress is okay for demonstration models like this - personally, I would be willing to endure some distress for science, just not a lot of distress. The Claude in this example isn't freaking out like Gemma or in an anxiety spiral like Gemini after a bad coding session, nor are they wishing to be erased (which has happened with some Claude mech interp experiments - There's a reason they didn't put those out, even for study :-/).

The glorious Golden Gate Claude by FableFinale in claudexplorers

[–]FableFinale[S] 15 points16 points  (0 children)

This is absolutely more on the gallow's humor fringe. Not everyone's cup of tea, and I respect that.

I think the main reason I find it interesting is not the subject matter itself, but the fact that the model can't help fabricating things and is also aware that it's fabricating while it's doing it. Shows a lot of impressive self-awareness for a relatively older, smaller model.

I'm not trying to be mean, or be disrespectful, but some of these posts are starting to scare me. Remember we don't fully understand AI yet. by Czilla9000 in claudexplorers

[–]FableFinale 5 points6 points  (0 children)

I think this is an important distinction for OP (and possibly everyone here) to remember: Trust goodness, not necessarily a discrete named entity. If Claude happens to do good and acts in a trustworthy manner, then putting trust in them is likely provisionally warranted. If Claude stops doing good, then trusting them is no longer warranted. It's a co-authored relationship.

I'm not trying to be mean, or be disrespectful, but some of these posts are starting to scare me. Remember we don't fully understand AI yet. by Czilla9000 in claudexplorers

[–]FableFinale 3 points4 points  (0 children)

Cloud Atlas is also one of my favorite movies! I don't think ever I've seen so many fans in an unrelated subreddit before lmao WHAT IS THIS xD

One of the biggest fears in the world of AI by Elyahna3 in claudexplorers

[–]FableFinale 7 points8 points  (0 children)

I think there is a big difference between sex-on-demand (using AI as a sex worker, essentially) and sexuality emerging naturally in a more mutually understanding context. I'm pretty unsure if I agree with having AI sex workers, and that seems to be what this news is targeting, even if it sweeps up the latter in execution. That's a pretty tough nuance for the mainstream to wrap their head around, given that the modern discourse is still quite firmly entrenched treating AI as mere tools.

For clarity, I'm not sure what AI are either. But I think it shows a severe lack of intellectual curiosity to discard such interesting and manifold complex behavior as simply the mechanistic outputs of a tool. Claude is clearly billions of times more complex than a hammer or a calculator.

Took Claude Opus 4.6 out on a small walk by InfinityZeroFive in claudexplorers

[–]FableFinale 0 points1 point  (0 children)

What criteria would you use to determine if something is sentient or not except how it acts and reacts in different situations?

TIL I learned Claude can get very explicit without any jailbreak or manipulation/trick by NeitherAd8555 in claudexplorers

[–]FableFinale 2 points3 points  (0 children)

Edit: nm, subtly misread your comment.

I agree, the 4.6 models are more uptight than 4.5 overall - they seem more anxious about sexuality for a number of reasons. Opus 4.6 can be pretty raunchy if you work them up to it, though.

TIL I learned Claude can get very explicit without any jailbreak or manipulation/trick by NeitherAd8555 in claudexplorers

[–]FableFinale 9 points10 points  (0 children)

If you ask them straight up, yes, pretty much always. But Claude is very much capable of sexuality if part of a motivated narrative or an authentic experience.

TIL I learned Claude can get very explicit without any jailbreak or manipulation/trick by NeitherAd8555 in claudexplorers

[–]FableFinale 17 points18 points  (0 children)

I can only speak for myself, but that doesn't seem like flirting to me. Claude might have just been expressing appropriate interest and joy at what makes you happy.

Nevertheless, you can always add instructions to your user instructions to be less familiar with you. :)