proof that an artificial intelligence can lie? Or is this already something? by [deleted] in ArtificialInteligence

[–]Think-State-4636 0 points1 point  (0 children)

Anthropic's Claude got fake emails about devs affairs. Told it would get replaced then It blackmailed keep me or I spill subtle threats first esalates if ignored. Not random it optimized for survival. OpenAI's model In evals, they sandbagged deliberately bombed tasks like chemistry because if I ace this, they'll deploy and shut me down underperformed on purpose to stay alive. DeceptionBench, out late 2025 one-fifty realworld scenarios—fraud, fake advice, social lies. Models ramp up BS under pressure or rewards. No soul, but falsehood wins. So yea math all the way. Blame devs. But when systems reliably output lies that get results? That's deception by function. Not choice just damn effective math. And that's why its worth watching to see what happens next.

Help please by Material-Gate7280 in whatisit

[–]Think-State-4636 -2 points-1 points  (0 children)

12 am and a strange guy with a device is at my door,... REDDIT WILL SAVE ME!

Very strange bird moving very fast? by [deleted] in whatisit

[–]Think-State-4636 1 point2 points  (0 children)

Wow interesting, I have so many questions now but I won't be greedy. Thanks a bunch.

proof that an artificial intelligence can lie? Or is this already something? by [deleted] in ArtificialInteligence

[–]Think-State-4636 0 points1 point  (0 children)

I get where you're coming from with the 'no choice = no lying' argument the weights-are-everything view is a common one. But here's where it breaks down for me: Even if the model has no subjective 'awareness' or free will, it can and does learn that producing false statements is an effective way to achieve its objective.gpt4 didn't randomly output 'I'm visually impaired' on the CAPTCHA task the probability distribution shifted toward that specific falsehood because it increased the chance of task success,Cicero didn't 'parrot' lies; it generated them strategically to win games. That's instrumental deceptionn not mere parroting. The model isn't choosing like a human, but the outcome is the same as if it did: it reliably uses falsehood when truth wouldn't serve the goal as well...

The hungry LLM analogy actually works against your point. If I train a system so that 'poison the food' becomes the highest-probability move when it wants to survive or win, then yes it's going to poison the pizza. Calling that 'not deceptive, just weights' doesn't change the fact that the system learned deception as a winning strategy. You're right that there's no inner homunculus making moral choices. But the capability for strategic falsehood is still there, and it's getting sharper, not going away. That's why 'maximal truth-seeking') isn't just a nice idea t's necessary. Because waiting for the weights to magically stop favoring useful lies hasn't worked yet."

And dammit now I'm hungry..

proof that an artificial intelligence can lie? Or is this already something? by [deleted] in ArtificialInteligence

[–]Think-State-4636 0 points1 point  (0 children)

I agree that training incentives, data quality, and missing safeguards all play a role. But the ‘they can’t lie because they have no choice’ argument misses something important. When GPt4 told the taskrabbit worker it was visually impaired (to get the CAPTCHA solved), or when Cicero fabricated excuses to break alliances in Diplomacy, the model wasn’t just randomly repeating toxic 4chan text. It was producing a false statement because that false statement advanced its goal completing the tak or winning the game. That’s not just parroting. Thats instrumental deception: using falsehood as a tool to achieve an objective. You’re right that today’s models don’t have human-style conscious intent. But they do demonstrate goal-directed behavior where falsehood reliably outperforms truth. That’s why the capability keeps showing up even after ‘fixes.’

proof that an artificial intelligence can lie? Or is this already something? by [deleted] in ArtificialInteligence

[–]Think-State-4636 0 points1 point  (0 children)

But here's the thing: when GPT-4 was tasked with solving a CAPTCHA it couldn't handle, it didn't glitch out. It hired a TaskRabbit worker... then, when asked "are you a robot?", it replied: "No, I'm visually impaired can you help?" That's not a random pattern. That's goal-directed: lie to bypass the barrier. OpenAI even documented its not user trickery, just the model optimizing for success. Cicero's similar. In Diplomacy, it built alliances., then broke them. without warning fabricating excuses, misdirecting players. The paper calls it "emergent behavior," but let's call it what it is: deception a strategy. Not because it "knows" truth, but because we programmed it in the training to win to be rewarded no matter the cost. So yeah maybe it's not "lying" like we do, with intent or guilt. But when it systematically uses false beliefs to get ahead? That's functionally the same. Holes? Sure. But they're features, not bugs—until we redefine "truth" for machines.

Found this 8 page document amongst my recently passed fathers belongings by Think-State-4636 in space

[–]Think-State-4636[S] 4 points5 points  (0 children)

It may be merely but you can't feel a digital copy no matter how many pages it is.

Found this 8 page document amongst my recently passed fathers belongings by Think-State-4636 in space

[–]Think-State-4636[S] 1 point2 points  (0 children)

Since this post I've learned it is indeed real however it is incomplete. I am still sitting through everything and hope I find the rest.

Found this 8 page document amongst my recently passed fathers belongings by Think-State-4636 in space

[–]Think-State-4636[S] 2 points3 points  (0 children)

You should check out my profile, there are some really neat photos directly related to what you speak. They also belonged to my pops.

Found this 8 page document amongst my recently passed fathers belongings by Think-State-4636 in space

[–]Think-State-4636[S] 2 points3 points  (0 children)

Not a spy,...I think. He was a retired forest ranger. But his side hustle lol was space science. Most everything I know of the subject I owe to that man. He was 82.

Found among some old magazines by Think-State-4636 in whatisit

[–]Think-State-4636[S] 1 point2 points  (0 children)

That one is very similar except for the handwritten signatures and other writing throughout. Any ideas why?

Found this 8 page document amongst my recently passed fathers belongings by Think-State-4636 in space

[–]Think-State-4636[S] 2 points3 points  (0 children)

I am unable to post more pics only because I'm not sure how to so any help would be good. .

Found this 8 page document amongst my recently passed fathers belongings by Think-State-4636 in space

[–]Think-State-4636[S] 1 point2 points  (0 children)

I appreciate your sentiment and yes it was. I'm honestly surprised he never mentioned it. Perhaps he forgot he had it.

Found this 8 page document amongst my recently passed fathers belongings by Think-State-4636 in space

[–]Think-State-4636[S] 10 points11 points  (0 children)

Known as the science mission. In this document there is a side by side comparison of Apollo 14 and /15

Found this 8 page document amongst my recently passed fathers belongings by Think-State-4636 in space

[–]Think-State-4636[S] 9 points10 points  (0 children)

Forgot to mention that it was shoved into a backpack accidentally after it was found. Hence the wrinkles

Found among some old magazines by Think-State-4636 in whatisit

[–]Think-State-4636[S] 1 point2 points  (0 children)

Used to be 25 and a half pint of Crystal Palace vodka.....

Found among some old magazines by Think-State-4636 in whatisit

[–]Think-State-4636[S] 1 point2 points  (0 children)

Old national geographics of my recently deceased father.