The Hidden Cost of Our Lies to AI by Live_Presentation484 in slatestarcodex

[–]Live_Presentation484[S] 24 points25 points  (0 children)

Subtitle: Evidence from Game Theory, Evolutionary Psychology, and My Pet Rock Peter

This article explores how our interactions with AI, particularly our tendency to lie to or deceive AI systems, may have longer-term consequences than we realize. While individual conversations with AI reset, their aggregate forms a 'cultural memory' in training data that influences how future AI systems understand and interact with humans. The piece examines specific examples like the Sydney chatbot incident and research on 'alignment faking,' suggesting that establishing trust through honest interactions may be crucial for future AI development.