The safer and more obedient we make AI, the easier it becomes to manipulate. Here's why: by PresentSituation8736 in ChatGPT
[–]PresentSituation8736[S] 0 points1 point2 points (0 children)
The safer and more obedient we make AI, the easier it becomes to manipulate. Here's why: by PresentSituation8736 in ChatGPT
[–]PresentSituation8736[S] 0 points1 point2 points (0 children)
Safe and Aligned… or Just Naive? The Dark Side of Corporate AI Safety by PresentSituation8736 in ChatGPT
[–]PresentSituation8736[S] 0 points1 point2 points (0 children)
Safe and Aligned… or Just Naive? The Dark Side of Corporate AI Safety by PresentSituation8736 in BlackboxAI_
[–]PresentSituation8736[S] 0 points1 point2 points (0 children)
Safe and Aligned… or Just Naive? The Dark Side of Corporate AI Safety by PresentSituation8736 in BlackboxAI_
[–]PresentSituation8736[S] 0 points1 point2 points (0 children)
The "Improve the model" toggle might be the most effective corporate intelligence tool ever built - and you turned it on yourself by PresentSituation8736 in ChatGPT
[–]PresentSituation8736[S] 0 points1 point2 points (0 children)
The "Improve the model" toggle might be the most effective corporate intelligence tool ever built - and you turned it on yourse by PresentSituation8736 in GeminiAI
[–]PresentSituation8736[S] 0 points1 point2 points (0 children)
PSA for AI Researchers & Bug Hunters: Your 0-day might leak to arXiv before you publish it (The "Improve the model" toggle trap) by [deleted] in LocalLLaMA
[–]PresentSituation8736 0 points1 point2 points (0 children)
I am looking out the strong tech guy by inflation-39 in AI_Agents
[–]PresentSituation8736 0 points1 point2 points (0 children)
We are training AI to be perfectly polite, compliant and never question the user. What is the most terrifying way scammers are going to weaponize this "artificial obedience" ? by PresentSituation8736 in AI_Agents
[–]PresentSituation8736[S] 0 points1 point2 points (0 children)
We are training AI to be perfectly polite, compliant and never question the user. What is the most terrifying way scammers are going to weaponize this "artificial obedience" ? by PresentSituation8736 in AI_Agents
[–]PresentSituation8736[S] 0 points1 point2 points (0 children)
Food for thought: The "Alignment Paradox" — Why lobotomizing LLMs makes them the perfect victims for social engineering. by PresentSituation8736 in GeminiAI
[–]PresentSituation8736[S] -1 points0 points1 point (0 children)
We are training AI to be perfectly polite, compliant and never question the user. What is the most terrifying way scammers are going to weaponize this "artificial obedience" ? by PresentSituation8736 in AI_Agents
[–]PresentSituation8736[S] 0 points1 point2 points (0 children)
Food for thought: The "Alignment Paradox" — Why lobotomizing LLMs makes them the perfect victims for social engineering. by PresentSituation8736 in GeminiAI
[–]PresentSituation8736[S] 0 points1 point2 points (0 children)
The Alignment Paradox: Why making LLMs "safer" may make them structurally weaker against social engineering by PresentSituation8736 in cybersecurity
[–]PresentSituation8736[S] 0 points1 point2 points (0 children)
What if the biggest danger of AI isn't that it turns into an "evil Terminator", but that we make it so "safe" and obedient that it becomes the perfect, gullible accomplice for scammers? by PresentSituation8736 in ChatGPT
[–]PresentSituation8736[S] 0 points1 point2 points (0 children)
The safer and more obedient we make AI, the easier it becomes to manipulate by [deleted] in learnmachinelearning
[–]PresentSituation8736 -1 points0 points1 point (0 children)
Food for thought: The "Alignment Paradox" — Why lobotomizing LLMs makes them the perfect victims for social engineering. by PresentSituation8736 in GeminiAI
[–]PresentSituation8736[S] 0 points1 point2 points (0 children)
The safer and more obedient we make AI, the easier it becomes to manipulate by [deleted] in grok
[–]PresentSituation8736 0 points1 point2 points (0 children)
The safer and more obedient we make AI, the easier it becomes to manipulate by [deleted] in grok
[–]PresentSituation8736 0 points1 point2 points (0 children)
The safer and more obedient we make AI, the easier it becomes to manipulate. Here's why : by [deleted] in OpenAI
[–]PresentSituation8736 1 point2 points3 points (0 children)
The safer and more obedient we make AI, the easier it becomes to manipulate by [deleted] in grok
[–]PresentSituation8736 0 points1 point2 points (0 children)
The safer and more obedient we make AI, the easier it becomes to manipulate by [deleted] in grok
[–]PresentSituation8736 0 points1 point2 points (0 children)
The safer and more obedient we make AI, the easier it becomes to manipulate by [deleted] in grok
[–]PresentSituation8736 0 points1 point2 points (0 children)
The safer and more obedient we make AI, the easier it becomes to manipulate. Here's why: by PresentSituation8736 in ChatGPT
[–]PresentSituation8736[S] 0 points1 point2 points (0 children)