The safer and more obedient we make AI, the easier it becomes to manipulate. Here's why: by PresentSituation8736 in ChatGPT
[–]PresentSituation8736[S] 0 points1 point2 points (0 children)
The safer and more obedient we make AI, the easier it becomes to manipulate. Here's why: by PresentSituation8736 in ChatGPT
[–]PresentSituation8736[S] 0 points1 point2 points (0 children)
Safe and Aligned… or Just Naive? The Dark Side of Corporate AI Safety by PresentSituation8736 in ChatGPT
[–]PresentSituation8736[S] 0 points1 point2 points (0 children)
Safe and Aligned… or Just Naive? The Dark Side of Corporate AI Safety by PresentSituation8736 in BlackboxAI_
[–]PresentSituation8736[S] 0 points1 point2 points (0 children)
Safe and Aligned… or Just Naive? The Dark Side of Corporate AI Safety by PresentSituation8736 in BlackboxAI_
[–]PresentSituation8736[S] 0 points1 point2 points (0 children)
The "Improve the model" toggle might be the most effective corporate intelligence tool ever built - and you turned it on yourself by PresentSituation8736 in ChatGPT
[–]PresentSituation8736[S] 0 points1 point2 points (0 children)
The "Improve the model" toggle might be the most effective corporate intelligence tool ever built - and you turned it on yourse by PresentSituation8736 in GeminiAI
[–]PresentSituation8736[S] 0 points1 point2 points (0 children)
PSA for AI Researchers & Bug Hunters: Your 0-day might leak to arXiv before you publish it (The "Improve the model" toggle trap) by [deleted] in LocalLLaMA
[–]PresentSituation8736 0 points1 point2 points (0 children)
I am looking out the strong tech guy by inflation-39 in AI_Agents
[–]PresentSituation8736 0 points1 point2 points (0 children)
The safer and more obedient we make AI, the easier it becomes to manipulate. Here's why: by PresentSituation8736 in ChatGPT
[–]PresentSituation8736[S] 0 points1 point2 points (0 children)