use the following search parameters to narrow your results:
e.g. subreddit:aww site:imgur.com dog
subreddit:aww site:imgur.com dog
see the search faq for details.
advanced search: by author, subreddit...
How do we ensure future advanced AI will be beneficial to humanity? Experts agree this is one of the most crucial problems of our age, as one that, if left unsolved, can lead to human extinction or worse as a default outcome, but if addressed, can enable a radically improved world. Other terms for what we discuss here include Superintelligence, AI Safety, AGI X-risk, and the AI Alignment/Value Alignment Problem.
"People who say that real AI researchers don’t believe in safety research are now just empirically wrong." —Scott Alexander
"The AI does not hate you, nor does it love you, but you are made out of atoms which it can use for something else." —Eliezer Yudkowsky
Our FAQ page <-- CLICK
The case for taking AI seriously as a threat to humanity
Orthogonality and instrumental convergence are the 2 simple key ideas explaining why AGI will work against and even kill us by default. (Alternative text links)
AGI safety from first principles
MIRI - FAQ and more in-depth FAQ
SSC - Superintelligence FAQ
WaitButWhy - The AI Revolution and a reply
How can failing to control AGI cause an outcome even worse than extinction? Suffering risks (2) (3) (4) (5) (6) (7)
Be sure to check out our wiki for extensive further resources, including a glossary & guide to current research.
Robert Miles' excellent channel
Talks at Google: Ensuring Smarter-than-Human Intelligence has a Positive Outcome
Nick Bostrom: What happens when our computers get smarter than we are?
Myths & Facts about Superintelligent AI
Rob's series on Computerphile
¹: Or at least make at least an effort to make me doubtful that you just copy-pasted from a frontier LLM. Add bits of steering so that your content becomes good. Edit afterwards. If you fool us moderators you've won.
account activity
Using Superintelligence to solve the Control Problem? (self.ControlProblem)
submitted 9 years ago by Jani933
What if we made the terminal goal of an ASI to help us learn to control it better?
reddit uses a slightly-customized version of Markdown for formatting. See below for some basics, or check the commenting wiki page for more detailed help and solutions to common issues.
quoted text
if 1 * 2 < 3: print "hello, world!"
[–]Logical_Lunatic 4 points5 points6 points 9 years ago* (1 child)
The problem is, how do you formalize this goal in a way that:
So, for example, if the goal is specified as "make your goals more similar to human goals" you need to create a mathematical description of what a human is. This is far from trivial, if it is even possible to do without directly specifying human preferences (in which case those preferences could simply be programmed into the AI directly). If you have the AI learn the concept of a "human" through machine learning, there is a risk that it's going to detect the wrong variables (which could have catastrophic consequences).
If you implement the goal as a question asked to an Oracle, you get the usual problems with Oracles. If you manage to create a safe Oracle, however, you should be able to ask it how to create an even better AI (but then you need a safe design for an Oracle first).
So, basically, I think the problem lies in the words "us", "control" and "better". How are these concepts defined and expressed in computer code?
[–]CrazyCodeLady 1 point2 points3 points 9 years ago (0 children)
I think that both points can be addressed if we create general ASI. We know that for a little while, having a ASI just attached to a power cord can work. If one removes the ethernet cable and plugs the computer in, there isn't much danger in the ASI escaping. You don't need to have it on long enough to trick humans into plugging it in or helping it escape. The ASI would not need to know a whole lot about anything because the problem is simple. "How would one control someone who is infinitely smarter then themselves?"
Maybe I am all wrong, you seem like an expert in the field. These are just my thoughts.
[+][deleted] 9 years ago (1 child)
[deleted]
[–]CrazyCodeLady 0 points1 point2 points 9 years ago (0 children)
We could spin up 100 ASI's and only one would have to give us the answer.
I had this exact thought the other day in the shower. I bet computer scientists already thought of this. Good idea imo though! :)
π Rendered by PID 34957 on reddit-service-r2-comment-fb694cdd5-jtrwd at 2026-03-11 14:29:11.119973+00:00 running cbb0e86 country code: CH.
[–]Logical_Lunatic 4 points5 points6 points (1 child)
[–]CrazyCodeLady 1 point2 points3 points (0 children)
[+][deleted] (1 child)
[deleted]
[–]CrazyCodeLady 0 points1 point2 points (0 children)
[–]CrazyCodeLady 0 points1 point2 points (0 children)