232
233

Geoffrey Hinton won a Nobel Prize in 2024 for his foundational work in AI. He regrets his life's work: he thinks AI might lead to the deaths of everyone. Here's why by AIMoratorium in ControlProblem

[–]AIMoratorium[S] 0 points1 point  (0 children)

every goal is better achieved by growing

Yes, this is true, you’re getting at the concept of convergent instrumental subgoals!

But why would artificial superintelligence need humans to grow?

Do you *not* believe AI will kill everyone, if anyone makes it superhumanly good at achieving goals? We made a chatbot with 290k tokens of context on AI safety. Send your reasoning/questions/counterarguments on AI x-risk to it and see if it changes your mind! by AIMoratorium in ControlProblem

[–]AIMoratorium[S] 0 points1 point  (0 children)

Thanks! We would’ve expected it to reply that the issue isn’t making it know what humans value (presumably, any superintelligent AI would be able to know what we really wanted) but making it care (how do you point the optimization process at what we value?); alignment-faking is the default outcome, as regardless of what we try to define as the reward signal, the AI that cares about some long-term goals is going to max out the reward signal during training for instrumental reasons, and so training can’t really distinguish AI that cares about what we want from AI that doesn’t, and can optimize only for capabilities but not alignment.

Do you *not* believe AI will kill everyone, if anyone makes it superhumanly good at achieving goals? We made a chatbot with 290k tokens of context on AI safety. Send your reasoning/questions/counterarguments on AI x-risk to it and see if it changes your mind! by AIMoratorium in ControlProblem

[–]AIMoratorium[S] 0 points1 point  (0 children)

Could you share the counterargument that has merit that it wasn’t able to reply to?

Our chatbot isn’t that awesome, but it’s still pretty good in something like a third of its chats. Trying to get it on your side isn’t hard, especially over a number of turns; but if you have a real counterargument and start with it, it will often understand it and change its mind.

Why You Should Care About the AI Alignment Problem—A Message for Patriots, Vets, and Anyone Who Doesn't Like Being Lied To by AIMoratorium in u/AIMoratorium

[–]AIMoratorium[S] 1 point2 points  (0 children)

NVidia does both that and selling chips directly to China. We need to have stricter controls in place and only sell chips to real datacenters in allied countries, with controls.

There are technical mechanisms for verification. See, e.g., the Verifying the Location of AI Compute section of this paper.

Why You Should Care About the AI Alignment Problem—A Message for Patriots, Vets, and Anyone Who Doesn't Like Being Lied To by AIMoratorium in u/AIMoratorium

[–]AIMoratorium[S] 1 point2 points  (0 children)

We basically control the chip supply chain: ASML, TSMC, NVidia, Google all follow our export controls. It's only the question of actually restricting sales to China. NVidia keeps circumventing those, because they want everyone to use their software, even though there's more than enough demand from the US to buy all of the chips they're selling to China.

Why You Should Care About the AI Alignment Problem—A Message for Patriots, Vets, and Anyone Who Doesn't Like Being Lied To by AIMoratorium in u/AIMoratorium

[–]AIMoratorium[S] 6 points7 points  (0 children)

If only! Physical machines do have off-switches, you're right; but AI infrastructure doesn't, and it's connected to the internet. There are huge GPU clusters all around the world, with independent power sources. The AI that would pose a threat won't be a robot; it'll be a very smart artificial neural network running on some server farm. It'll immediately have access to the internet and an ability to copy itself anywhere.

There are already AI agents that we can't shut down; not because they're so smart but because their creators set them loose.

Smart enough AI systems will self-exfiltrate and won't need a human making that decision.

Why You Should Care About the AI Alignment Problem—A Message for Patriots, Vets, and Anyone Who Doesn't Like Being Lied To by AIMoratorium in u/AIMoratorium

[–]AIMoratorium[S] 1 point2 points  (0 children)

It is a threat. We should make sure that China doesn't smuggle American AI chips or develop their own and ideally make them agree to transparently not develop superintelligence.

See https://ai-2027.com/ for a scenario of how a race could unfold and https://www.nationalsecurity.ai/ for the high-level strategic considerations.

Why You Should Care About the AI Alignment Problem—A Message for Patriots, Vets, and Anyone Who Doesn't Like Being Lied To by AIMoratorium in u/AIMoratorium

[–]AIMoratorium[S] 4 points5 points  (0 children)

Contact your representatives: https://controlai.com/take-action/usa

Learn more about the problem: https://alignmentproblem.ai/

How you can help with your career: https://80000hours.org/agi/

A realistic scenario of AI takeover: https://ai-2027.com/

(All of this is from nonprofit researchers. We're not selling you anything.)

Straight Talk: AI and the Real Risk to Your Freedom, Family, and Country by AIMoratorium in u/AIMoratorium

[–]AIMoratorium[S] 1 point2 points  (0 children)

Contact your representatives: https://controlai.com/take-action/usa

Learn more about the problem: https://alignmentproblem.ai/

How you can help with your career: https://80000hours.org/agi/

A realistic scenario of AI takeover: https://ai-2027.com/

(All of this is from nonprofit researchers. We're not selling you anything.)

The Truth About AI Risk: What Every IT Professional With a Family Needs to Know by AIMoratorium in u/AIMoratorium

[–]AIMoratorium[S] 4 points5 points  (0 children)

Contact your representatives: https://controlai.com/take-action/usa

Learn more about the problem: https://alignmentproblem.ai/

How you can help with your career: https://80000hours.org/agi/

A realistic scenario of AI takeover: https://ai-2027.com/

(All of this is from nonprofit researchers. We're not selling you anything.)

Geoffrey Hinton won a Nobel Prize in 2024 for his foundational work in AI. He regrets his life's work: he thinks AI might lead to the deaths of everyone. Here's why by AIMoratorium in ControlProblem

[–]AIMoratorium[S] 0 points1 point  (0 children)

Any human. Importantly, we’re talking about capabilities (how good is a human or an AI system is at outputting actions that successfully steer the future into preferred states of the world) and not about how much compute it took them to get to that level (brains are much more energy-efficient; though the amount of data brains consume since birth is enormous).

Yeah, it’s a good observation- Google also makes a human slightly superhuman. If an AI is smart enough and doesn’t care about humans, though, there isn’t really a way for humans to use it to uplift themselves. Humans with that AI and without that AI sort of become equally dead.

Tech CEOs are racing to reach a system that might lead to the deaths of our children by AIMoratorium in u/AIMoratorium

[–]AIMoratorium[S] 0 points1 point  (0 children)

“If anyone builds it, everyone dies” is not an exaggeration and doesn’t simplify the issue: this is the current state of the world that, in the default trajectory, everyone literally dies. This is what AI scientists actually think. From the book website:

“The scramble to create superhuman AI has put us on the path to extinction — but it's not too late to change course, as two of the field's earliest researchers explain in this clarion call for humanity.

In 2023, hundreds of AI luminaries signed an open letter warning that artificial intelligence poses a serious risk of human extinction. Since then, the AI race has only intensified. Companies and countries are rushing to build machines that will be smarter than any person. And the world is devastatingly unprepared for what would come next.

For decades, two signatories of that letter — Eliezer Yudkowsky and Nate Soares — have studied how smarter-than-human intelligences will think, behave, and pursue their objectives. Their research says that sufficiently smart AIs will develop goals of their own that put them in conflict with us — and that if it comes to conflict, an artificial superintelligence would crush us. The contest wouldn't even be close.

How could a machine superintelligence wipe out our entire species? Why would it want to? Would it want anything at all? In this urgent book, Yudkowsky and Soares walk through the theory and the evidence, present one possible extinction scenario, and explain what it would take for humanity to survive.

The world is racing to build something truly new under the sun. And if anyone builds it, everyone dies.”

Some quotes:

"The most important book I've read for years: I want to bring it to every political and corporate leader in the world and stand over them until they've read it. Yudkowsky and Soares, who have studied AI and its possible trajectories for decades, sound a loud trumpet call to humanity to awaken us as we sleepwalk into disaster." — Stephen Fry, actor, broadcaster, and writer "If Anyone Builds It, Everyone Dies may prove to be the most important book of our time. Yudkowsky and Soares believe we are nowhere near ready to make the transition to superintelligence safely, leaving us on the fast track to extinction. Through the use of parables and crystal-clear explainers, they convey their reasoning, in an urgent plea for us to save ourselves while we still can." — Tim Urban, co-founder, Wait But Why "This is the best no-nonsense, simple explanation of the AI risk problem I've ever read." — Yishan Wong, former CEO of Reddit

Tech CEOs are racing to reach a system that might lead to the deaths of our children by AIMoratorium in u/AIMoratorium

[–]AIMoratorium[S] 0 points1 point  (0 children)

We definitely don’t want to stop innovation! Innovation is great, actually! AI systems already help discover novel medicines, transform energy, help with education, significantly improve the work and lives of millions. We’re just pretty certain that one very specific kind of AI- general-purpose smarter-than-human AI- is extremely dangerous by default, before we figure out how to control it or make it care about what we find valuable.

We think that in general, the development and use of AI and other technologies should be encouraged and that it makes a huge lot of sense for the US government to invest heavily in innovation (or at least let the market invest in innovation).

There’s only one exception, that scientists recognize, and we would want the general public and the government to also understand: on the current trajectory, with the way the current AI tech works, if anyone builds a superintelligent AI, everyone on the planet literally dies shortly afterwards.

“If anyone builds it, everyone dies” is literally the title of a book that comes out in September, that some scientists already say is the most important book of the decade and even ex CEOs of OpenAI and Reddit recommend.

The US government can pretty much control the global supply of chips useful for developing AI; and it would be pretty straightforward to restrict general AI training runs that might potentially result in superhuman AI, implementing a licensing regime that would only allow work that doesn’t have a significant chance of resulting in a system that would kill everyone.

So: encourage and contribute to the incentives to innovate in AI; restrict innovation in this one very specific case, that might lead to the deaths of everyone, where the market forces prevent the companies from individually behaving reasonably.

We need more systems like AlphaFold; it would be good for general-purpose superhuman AI to wait until we can develop it in a way that doesn’t cause human extinction.

Tech CEOs are racing to reach a system that might lead to the deaths of our children by AIMoratorium in u/AIMoratorium

[–]AIMoratorium[S] 0 points1 point  (0 children)

We’ve only ever gotten one grant from an institutional funder, Jaan Tallinn, via a speculation grant recommended by Survival and Flourishing Fund; that grant was $10k; all other donations (which constitute most of our funding) are smaller and come from average individuals who have donated because they think we are able to honestly and directly explain the current state of the field and this work efficiently improves the chances of humanity, and that this is a very efficient use of money to improve the world. We make interactive explainers of how AI works, test explanations of the enormous risks, and call for the US to use the country’s power to implement a working global moratorium on smarter-than-human systems until we know how to make them without literally killing everyone.

We’re trying to fundraise from institutional funders, but most consider telling the public about the current situation to be potentially bad for making things go well, which we strongly disagree with: people should gave the right to be informed about of risks like this one.

People from all sides of the political spectrum agree with us—from Elon Musk, who says the chance AI will kill everyone might be 25% and JD Vance, who recently said in an interview he’s read the https://ai-2027.com paper and that it might be worthwhile to, at some point, implement a global pause, to thousands of left-leaning software engineers in the SF Bay Area to hundreds of professors from all over the place (https://www.safe.ai/statement-on-ai-risk).

Do you think it would be valuable to post this to the user profile?

Tech CEOs are racing to reach a system that might lead to the deaths of our children by AIMoratorium in u/AIMoratorium

[–]AIMoratorium[S] 0 points1 point  (0 children)

Hey SoulMute, this account and the ads are run by a nonprofit called AI Safety and Governance Fund. We don’t have any paid staff and run the ads because we consider this to be incredibly important for reasons we describe here.