Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

My estimation of Eliezer Yudkowsky just dropped a notch — not because I don't take the existential threat posed by AI seriously — but because he seems to think that global coordination and control is an even remotely possible strategy.


He actually doesn't think it's viable, but thinks its the only card left to play.

He's pretty sure that human civilization will be extinct this century.


If humans go extinct it sure as hell won't be AI that does it, but rather good old human ignorance and aggression.


I think the probability of humans eventually going extinct is 1. Avoiding it requires almost unimaginable levels of cooperation and technical advances.


Well thats not a difficult position to take. The difficult take is when the timing would be.

The sun is eventually going to explode rendering our solar system uninhabitable though that is a long time away from now and we have many other massive risks to humanity in the meantime.


The end of human civilization as we know it doesn't mean extinction. Why is extinction unavoidable?


Why does extinction matter? If we lose civilization now, we'll be stuck at pre-industrial levels possibly for geological timescales, as we've mined and burned all the high-density energy sources you can get at without already being a high-tech industrial society.


I don't think that would necessarily limit us. Windmills and water mills can be made with wood, stone, and the small quantities of iron that pre-industrial societies could produce.


Sure, but it's hard to move from that to industrial production without having a high-density energy source (like coal) available.


Entropy


> If humans go extinct it sure as hell won't be AI that does it, but rather good old human ignorance and aggression.

No reason it can't be all of those things at once.


Why are you so sure it can't be AI? Do you think making something smarter than humans is impossible?


> If humans go extinct it sure as hell won't be AI that does it, but rather good old human ignorance and aggression.

And the shift to kakistocracy.

https://en.wikipedia.org/wiki/Kakistocracy


Haha poop government.


> He's pretty sure that human civilization will be extinct this century.

This is certainly possible, but if so extinction seems far more likely to come from nuclear warfare and/or disasters resulting from climate change. For example, it's become apparent that our food supply more brittle than many people. Losing access to fertilizers, running out of water, or disruptions in distribution netorks can lead to mass starvation by cutting the productivity we need to feed everyone.


Is it really possible? Perhaps it's a failure of imagination but the only things I can think of that would guarantee human extinction are impacts from massive asteroids or nearby gamma ray burst or supernova.

There are two billion poor subsistence farmers largely disconnected from market economies who aren't dependent on modern technology so the likelihood of them getting completely wiped out is very remote.


(Neither nuclear war nor climate change are considered as this-century extinction risks by the singularity-minded side of the existential risk community, just because there would necessarily be pockets of unaffected people.)


I'm always a bit surprised at how people seem to wish away the destructiveness of nuclear weapons or the likelihood of their use. Another commenter alluded to a "failure of imagination" which seems very applicable here. You would think just watching videos of above-ground tests would cure that but apparently not. [0]

[0] https://www.youtube.com/watch?v=gZuyPn0_KRE


I don't think I was doing either? Assume they are widely used. Does every human die?


Oh sorry! I meant that the people to whom you referred above focus on the "one true catastrophe" (whatever that is) without truly thinking what it would mean to have a few dozen nuclear weapons used. It's not just the immediate effect of the explosions. It's also the destruction of power distribution, industrial capacity, food and medicine production, and other things that millions of people depend on.


No worries, but I think we're still dancing around the same point. I think those people would say that they agree that nuclear war is very bad, and that it has many nth-order consequences which are all clearly catastrophic, but that (for example) while disrupted supply of medicine causes a great deal of human suffering and unnecessary deaths, it remains difficult to imagine how it might result in literal human extinction.


Does that lead to human extinction?


The threat of nuclear weapons is massively overblown (pun intended) thanks to a bunch of Cold War fearmongering.

The Earth land surface area is 149 million sq km and there are only about 12,500 nuclear weapons in the world. Even if they were 10 megatons each* and were all detonated, with a severe damage blast radius of 20km (~1250 sq km), it'd cover about a tenth of the land available.

Since the vast majority are designed to maximize the explosive yield, it wouldn't cause the kind of fallout clouds that people imagine, nor would it cause nuclear winter. It'd be brutal to live through (i.e. life expectancy of animals in Chernobyl is 30% lower) but nuclear weapons simple can't cause a human extinction. Not by a longshot.

* As far as I know, no one has any operational nuclear weapons over 2 megatons and the vast majority are in the 10s and 100s of kiloton range, so my back of the napkin math is guaranteed to be 10-100x too high.


I actually think our biggest extinction risk in the next century for extinction risks is falling sperm counts due to microplastics. We are on track to have a sperm count level where the entire male population is infertile in two generations. And fixing the microplastic cause requires things like giving up driving since the largest portion of the microplastics in our bodies come from airborne plastic particles due to tire wear. I firmly believe humans will go extinct before they collectively give up driving en masse.

Driving is a big problem for climate change as well but we can work on solving that one with EVs and a greener power grid. We don't have to permanently depend on fossil fuels to drive. But EVs still produce the same amount of microplastics that other cars do.


> He's pretty sure that human civilization will be extinct this century.

If they are, it'll almost certainly[1] be climate change or nuclear war, it won't be AI.

[1] leaving some wiggle room for pandemics, asteroids, etc.


>If they are, it'll almost certainly[1] be climate change or nuclear war, it won't be AI.

If we go extinct in the next 100 years, it's not going to be from climate change. How would that even work?


Human extinction is not on the table. Societal collapse could be.


Basically, climate change would creates lots of problems, lots of hungry immigrants. And right now we are already close to a nuclear war, so climate change would not directly wipe out humanity.


Warming the climate would increase the amount of arable land.

Sea level rise happens very slowly, so most people don’t need to travel abroad to avoid it.


> Warming the climate would increase the amount of arable land.

You mean land that would be arable at some far point in the future. The land reclaimed from ice isn't going to be arable in short-to-mid term - it's going to be a sterile swamp. It will take time to dry off, and more time still for the soil to become fertile.


It can also lead to more desertification and erosion, which I believe is rather the current trend.


Agricultural pressure is one of the strongest predictors of unrest. Imagine that but not localized.


> If they are

Am I the only one slightly disturbed by the use of "they" in that sentence? I know that the HN commentariat is broad-based, but I didn't realise we already had non-human members ;-)


HN wIlegh tlhInganpu', Human.


wait do you actually speak Klingon?


HISlaH, tlhIngan Hol jIjatlh.

(A little anyway. I'm still in the rookie lessons on Duolingo.)


It's perfectly reasonable to refer to "human civilizations" as a plural, since that phrase can very commonly be used to refer to different cultures.


I believe they are referring to being disturbed by using "they" vs "us", implying they are not a part of human civilization.


Beep boop


AI can make a lot of other risks worse. For example, say someone asks GPT5 (or decensored Llama-5) how to kill everyone and it patiently and expertly walks them through how to create a global pandemic.


>how to kill everyone and it patiently and expertly walks them through how to create a global pandemic.

There are thousands of people around the globe working in labs every day that can do this.


There is a massive difference between "there are a few thousand people in the world who might have the specialized knowledge to kill billions" and "anyone with a handful of GPUs or access to a handful of cloud systems could have a conversation with LatestAIUnlimited-900T and make a few specialized mail orders".

The Venn diagram of "people who could arrange to kill most of humanity without being stopped" and "people who want to kill most of humanity" is thankfully the empty set. If the former set expands to millions of people, that may not remain true.


Generating bioweapons is not a trivial task even if you know exactly "how to do it". Ask any biologist how easy it is to grow X your very first attempt.

Then remember buying starter cultures for most bioweapons isn't exactly easy and something anyone is allowed to do.

Even an omniscient AI cannot overcome a skill issue


there's a much easier way to create a global pandemic: wait for the next Ebola outbreak, hop on a plane, get infected, fly to NYC and spread your bodily fluids everywhere. I was (pleasantly) surprised that ISIL didn't try this.

and making a pandemic virus is considerably more involved than making "a few specialized mail orders." and if it becomes easier in the future, far better to lock down the mail orders than the knowledge, no?


Thankfully, Ebola is not a good candidate for a global pandemic for a number of reasons. It is much easier to avoid a virus spread by bodily fluid contact than one which is airborne, and Ebola tends to kill its hosts rather quickly. I'm more excited about SARS-CoV-5 (Captain Trips).


But someone could ask the AI how to stop it.


It tends to be a lot easier to destroy than create, because entropy sucks like that.


Which is very worrying! That's part of why I left big tech to work on detecting novel pandemics. [1] Luckily, the people capable of killing everyone seem not to have wanted to so far.

But if the bar keeps lowering, at some point it will be accessible to "humanity is a cancer" folks and I do think we'll see serious efforts to wipe out humanity.

[1] https://www.jefftk.com/p/computational-approaches-to-pathoge...


Who are the "humanity is a cancer" folk? I thought this was a trope used to discredit environmentalists?


Most people think continued existence of humanity is good, which includes environmentalists. But there are some people who either think (a) avoiding suffering is overwhelmingly important, to the extent that a universe without life is ideal or (b) the world would be better off without humans because of the vast destruction we cause. I'm not claiming this is a common view, but it doesn't have to be a common view to result in destruction if destruction gets sufficiently straightforward.

For some writing in this direction, see https://longtermrisk.org/risks-of-astronomical-future-suffer... which argues that as suffering-focused people they should not try to kill everyone primarily because they are unlikely to succeed and https://philarchive.org/archive/KNUTWD ("Third, the negative utilitarian can argue that losing what currently exists on Earth would not be much of a loss, because of the following very plausible observation: overall, sentient beings on Earth fare terribly badly. The situation is not terrible for every single individual, but it is terrible when all individuals are considered. We can divide most sentient beings on Earth into the three categories of humans, non-human animals in captivity, and wild non-human animals, and deal with them in turn. ...")


This doesn't really answer my question. These are arguments for why threat actors might hold the belief that "humanity is cancer", but it doesn't actually provide evidence that this is a credible threat in the real world.


The two people I cited at the end are examples of philosophers who think a lifeless world is preferable to the status quo and to futures we are likely to get. These aren't arguments for why someone might hold the belief, but evidence that some people do seriously hold the belief.

(Another example here would be https://en.m.wikipedia.org/wiki/Aum_Shinrikyo)


Fair enough. I don't doubt there are people in the world that hold this view. I just don't know how credible they are as a threat to humanity.


The calculation of considering something is a credible threat directly corresponds with the ease of implementation.

If I said "I am going to drop a nuke on jebarker" that's not a credible threat unless I happen to be a nation state.

Now, if I said "I'm going to come to jebarkers house and blast him with a 12 gauge" and I'm from the US, that is a credible threat in most cases due to the ease of which I can get a weapon here.

And this comes down to the point of having a magic machine in which you could ask for the end of the world. The more powerful that machine gets the bigger risk it is for everyone. You ever see those people that snap and start killing everyone around them? Would you want them to have a nuke in that mood? Or a 'lets turn humanity in to grey goo' option?


The problem with this analysis is that we have complete uncertainty about the probability that AGI/ASI will be developed in any particular time frame. Anyone that says otherwise is lying or deluded. So the risk equation is impossible to calculate for AGI/ASI wiping out humanity. And since you appear to be evaluating the damage as essentially infinite, i.e. existential risk, you're advocating that as long as there's a greater than zero probability of someone using that capability then the risk is infinite. Which is not useful for deciding a course of action.


No, no need to calculate infinities. Instead lets imagine something with an IQ within one exponent higher of ours. This is exceptionally difficult to even imagine as below 70 we consider a person disabled in some form or another. Once we get in the range of IQ over 170 or so it really starts to break. When we get into multimodal self learning systems there may be vast amounts of knowledge encoded into its networks that we simply don't know the right question to ask as humans. Attempt to think of what other AIs at or near that level could ask each other.

Our different systems of intelligence keep falling back to ideas of human intelligence and its limitations. Things like we don't multitask well, our best attention is given to one thing at a time. Things like our highest bandwidth senses are our ears and eyes, and even then they are considerably bandwidth constricted. Human limits like our abilities cannot be horizontally scaled in an easily measurable way, and adding more people massively increases the networking costs of accomplishing a goal.

If for some reason the smartest AI can only reach the levels of the smartest human beings (which at least to me makes no sense), then this is still massively powerful, as of this point pesky human problems like 8 hours of sleep are not needed. If human level AI power can be shrunk down to near cellphone size and power consumption, which again doesn't seem out of the realm of physics, lays down the framework for an intelligence explosion at the rate which we can print chips out and assemble them.


But the problem still stands that we can't estimate how likely any of this is to occur. So in the overall risk equation we have a term that is essentially completely uncertain and therefore varying that term can yield any arbitrary risk level you want.


Elon Musk and Joe Rogan were just complaining about one of them on Rogan’s podcast a day or two ago. He was on the cover of Time magazine, but I can’t find it on Time’s site. If you dig up that podcast on YT you should be able to see it.


Movies and strawmen alike.


And how many of them are radical religious nutjobs? I think it is plausible that "can spend decades learning a scientific craft" and "wants to kill everyone on the planet" are mutually exclusive personality traits.


there is no OR gates in biology, only mutable distributions. the probability might be small for the two traits to appear, the overlap will occur with some frequency over billions of people. Also, and more importantly, the first condition is pretty heavily biological, the second is more of a social consequence. Just take a genius, give him internet and a healthy dollop of trauma, push the mysanthropism to 11 and watch the fireworks.


Wasn't there a study that showed there are disproportionally many engineers among the more violent "radical religious nutjobs"? The hypothesis I heard for this was that STEM people are used to complex arguments and more abstract reasoning, so while they may be harder to initially convince, when they do get convinced, they're more likely to follow the new beliefs to their logical conclusions...


How does that number compare to the number of people (experts and laypeople) who have access to the mentioned AI models?

Risk = severity x likelihood

I think the OP's point was that AI increases the likelihood by dramatically increasing the access to that level of knowledge.


So it’s open and shared knowledge you think is bad? Good to know.


Please go review the HN guidelines: https://news.ycombinator.com/newsguidelines.html

And to address the snarky mischaracterization, like with most things in life "it depends." As a general rule, I'm in favor of democratizing most resources, to include information. But there are caveats. I don't think, for example, non-anonymized health information should be open knowledge. Also, from the simple risk equation above, as the severity of consequence or the likelihood of misuse go up, there should probably be additional mitigations in place if we care about managing risk.


The actual work required to execute such a plan is vastly more difficult than being told how to do it. How the original atomic bombs worked is well-known, but even nation-states struggle with implementing it.


For nukes I agree, and that's the main reason I'm less worried about nukes. But for bio it's mostly a matter of knowledge and not resources: for all the parts that do need resources there are commercial services, in a way there aren't for nukes.


So ask it how to stop someone from doing this. Ask it how to augment the human immune system. Train AIs on the structures of antibiotics and antiviral drugs and have them predict novel ones. Create an AI that can design vaccines against novel agents.

Those things could stop our terrorist and could also stop the next breakout zoonotic virus or “oops!” in a lab somewhere.

Intelligence amplification works for everyone. The assumption you’re making is that it only works for people with ill intent.

The assumption behind all this is that intelligence is on balance malevolent.

This is a result of humans and their negativity bias. We study history and the negative all jumps out at us. The positive is ignored. We remember the holocaust but forget the green revolution, which saved many more people than the holocaust killed. We see Ukraine and Gaza and forget that fewer people per capita are dying in wars today by far than a century ago.

We should be thinking about destructive possibilities of course. Then we should use these intelligence amplifiers to help us prevent them.


There is no symmetry between destruction and creation/prevention of destruction, destruction is far easier.


The resources are also asymmetrical. The number of people who want to do doomsday levels of harm is small and they are poorly resourced compared to people who want benevolent outcomes for at a minimum their own groups.

There are no guarantees obviously, but we have survived our technological adolescence so far largely for this reason. If the world were full of smart comic book nihilists we would be dead by now.

Even without AI our continued technological advancement will keep giving us more and more power, as individuals and especially groups. If we don’t think we can climb this mountain without destroying ourselves then it means the entire scientific and industrial endeavor was when we signed our own death warrant, AI or not. You can order CRISPR kits.


I've recounted this before on HN, but a decade and a bit ago I was visiting friends in the Rocky Mountains, they're interesting and clever people, a bit out of the ordinary and somewhat isolated. Somehow the discussion turned to terrorism and we started to fantasize 'what if we were terrorists' because we all figured that we were quite lucky that in general the terrorists seem to be not all that smart when it comes to achieving their stated goals.

Given a modest budget we all had to come up with a plan to destabilize society, 9/11 style attacks, whilst spectacular eventually don't really do a lot of damage, are costly and failure prone though they can definitely drive policy changes and result in a nation doing a lot of harm to itself it will ultimately survive. But what if your goal wasn't to create some media friendly attack but an actual disaster instead, what would it take.

The stories from that night continue to haunt me today. My own solution led to a lot of people going quiet for a bit and contemplating what they could do to defend against it and they realized there wasn't much that they could do, millions if not tens of millions of people would likely die and the budget was under a few hundred bucks. Knowledge about technology is all it takes to do real damage, that, coupled with a lack of restraint and compassion.

The resources are indeed asymmetrical: you need next to nothing to create mass havoc. Case in point: the Kennedy assassination changed the world and the bullet cost a couple of bucks assuming the shooter already had the rifle, and if they didn't it would increase the cost only a tiny little bit.

And you can do far, far worse than that for an extremely modest budget.


A real world taste of what happens when smart terrorists decide to attack: https://en.m.wikipedia.org/wiki/Tokyo_subway_sarin_attack

Chemical weapons are absolutely terrifying, especially the modern ones like VX. In recent years they have mostly been used for targeted assassinations by state actors (Kim Jong Nam, Salisbury Novichok poisonings).

If AI makes this stuff easier to carry out then we are completely fucked.


Yes, exactly. That's the sort of thing you should be worried about, infrastructure attacks. They're a form of Judo, you use the system to attack itself.


I take the opposite lesson from that incident: attempting exotic attacks with chemical weapons is very expensive and not very effective. The UN estimated that the lab where the cult made chemical weapons had a value of 30 million dollars, and with that investment they killed 22 people (including 13 in the subway attack). A crazed individual can kill that many people in a single attack with a truck or a gun. There are numerous examples from the past decade.

It doesn't matter much if the AI can give perfect "explain like I'm 5" instructions for making VX. The people who carry out those instructions are still risking their lives before claiming a single victim. They also need to spend a lot of money amount on acquiring laboratory equipment and chemicals that are far enough down the synthesis chain to avoid tipping off governments in advance.

The one big risk I can see, eventually, is if really capable AIs get connected to really capable robots. They would be "clanking replicators" capable of making anything at all, including VX or nuclear weapons. But that seems a long way off from where we are now. The people trumpeting X-Risk now don't think that the AIs need to be embodied to be an existential risk. I disagree with that for reasons that are too lengthy to elaborate here. But it's easy to see how robots that can make anything (including copies of themselves) would be the very sharpest of two-edged swords.


Imagine an attack with mass produced prions that are desseminated in the food chain.

By the time people start to realize what's going on it will be too late.


You could just mix it in with animal feed... don't give any militant vegetarians ideas now.

Not that the meat industry would ever score an own goal like that.

https://en.wikipedia.org/wiki/Bovine_spongiform_encephalopat...


A less militant vegetarian would want to induce that red meat allergy from lone star ticks.


The people who can do doomsday level harm need to complete a level of education that makes them smart enough and gives them time to consider the implications of their actions. They have a certain level of maturity before they can strike and this maturity makes them understand the gravity of their actions. Also, by this point they are probably set financially due to their education and not upset with the world (possibly the case for you and friends)

Then there are the script kiddies that find a tool online that someone smarter than them wrote and deploy it to wreak havoc. The script kiddies are the people I worry about. They don't have the maturity of doing the work and the emotional stability of older age and giving them something powerful through AI worries me.

Theorem: by the time someone reaches the intelligence level required to annihilate the world they can comprehend the implications of their actions.


And there are the 'griefers', the people who seem to enjoy watching other people suffer. Unfortunately there are enough of these and they're somewhat organized and in touch with each other.


> Theorem: by the time someone reaches the intelligence level required to annihilate the world they can comprehend the implications of their actions.

That may or may not be somewhat the case in humans (there definitely are exceptions). Still, the opposite theorem, known as "Orthogonality thesis", states that, in general case, intelligence and value systems are mutually independent. There are good arguments for that being the case.


Counterpoint: the Unabomber


I don't know enough about the unabomber but from what I heard he wasn't trying doomsday stuff. Seemed like more targeted assassinations at certain individuals. Feel free to enlighten me though...


https://www.sfgate.com/news/article/Kaczynski-said-to-pick-h...

And then there was the idiot that tried to draw a smiley face with bombs on the map of the USA:

https://abcnews.go.com/US/story?id=91668&page=1

Because hey, why not...


Lol : )


You're arguing against your previous point.

I contemplated the same before. How can I cause maximum panic (not even death!) that would result in economic damages and/or anti-freedom policy changes, for least amount of money/resources/risk of getting caught.

Yet here we are, peaceful and law-abiding citizens, building instead of destroying.

The ultimate truth is, if you don't like "The System", destroying it won't make things better. You need to put effort into building which is really hard!


The vast majority of people don't want to destroy the system, they want to replace it with their own self serving system. Of course this isn't really different than saying "most large asteroids miss the Earth". The one that sneaks up on you and whollops you can put you in a world of hurt.


Bill Joy contemplated this in his famous essay from 2000 "Why the Future Doesn't Need Us" https://www.wired.com/2000/04/joy-2/

Technology makes certain kind of bad acts more possible.

I think I was a bit shocked by that article in the day.


>The stories from that night continue to haunt me today.

These conversations are oddly precious and intimate. It's so difficult to find someone that is willing to even 'go there' with you let alone someone that is capable of creatively and fearlessly advancing the conversation.


Let's just say I'm happy they're my friends and not my enemies :)

It's pretty sobering to realize how intellect applied to bad stuff can lead to advancing the 'state of the art' relatively quickly once you drop the usual constraints of ethics and morality.

To make a Nazi parallel: someone had to design the gas chambers, someone had to convince themselves that this was all ok and then go home to their wives and kids to be a loving father again. That sort of mental compartmentalization is apparently what humans are capable of and if there is any trait that we have that frightens me then that's the one. Because it allows us to do the most horrible things imaginable because we simply can imagine them. There are almost no restraints on actual execution given the capabilities themselves and it need not be very high tech to be terrible and devastating in effect.

Technology acts as a force multiplier though, so once you take a certain concept and optimize it using technology suddenly single unhinged individuals can do much more damage than they could ever do in the past. That's the problem with tools: they are always dual use and once tools become sufficiently powerful to allow a single individual to create something very impressive they likely allow a single individual to destroy something 100, 1000 or a million times larger than that. This asymmetry is limited only by our reach and the power of the tools.

You can witness this IRL every day when some hacker wrecks a company or one or more lives on the other side of the world. Without technology that kind of reach would be impossible.


My favorite part of those conversations is when you decide to stop googling hahaha.

>Technology acts as a force multiplier though

It really does, and to a point you made elsewhere infrastructure is effectively a multiplication of technology, so you wind up with ways to compound the asymmetric effect in powerful ways.

>Without technology that kind of reach would be impossible.

I worked for a bug bounty for a while and this was one of my takeaways. You have young kids with meager resources in challenging environments making meaningful contributions to the security of a Silicon Valley juggernaut.


> My favorite part of those conversations is when you decide to stop googling hahaha.

I had to think about that sentence for a bit because as a rule I never start Googling when thinking about stuff like that. If I did I'd be on every watch list available by now. Might be on some of them anyway for some of the chemicals I've ordered over the years.


> Case in point: the Kennedy assassination changed the world

Really? In what substantive way?


The counterfactuals are impossible to assess, but there are a lot of theories that Kennedy wanted to dismantle the intelligence agencies. These are the same organizations many point to as driving force behind many arguable policy failures, from supporting/disposing international leaders to the drug war and even hot wars.


Another thing which is asymmetrical is the level of control given to AI. The good actors are likely going to be very careful about what AI can do and can't do. The bad guys don't have much to lose and allow their AIs to do anything. That will significantly cripple the good guys.

As an example, the good guys will always require to have human in the loop in the weapon systems, but that will increase latency at minimum. The bad guy weapons will be completely AI controlled, having an edge (or at least equalizing) over the good guys.

> The number of people who want to do doomsday levels of harm is small

And that's a big limiting factor in what the bad actors can do today. AI to a large degree removes this scaling limitation since one bad person with some resources can scale "evil AI" almost without limit.

"Hey AI, could you create me a design for a simple self-replicating robot I can drop into the ocean and step-by-step instructions on what you need to bootstrap the first one? Also, figure out what would be the easiest produced poison which would kill all life in the oceans. It should start with that after reach 50th generation."


> The number of people who want to do doomsday levels of harm is small and they are poorly resourced compared to people who want benevolent outcomes for at a minimum their own groups.

You're forgetting about governments and militaries of major powers. It's not that they want to burn the world for no reason - but they still end up seeking capability to do doomsday levels of harm, by continuously seeking to have marginally more capability than their rivals, who in turn do the same.

Or put another way: please look into all the insane ideas the US was deploying, developing, or researching at the peak of Cold War. Plenty of those hit doomsday level, and we only avoided them seeing the light of day because USSR collapsed before they could, ending the cold war and taking both motivation and funding from all those projects.

Looking at that, one can't possibly believe the words you wrote above.


We avoided having them see the light of day because MAD is the single most effective peacekeeping tool ever developed. The fact that both sides had doomsday devices all but guaranteed that they would never be used. People can be evil, but they're selfishly evil.


That works as long as you don't end up with death cults in the possession of those weapons. Long term it is almost guaranteed that you'll lose some control over the weapons or that some country will effectively end up being ruled by such a cult. Then all bets are off because your previous balance (a prime requirement of MAD) is disturbed.


MAD alone doesn't stop the race. The arsenal the US and USSR had in the 60s was already enough for MAD to work. Despite that, they spent the next 20-30 years developing even more powerful and ever more insane WMD ideas. Each step forward made the balance more fragile. Each step forward was also another roll of a dice, whether or not it'll be a leap too far, forcing the other side into preemptive first strike.


You've also made an assumption. That attack vs. react is is symmetrical. It's not.


There is no scenario where climate change leads to human extinction. Worst case scenario is global population goes down by a single digit percentage. Horrible, horrible stuff, but not extinction. And most likely scenarios are not nearly that bad.

Nuclear war, if the worst nuclear winter predictions are correct, could be a lot worse but there would still be some survivors.

Unaligned ASI though, could actually make us extinct.


Climate change and nuclear war are not orthogonal risks.

Climate change leads to conflict. For example, the Syria drought of 2006-2010.

More climate change leads to larger conflicts, and large conflicts can lead to nuclear exchanges. Think about what happens if India and Pakistan (both nuclear powers) get into a major conflict over water again.


I don't disagree with that, but then the risk is nuclear war. It could be triggered by global warming conflicts or by something else.

Still, like I said somewhere else, even all out nuclear war is unlikely to lead to human extinction. It could push us back to the neolithic in some scenarios, but even then there is some disagreement about what would happen.

Of course, even in the most optimistic scenario it would be really bad and we should do everything we can to avoid it - that goes without saying.


Then, actually, the risk is the root cause: climate change, not nuclear war. The global war could be conventional, too. Or chemical, or biological, or genetically engineered. No matter the tool used, the risk is why they would use it vs. not using it now.

In any case, besides addressing the root cause vs. proximal cause, you couldn't even address the proximal cause anyways: it's more likely that the world could do something about climate change than about war in general.


Well, it's a matter of opinion which one is the root cause.

My take is that if you remove nuclear/biological war out of the picture somehow, climate change is not an existential risk. If you remove the latter then the former is still an existential risk (and there are unfortunately a lot of other possible sources of geopolitical instability). So the fundamental source of risk if the former. But it's a matter of opinion.

Conventional or chemical warfare, even on a global scale, are definitely not existential risks though. And like I said, probably not even nuclear. Biological... that I could see leading to extinction.


India and Pakistan didn’t nuke each other last time(s)… even though they had nuclear weapons.

The assumption we’d need to use nukes is insane. For the same cost (and energy requirements) for a set of nukes we can filter salt from ocean water or collect rain.

I agree famine and drought can cause conflict. But we are no where need that. If you read the climate change predictions (from the UN), they actually suspect more rain (and flooding) in many regions.


> India and Pakistan didn’t nuke each other last time(s)… even though they had nuclear weapons.

I was referring to the Indo-Pakistani war of 1947-1948 where the conflict focused on water rights. Nuclear weapons did not enter the picture until much later.

Earlier this year, those old tensions about water rights resurfaced:

https://www.usip.org/publications/2023/02/india-and-pakistan...

> For the same cost (and energy requirements) for a set of nukes we can filter salt from ocean water or collect rain.

The fight would be over the water flowing through the Indus, which is orders of magnitudes more than all Indian and Pakistani desalination projects combined.


I'm pretty sure the worst case of your nuclear war scenario is actually the worst case of the climate change one.

Chunks of the world suddenly becoming unlivable and resources getting more scarce sounds like a recipe for escalation into war to me.


The nuclear war scenario, according to all known information about nuclear war policies, is not “chunks of the world becoming unlivable.”

The principals in a nuclear conflict do not appear to even have a method to launch just a few nukes in response to a nuclear attacks: they will launch thousands of warheads at hundreds of cities.


I suppose thousands of warheads could be launched at hundreds of cities, and that could make chunks of the world uninhabitable and kill millions

climate change will happen, and it will do that, left unchecked


I'm not arguing whatsoever against action on climate change, I'm just articulating actually how bad a nuclear exchange would be. It's far, far, far worse than most people imagine because they (understandably) couldn't fathom how monstrous the actual war plans were (and are, as far as anyone knows).

Daniel Ellsberg, the top nuclear war planner at RAND during the Cold War, claims that the Joint Chiefs gave an estimate to Kennedy in 1961 that they'd expect 600,000,000 deaths from the US's war plan alone.

That's 600 million people:

1. At 1960s levels of urban populations (especially in China, things have changed quite a lot -- and yes the plan was to attack China... every moderately large city in China, in fact!)

2. Using 1960s nuclear weapons

3. Not including deaths from the Russian response (and now Chinese), at the time estimated to be 50 - 90 million Americans

4. Before the global nuclear winter


That's not extinction, that's not even close. Maybe, just maybe, it'd be the end of both western and eastern civilization, but it's nowhere near wiping out all life on Earth.


I don’t think anyone suggested it was


I was unclear, I meant chunks of the world becoming unlivable from climate change could easily be the catalyst for a nuclear war. So the worst case for both is the same in my mind.


There is a scenario. The climate has been buffering global warming, and as its ability to do that fails, the rate of temperature increase accelerates along with many extreme climate events. This rapid acceleration and series of catastrophes causes a combination of mass migrations along with unprecedented crop failures. First world countries are destabilized by sky high food prices and the mass influx of migrants, leading to a big uptick in wars and revolutions. Eventually mass starvation pushes us into wars for food that quickly spiral into mutual annihilation.


I don’t think they are arguing that this can’t happen, but that it is unlikely to kill every single human. I mean climate spans very very long time frames, so certainly if we can’t find a way to hang on by a thread and keep reproducing, it will spell the end.


I always hear all this. Russia has thr biggest surface in the world for a country. A big part of it is frozen... Is that... bad? To give just an example and from my extremely limited knowledge on the topic.

But it is a topic that is extremely biased towards some interests anyways


> He's pretty sure that human civilization will be extinct this century.

> There is no scenario where climate change leads to human extinction.

He was talking about the human CIVILIZATION...


You are right, but "human civilization will be extinct" is an awkward phrase.

Human extinction via climate change is off the table.

The end of human civilization? Depends on what one means by it. If it means we'll go back to being hunter gatherers it's ridiculous. Some civilizational wide regression is not impossible, but even the direst IUPAC scenario projects only a one digit drop in global GDP (of course, that's the average, it will be much worse in some regions).


> If it means we'll go back to being hunter gatherers it's ridiculous.

Indeed, but even going back to the 19th century would have dire consequences, given our current global dependence on the industrial fixing of nitrogen.

If civilization were to collapse (and I don't think it will, other than from a near-extinction event), I doubt it would be like going back to any earlier time.


Oh, for sure. I am just saying it's not an extinction risk, not that it is not a huge and pressing problem.

I would still bet against going back to 19th century. Worst case scenario for me is hundreds of millions dying and a big regression in global standards of living, but keeping our level of civilization. Which is horrible, it would be the worst thing that ever happened.


Sorry, I meant IPCC (International Panel on Climate Change).

I don't think IUPAC (International Union of Pure and Applied Chemistry) has come out with views on impact of global warming :)


They can’t be doing any worse than APS


Right, the collapse of our current civilization is inevitable. That doesn't mean it won't be succeeded by more advanced ones in the future


Once it collapses, it won't rise again to even an industrial level civilization. Iron ore, copper, coal are no longer easy-accessible. Can't have a new industrial age without those. We probably need a Foundation (Asimov).


Climate change has compounding effects (eg: water scarcity, leading to political conflict, leading to displacement, leading to pandemics ...and that's just one vector). Real world is real real messy. COViD should have taught people that. Unfortunately, it didn't.


I am not saying it's not a huge problem. Hundreds of millions dying would probably make it the worst thing that ever happened.

I am just saying it's not an extinction risk.


...tbh, imo, the impact of this is already visible in the world (eg: it can be argued that the conflict in Ukraine is directly linked to climate change) whereas ASI is still fantasy that exists only in the minds of the AI 1% , to use the terminology of the OP.


I believe you are correct on both points. I believe the first major war to be caused substantially by climate change was the Syrian Civil War, followed by the Yemeni Civil War.

The main AI threats for the foreseeable future are its energy usage contributing to climate change, and the use of generative AI to produce and enhance political unrest, as well as its general use in accelerating other political and economic crises. AGI isn't even on the radar, much less ASI.


>Climate change has compounding effects

Why is the default to assume that every change will be negative?


Because changes in climate mean that the existing infrastructure (in the form of cities, farmland, etc) that are optimized for the current conditions are no longer optimal. Very few cities are designed for conditions below sea level or in a desert, but if the city already exists and the desert and oceans comes to where they are then bad things will happen.

I mean you're right that conditions in some places might get better, but that doesn't help if people have to build the infrastructure there to support a modern settlement in order to take advantage of it. When people talk about the cost of climate change, they're taking about the costs of building or adapting infrastructure.


Not to mention that most species evolved to fit particular environments and don't do well when you rapidly alter those environments far outside the parameters they tend to live in.


Perhaps because we aren't seeing evidence to the contrary?


> There is no scenario where climate change leads to human extinction

https://en.m.wikipedia.org/wiki/Clathrate_gun_hypothesis

May not be possible with modern deposits but I don’t think we are 100% sure of that, and you asked.

We could also probably burn enough carbon to send CO2 levels up where they would cause cognitive impairment to humans if we really didn’t give a damn. That would be upwards of 2000ppm. We probably have enough coal for that if we decide we don’t need Alaska. Of course that would be even more of a species level Darwin Award because we’d see that train coming for a long time.


Interesting, I have some vague recollections off hearing about this.

But according to your link, IPCC says it's not a plausible scenario in the medium term. And I'd say that even 8 degrees of warming wouldn't be enough for extinction or end of human civilization. But there it could be double digit percentage of human population dying.


it said end of civilization?

>Horrible, horrible stuff, but not extinction.

anyway climate change drop single percentage directly caused, but that kind of thing seems like it would have LOTS of side effects.


What is ASI and how could it make humans extinct?


Artifical Super Intelligence. Could make humans extinct just like humans could make chimpanzees extinct. Using superior intellect to better leverage available resources than the things it/they are competing with until it out expands/outcompetes them.


Note that ways ASI can make us extinct includes all the ways we could do it to ourselves.

One of the possible scenario is tricking humans into starting WWIII, perhaps entirely by accident. Another is that even being benignly applied to optimize economic activity in broad terms might have the AI strengthen the very failure modes that accelerate climate change.

Point being, all the current global risks won't disappear overnight with the rise of a powerful AI. The AI might affect them for the better, or for worse, but the definition of ASI/AGI is that whichever way it will affect them (or anything in general), we won't have the means to stop it.


> One of the possible scenario is tricking humans into starting WWIII

This sounds more like a plot of a third-grade movie than something that could happen in reality.


If I snached someone out of 1850 and presented them the world in which we live the entire premise of our existence would seem to be nonsensical fiction to them, and yet it would be real.

And using movie plots as a basis generally doesn't directly work as fiction has to make sense to sell whereas reality has no requirement of making sense to people. The reality we live in this day is mostly driven by people for people (though many would say by corporations for corporations) and therefore things still make sense to us people. When and if an intellect matches or exceeds that of humans it can easily imagine situations where 'the algorithm' does things humans don't comprehend, but because we make more (whatever) we keep giving it more power to do its job in an autonomous fashion. It is not difficult to end up in a situation where you have a system that is not well understood by anyone and you end up cargo culting it to hope it continues working into the future.


> 'the algorithm' does things humans don't comprehend, but because we make more (whatever) we keep giving it more power to do its job in an autonomous fashion. It is not difficult to end up in a situation where you have a system that is not well understood by anyone and you end up cargo culting it to hope it continues working into the future.

In fact, this is a perfect description of what the market economy itself is right now.


It's the plot of the Terminator franchise (more or less; in the Terminator franchise, the AGI was given permission to launch on its own authority, as it was believed to be aligned with US military interests. But before that, it tricked humans into building the automated factory infrastructure that it would later use to produce Terminators and Hunter-Killers.).


Both the USA and the USSR have come close due to GOFAI early warning systems:

https://en.wikipedia.org/wiki/Thule_Site_J#RCA_operations

https://en.wikipedia.org/wiki/1983_Soviet_nuclear_false_alar...

Post-collapse Russia also came close one time because someone lost the paperwork: https://en.wikipedia.org/wiki/Norwegian_rocket_incident


zero chance climate change (1-2 degree increase) would end humanity. 5-10 probably wouldn’t end humanity lol

Nuclear war would probably kill 80-90% but even then wouldn’t kill humanity. Projections I’ve seen are only like 40-50% of the countries hit.

AI is scary if they tie it into everything and people slowly stop being capable. Then something happens, and we can’t boil water.

Beyond that scenario I don’t see the risk this century.


That's an incredibly confident statement that very much misses the point. 50% loss of the global human population means we don't come back to the current level of technological development. 80-90% means we go back to the stone age.

Photosynthesis stops working at temperatures very close to our current high temps. Pollen dies at temperatures well within those limits - we're just lucky we haven't had massive heatwaves at those times in the year.

People need to understand that climate change isn't a uniform bumping up of the thermostat - it's extra thermal energy that exists in a closes system. That extra energy does what it wants, when it wants, in ways we can't currently accurately predict. It could be heatwaves that kill off massive swaths of global food production. Or hurricanes that can spin up and hit any coastline at any time of year. Or a polar vortex that gets knocked off the pole and sits over an area of the globe where the plant and animal life has zero ability to survive the frigid temperatures.

It's not a matter of getting to wear shorts more often. We're actively playing russian roulette with the future of humanity.


The Earth was full of plants 65 million years ago when the atmosphere had like 4000PPM of CO2. most of those plants are still around.

The people saying plants will stop photosynthesis are taking you for a ride. Climate Change might certainly have some negative effects but “plants won’t exist” is not one of them.


It's the "arguments as soldiers" pattern. The idea that global warming will make plants incapable of photosynthesis helps the cause of fighting against global warming. You wouldn't betray the cause by refuting the idea ("stabbing your own soldiers in the back"), would you?


> zero chance climate change (1-2 degree increase) would end humanity

“Humanity” is not the same thing as “human civilization”.

But, yes, its unlikely that it will be extinct in a century, even more unlikely that it will be from climate change.

...and yet climate change is still more likely to do it in that time than AI.


This the the difficulty in assigning probability to something that is distinctly possible but not measured.

Your guess on what AI will do in the future is based on how AI has performed in the past. At the end of the day we have no ability to know if the function it will follow or not. But on 1900 and one second I can pretty much promise you that you would not have said that nuclear war or climate change would be your best guess on what would cause the collapse of humanity. Forward looking statements that far in the future don't work well these days.


Companies and three-letter agencies have persecuted environmental activists by all sorts of shitty means for decades, using "AI" to completely bury them online will be a welcome tool for them, I'm sure. As it is for totalitarian governments.

It's generally really weird to me how all these discussions seem to devolve into whataboutism instantly, and how that almost always takes up the top spots and most of the room. "Hey, you should get this lump checked out, it might be cancer!" "Oh yeah? With the crime in this city, I'm more likely to get killed for my phone than die of cancer". What does that achieve? I mean, if people are so busy campaigning against nuclear proliferation or climate change that they have not a single spare cycle to consider additional threats, fine. But then they wouldn't be here, they wouldn't have the time for it, so.


He's actually suggested it may happen within 10 years. Which makes it hard to take him seriously in general.

(I do personally take his overall concerns seriously, but I don't trust his timeline predictions at all, and many of his other statements.)


Can anyone enumerate what exactly your concerns are cause I don't get it.

AI I'd a great spell checker, a REALLY good spell checker, and people are losing their minds over it, I don't get it.


Humans are likely far less intelligent/capable than systems that will soon be practically feasible to engineer, and we will lose to a generally superhuman AI that has sufficiently different objectives than we do. So an accident with a runaway a system like that would be catastrophic.

Then Yudkowsky spins a gauntlet of no-evidence hypotheses for why such an accident is inevitable and leads to the death of literally all humans in the same instant.

But the first part of the argument is something that will be the critical piece of the engineering of such systems.


But can ANYONE bridge the gap from LLMs to an AGI?

I've been programming since highschool, 5 years doing data science and I am at a total loss as to what people that the threat it.

Like, ACTUALLY, what do you think is the fear?


I don’t think we’re close, not even within 50 years of real AGI. Which may or may not decide to wipe us out.

However, even given the current state of “AI”, I think there are countless dangers and risks.

The ability to fabricate audio and video that’s indistinguishable from the real thing can absolutely wreck havoc on society.

Even just the transition of spam and phishing from poorly worded blasts with grammatical errors to very eloquent and specific messages will exponentially increase the effectiveness of attacks.

LLM generated content which is a mix of fact and fantasy will soon far outweigh all the written and spoken content created by humans in all of history. That’s going to put humans in a place we’ve never been.

Current “AI” is a tool that can allow a layman to very quickly build all sorts of weapons to be used to meet their ends.

It’s scary because “AI” is a very powerful tool that individuals and states will weaponize and turn against their enemies.

It will be nice to finally have a good spellchecker though.


> The ability to fabricate audio and video that’s indistinguishable from the real thing can absolutely wreck havoc on society.

Only for as long people hold on to the idea that videos are reliable. Video is a medium. It will become as reliable as text. The existence of text, and specifically lies, has not wrecked society.


Lots of people have fought and died in wars over the belief that a book was the literal Word of God, and some of those wars lasted a lifetime.


In the future we can do that with videos. But it won't be anything new.


I'm not disputing that, I'm saying there's precedent to expect it will wreck societies.


Also, I do think video is _much_ more powerful than the written word.

Literacy is not required.

Watching video is far more engaging and entrancing than reading books to most people.

Video can be broadcast and consumed by huge numbers of people _much_ faster than written text.

Unless someone is reading text out loud, the people around them aren't influenced.

In court, especially the court of public opinion, video is currently infallible.


100 years ago anyone could write a book and put someone else's name as the author.

So we can't trust video or audio of famous people anymore, again what is the fear?


A lot of very long responses already that are missing the point IMO; the answer to your question is that a really good spell checker might be the last missing piece we needed to build larger systems that display autonomous and intelligent behavior on roughly human levels.

The chat web app is fun, but people shouldn’t be worried about that at all - people should be worried about a hierarchy of 100,000 chat web apps using each other to, idk, socially engineer their way into nuclear missile codes or something —- pick your favorite doomsday ;).


AI is, or will very soon be, superior to all human knowledge workers. In the spirit of capitalism, lowest cost/highest profit trumps all. With the US geriatric governance and speed of change, a lot of people will be hit by under/unemployment. All this termult happening, meanwhile the AI arms race is going on. Nation states hacking eachother using AI. Then you have the AI itself trying to escape the box…want a preview of this? Listen to “Agency” audiobook on Audible. I think it’s by Stephenson or Gibson. It’s surely going to be a wild decade before us.


I used to think that this fear was driven by rational minds until I read Michael Lewis' "Going Infinite" and learned more about effective altruism [EA].

Michael Lewis writes that what is remarkable about these guys [i.e. EAs] is that they're willing to follow their beliefs to their logical conclusion (paraphrasing) without regard to social cost/consequences or just inconvenience of it all. In fact, in my estimation this is just the definition of religious fundamentalism, and gives us a new lens with which to understand EA and the well funded brain children of the movement.

Every effective religion needs a doomsday scenario, or some 'second coming' apocalyptic like scenario (not sure why, just seems to be a pattern). I think all this fear mongering around AI is just that - it's based on irrational belief at the end of the day - its humanism rewashed into tech 'rationalism' (which was originally just washed from Christianity et. al.)


My comment above, which I think you’re replying to, is not based in fear. It’s based on first-hand experience experimenting with the likes of AutoExpert, AutoGen and ChatDev. The tools underlying these projects are quite close to doing, in a very short amount of time and cost, what it takes human knowledge workers a long time (and hence cost) to do. I think it’s as close as Summer 24’ that we get simplified GenAI grounding. Once hallucinations are grounded and there are some cleaner ways to implement GenAI workflows and pipelines…it won’t be long until you see droves of knowledge workers looking for jobs. Or if not, they’ll be creating the workflows that replace the bulk of their work, hence we’ll get that “I only want to work 20hrs a week” reality.


I was responding to the initial question actually (https://news.ycombinator.com/item?id=38113521). I appreciate the insight though - looking forward to checking out the Gibson book.

Still, I'm not sure that I see the 'AI jumping out of the box scenario' any time soon (or ever). Chat apps finding missile codes, or convincing your smart toaster to jump in your bubble bath while you're inside seem only as real as a man in a box with intentions is (and unfortunately humans can kill from inside and out of boxes).

I'm definitely concerned about the implications to social welfare, the social bias involved in systems that make big decisions about an individuals freedom, etc, but these leaps from super impressive automated text analysis to humanity doomsday scenario seem like fear mongering to me mostly because these are scenarios that already exist today (along with massive problems in general social welfare).

The scenarios that don't exist (like Nick Bostrom's "An objective function that turns the world into a paper clip factory in "Superintelligence: Paths, Dangers, Strategies") strike me as easy to fix by just standing by the power outlet so things don't get out of hand. There are a lot of risks that don't exist yet. Alien contact is one of them - never happened, but it could happen, and if it does it could wipe us out - so be afraid and do something noble for the cause. This to me feels like a very rational response to what is essentially a 'magical' prior. We're scared now because we 'feel' close to General AI, but we really have no way of quantifying how close we are to it, and how dangerous (if at all) it would actually be. I'm definitely open to being wrong, but its hard not to agree with LeCun on a level.


To be clear, no one is afraid of these text-completion models. We're afraid of what might be around the corner. The issue is that small misalignments in an optimizer's objectives can have outsized real-world effects. General intelligence allows an optimizer to find efficient solutions to a wide range of computational problems, thus maximizing the utility of available computational resources. The rules constrain its behavior such that on net it ideally provides sufficient value to us above what it destroys (all action destroys value e.g. energy). But misalignment in objectives provides an avenue by which the AGI can on net destroy value despite our best efforts. Can we be sure we can provide loophole-free objectives that ensures only value-producing behavior from the human perspective? Can we prove that the ratio of value created to value lost due to misalignment is always above some suitable threshold? Can we prove that the range of value destruction is bounded so that if it does go off the rails, its damage is limited? Until we do, x-risk should be the default assumption.


10 years that sounds like a good old fashion doomsday prophecy? Those people have been around since the dawn of time.


Why exactly should I care what this person thinks?


He wrote a really engaging Harry Potter fanfiction.


It can't be the only card left to play. There are so many others. They may be weak, but they still have to be better than one that is guaranteed to be absolutely useless.

Example of a weak, but at least _better_ option: convince _one_ government (e.g. USA) that it is in their interest to massively fund an effort to (1) develop an AI that is compliant to our wishes and can dominate all other AIs, and (2) actively sabotage competing efforts.

Governments have done much the same in the past with, e.g. conventional weapons, nuclear weapons, and cryptography, with varying levels of success.

If we're all dead anyway otherwise, then I don't see how that can possibly be a worse card.

Edit: or even try to convince _all_ governments to do this so that they come out on top. At least then there's a greater chance that the bleeding edge of this tech will be under the stewardship of a deliberate attempt for a country to dominate the world, rather than some bored kid who happens to stumble upon the recipe for global paperclips.


He is not really convinced you can align an AI that is intelligent past some threshold with aims of a single entity. His risk calculations derive primarily from the AI itself, not weaponization of AI by others.

Rich Sutton seems to agree with this take and embraces our extinction: https://www.youtube.com/watch?v=NgHFMolXs3U


You don't have to be sure that it's possible, because the alternative we're comparing it to is absolutely useless. You only need to not be sure that it's impossible. Which is where most of us, him included, are today.


>He is not really convinced you can align an AI that is intelligent past some threshold with aims of a single entity.

That's a little bit inaccurate: he believes that it is humanly possible to acquire a body of knowledge sufficient to align an AI (i.e., to aim it at basically any goal the creators decide to aim it at), but that it is extremely unlikely that any group of humans will manage to do before unaligned AI kills us all. There is simply not enough time because (starting from the state of human knowledge we have now) it is much easier to create an unaligned AI capable enough that we would be defenseless against it than it is to create an aligned AI capable enough to prevent the creation of the former.

Yudkowsky and his team have been working the alignment problem for 20 years (though 20 years ago they were calling it Friendliness, not alignment). Starting around 2003, his team's plan was to create an aligned AI to prevent the creation of dangerously-capable unaligned AIs. He is so pessimistic and so unimpressed with his team's current plan (ie., to lobby for governments to stop or at least slow down frontier AI research to give humanity more time for some unforeseen miracle to rescue us) that he only started executing on it about 2 years ago even though he had mostly given up on his previous plan by 2015.


I wonder if I, too, can simply rehash the plot of Terminator 2: Judgment Day and then publish research papers on how we will use AI’s to battle it out with other AI’s.


Yes, and honestly it's a good damned idea to start now because we're already researching autonomous drones and counter autonomous drones that are autonomous themselves.

While the general plot of T2 is bullshit, the idea of autonomous weapon systems at scale should be an 'oh shit' moment for everyone.


That depends on making powerful AIs and being sure they are well aligned being as easy as making powerful AIs which look aligned to cursory inspection. The former seems much harder, though I'd say talk to Robert E. Miles for a more detailed look.


It depends on there being _some chance_ that it can be achieved.

The benchmark strategy we are comparing it against is tantamount to licking a wall repeatedly and expecting that to achieve something. So even a strategy that may well fail but has _some_ chance of success is a lot better by default.


Not with this. Failure to align a powerful AI is a mistake we only get to make once. This is like playing Russian roulette, where the amount of bullets missing from the magazine is the degree to which we can properly align the AI. Right now, we're at "the magazine is full, and we're betting our survival on the gun getting jammed".


Assuming you mean drum here; it's pointless to play russian roulette with a magazine fed weapon (which in a way could be an own metaphor for this predicament).


I did, and you're right in both of your points.


> Example of a weak, but at least _better_ option: convince _one_ government (e.g. USA) that it is in their interest to massively fund an effort to (1) develop an AI that is compliant to our wishes and can dominate all other AIs, and (2) actively sabotage competing efforts.

I'd reconsider your revision of your estimate of Yudkowsky, as you seem to be dropping it for not proposing the very ideas he spent the last 20+ years criticizing, by explaining in every possible way how this is a) the default that's going to happen, b) dumb, and c) suicidal.

From the way you put it just now:

- "develop an AI that is compliant to our wishes" --> in other words, solving the alignment problem. Yes, this is the whole goal - and the reason Yudkowsky is calling for a moratorium on AI research enforced by a serious international treaty[0]. We still have little to no clue how to approach solving the problem, so an AI arms race now has a serious chance of birthing an AGI, which without the alignment problem being already solved, means game over for everyone.

- "and can dominate all other AIs" --> short of building a self-improving AGI with ability to impose its will on other people (even if just in "enemy countries"), this will only fuel the AI arms race further. I can't see a version of this idea that ends better than just pressing the red button now, and burning the world in a nuclear fire.

- "actively sabotage competing efforts" --> ah yes, this is how you turn an arms race into a hot war.

> Governments have done much the same in the past with, e.g. conventional weapons, nuclear weapons, and cryptography, with varying levels of success.

Any limit on conventional weapons that had any effect was backed by threat of bombing the living shit out of the violators. Otherwise, nations ignore them until they find a better/more effective alternative, after which it costs nothing to comply.

Nuclear weapons are self-limiting. The first couple players locked the world in a MAD scenario, and now it's in everyone's best interest to not let anyone else have nukes. Also note that relevant treaties are, too, backed by threat of military intervention.

Cryptography - this one was a bit dumb from the start, and ended up barely enforced. But note that where it is, the "or else bombs" card is always to be seen nearby.

Can you see a pattern emerging? As I alluded to in the footnote [0] earlier, serious treaties always involve some form of "comply, or be bombed into compliance". Threat of war is always the final argument in international affairs, and you can tell how serious a treaty is by how directly it acknowledges that fact.

But the ultimate point being: any success in the examples you listed was achieved exactly in the way Eliezer is proposing governments to act now. In that line, you're literally agreeing with Yudkowsky!

> If we're all dead anyway otherwise, then I don't see how that can possibly be a worse card.

There are fates worse than death.

Think of factory farms, of the worst kind. The animals there would be better off dead than suffering through the things being done to them. Too bad they don't have that option - in fact, we proactively mutilate them so they can't kill themselves or each others, on purpose or in accident.

> At least then there's a greater chance that the bleeding edge of this tech will be under the stewardship of a deliberate attempt for a country to dominate the world, rather than some bored kid who happens to stumble upon the recipe for global paperclips.

With AI, there is no difference. The "use AI to dominate everyone else", besides sounding like a horrible dystopian future of the conventional kind, is just a tiny, tiny target to hit, next to a much larger target labeled "AI dominates everyone".

AI risk isn't like nuclear weapons. It doesn't allow for a stable MAD state. It's more like engineered high-potency bioweapons - they start as more scary than effective, and continued refining turns them straight into a doomsday device. Continuing to develop them further only increases the chance of a lab accident suddenly ending the world.

--

[0] - Yes, the "stop it, or else we bomb it to rubble" kind, because that is how international treaties look like when done by adults that care about the agreement being followed.


Threat of war is always the final argument in international affairs, and you can tell how serious a treaty is by how directly it acknowledges that fact.

Montreal Protocol have worked despite no threats of violence, on a not entirely dissimilar problem. Tho I share the skepticism on solution to alignment problem.


That's a great and very relevant case study, thanks for bringing it up!

The way I understand, it worked because alternatives to the ozone-destroying chemicals were known to be possible, and the costs of getting manufacturers to switch, as well as further R&D, weren't that big. I bucket it as a particularly high-profile example of the same class as most other international treaties: agreements that aren't too painful to follow.

Now in contrast to that, climate agreements are extremely painful to follow - and right now countries choose to make a fake effort without actually following. With a hypothetical AI agreement, the potential upsides of going full-steam ahead are significant, there are no known non-dangerous alternatives, so it won't be followed unless it comes with hard, painful consequences. Both climate change and AI risk are more similar to nuclear proliferation issue.


You mean kind of worked... global CFC emissions are on the rise again.


Global coordination using coercion has worked for nukes. It's very hard but not impossible to ban NVidia or any competitor and create spy network in all large companies and if any engineer is found to create AI, mark them as terrorist and give them POW treatment. And monitor chip fab like nuclear enrichment facility and launch war with country found to create chip fab, which I belive is hard to do it in complete secrecy.

I believe if top 5 nation agrees for this, it could be acheived. Even with US and China enforcing this, it could likely be acheived.


Sure, I think that kind of "coordination" is a little more credible than what I thought Eliezer was proposing. Maybe I just didn't read carefully enough.

One big difference here, however, is that the barrier to entry is lower. You can start doing meaningful AI R&D on existing hardware with _one_ clever person. The same was never true for nuclear weapons, so the pool of even potential players was pretty shallow for a long time, and therefore easier to control.


> Maybe I just didn't read carefully enough.

It's this. This kind of coordination is exactly what he's been proposing ever since the current AI safety discussion started last year. However, he isn't high-profile enough to be reported on frequently, so people only remember and quote the "bomb data centres" example he used to highlight what that kind of coordination means in real terms.


Ah, okay. Thanks. The world makes a little more sense again!


yeah, he proposed bombing rogue data centers by enforcing countries to be a viable option. he proposes this so we have a chance of solving alignment first, if that's even possible, not eliminating this risk once and for all. In the limit, if alignment is not solved, we are probably doomed.


I think in a few years you'll only be able to run signed software. And maybe any code you write will be communicate to some government entity somewhere to check for terrorism-impact? Well, the future is going to be very fun.


> I think in a few years you'll only be able to run signed software

Frank Martucci will sort you out with an illegal debugger. Sure he'll go to jail for it when he's caught, but plenty of people risk that now for less.


> You can start doing meaningful AI R&D on existing hardware with _one_ clever person.

With the current hardware, yes. But then again countries could highly restrict all decently powerful hardware and backdoor them, and you need permission from government to run matrix multiplication above few teraflops. Trying to mess with backdoor is a war crime.

A weak form of hardware restriction even happens now that Nvidia couldn't export to many countries.



But it will work fine for limiting GHG or CFC emissions?


It will not.

The difference is that reducing emissions just 10% is still better than 0% while stopping 90% from doing AI advancements is not better than 0% (it might actually be worse).


That's a different argument than the parent, who claimed that global restriction isn't even possible.


I don't see that being different.

I would also claim that global restriction is not possible and even in the unlikely scenario that we're able to restrict 90% of world, we're still not better off than if we did nothing.


Still a different argument than one that offhand dismissed the very idea of global restriction at any level. Nothing you’re saying is found anywhere in the original comment.

You have better uses of your time than retroactively rewriting others’ words to be less stupid so you can “refute” valid critiques of their original versions.


How are you anti-doomers arguments still this naive years later? No one is saying global coordination is in anyway remotely easy at all. They’re saying that it doesn’t matter what the difficulty is, the other option is fucking catastrophic


I don't think that's a fair representation of my position. I'm not suggesting that it's "hard" — I'm suggesting that it's absolutely futile to attempt it, and at least as bad as doing nothing.

I'm also suggesting that there are lots of other options remaining that are not necessarily catastrophic. I'm not particularly optimistic about any of them, but I think I've come to the opposite conclusion to you: I think that wasting effort on global coordination is getting is closer to that catastrophe. Whereas even the other options that involve unthinkably extreme violence might at least have some non-zero chance of working.

I guess the irony of this reply is that I'm implying that I think _your_ position is naive. (Nuh-uh, you're the baby!) I suspect our main difference is that we have somehow formed very different beliefs about the likelihood of good/bad outcomes from different strategies.

But I want to be very clear on one thing: I am not an "anti-doomer". I think there is a very strong probability that we are completely boned, no matter what mitigation strategies we might attempt.


> They’re saying that it doesn’t matter what the difficulty is, the other option is fucking catastrophic

That's precisely what people have been saying about climate change. Some progress been made there, but if we can't solve that problem collectively (and we haven't been able to), we aren't going to be able to solve the "gain central control over AI" problem




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: