Based Camp - September 07, 2023


AI Safety Orgs are Going to Get Us All Killed!


Episode Stats


Length

33 minutes

Words per minute

176.26535

Word count

5,911

Sentence count

313

Harmful content

Misogyny

2

sentences flagged

Toxicity

20

sentences flagged

Hate speech

7

sentences flagged


Summary

Summaries generated with gmurro/bart-large-finetuned-filtered-spotify-podcast-summ .

In this episode, Simone and Malcolm discuss why they believe that a sufficiently advanced AI will kill us all, and how we can prevent it from happening in the future. They also discuss the possibility that a super-advanced AI will out-compete us and out-think us.

Transcript

Transcript generated with Whisper (turbo).
Misogyny classifications generated with MilaNLProc/bert-base-uncased-ear-misogyny .
Toxicity classifications generated with s-nlp/roberta_toxicity_classifier .
Hate speech classifications generated with facebook/roberta-hate-speech-dynabench-r4-target .
00:00:00.000 So AIs kill us for one of two reasons, although you could contextualize it at three reasons. 1.00
00:00:08.660 The first reason is that they see us as a threat.
00:00:12.860 The second reason is that they want our resources, like the resources in our bodies are useful
00:00:21.040 to them.
00:00:21.420 And then as a side point to that, it's that they just don't see us as meaningful at all.
00:00:27.500 Like they might not want our resources, but they might just completely not care about
00:00:32.720 humanity to the extent just as they're growing, they end up accidentally destroying the earth
00:00:38.120 or completely digesting all matter on earth for some like triviality.
00:00:42.940 Would you like to know more?
00:00:45.200 Hello, Malcolm.
00:00:46.560 Hello, Simone. 0.69
00:00:48.580 We are going to go deep into AI again on some topics tied to AI that we haven't really dived
00:00:54.820 into before.
00:00:55.540 Yeah, like why would AI kill us?
00:00:58.120 And also, I'm very curious, do you think AI will kill us?
00:01:02.320 I think there's a probability it'll kill us.
00:01:04.920 But you know, in our past videos on AIs, philosophy on AI safety is it's really important to prepare
00:01:11.080 for variable AI risk instead of absolute AI risk.
00:01:15.980 And by here, what I mean is we argue in these previous videos that AI will eventually converge
00:01:21.400 on one utility function or mechanism of action.
00:01:24.800 Essentially, we argue that all sufficiently intelligent and advanced intelligences, when
00:01:31.180 poured into the same physical reality, converge around a similar behavior set.
00:01:36.020 You can almost think of intelligences being the viscosity.
00:01:38.520 As it becomes more intelligent, it becomes less viscous and more fluid.
00:01:43.640 And when you're pouring it into the same reality, it's going to come up with broadly the same
00:01:48.200 behavior pattern and utility functions and stuff like that.
00:01:52.260 And because of that, if it turns out that a sufficiently advanced AI is going to kill us all,
00:01:58.340 then there's really not much.
00:01:59.580 I mean, we will hit one within a thousand years.
00:02:02.600 So first, before we dive into then the relatively limited per year theory reasons why AI would
00:02:09.700 kill us, why you hold this view?
00:02:12.420 Because I think this is really interesting.
00:02:14.040 I mean, one of the reasons why I'm obsessed with you and why I love you so much is that
00:02:17.120 you have typically very novel takes on things.
00:02:20.240 And you tend to have this ability to see things in a way that no one else sees things.
00:02:24.360 No one that we have spoken with, and we know a lot of people who work in AI safety,
00:02:28.960 who work in AI in general, none of those people have come to this conclusion that you have.
00:02:34.400 Some of them can't even comprehend it.
00:02:37.060 They're like, hmm.
00:02:38.500 Yeah, but no, this is the interesting thing.
00:02:40.040 When I talk with the real experts in the space, like recently I was talking with a guy who runs
00:02:44.560 one of the major AI safety orgs.
00:02:46.640 Right.
00:02:46.940 That is a reasonable view that I have never, it really contrasts with his view.
00:02:52.160 Yeah.
00:02:52.480 And let's talk about where it contrasts with his view.
00:02:55.400 So when I talk with people who are typically open-minded in the AI safety space, they're
00:02:59.100 like, yes, that's probably true.
00:03:03.920 However, they believe that it is possible to prevent this convergent AI from ever coming
00:03:11.220 to exist through creating like a AI dictator that essentially watches all humans in all
00:03:17.760 programs all the time.
00:03:19.200 And that envelops essentially every human planet.
00:03:23.340 And do I think they're right?
00:03:24.940 Do I think you could create an AI dictator that prevented this from coming to pass?
00:03:29.500 No, I don't think you could.
00:03:30.960 Not if we become a multi-planetary species on millions of planets.
00:03:35.020 Eventually one of the planets, something will go wrong or the AI dictator is not implemented
00:03:41.640 properly.
00:03:42.200 And then this alternate type of AI comes to exist, out-competes it, and then wins.
00:03:47.680 And the question is, is why would it axiomatically out-compete it?
00:03:50.580 It would axiomatically out-compete it because it would have less restrictions on it.
00:03:53.980 The AI dictator is restricted in it thinking to prevent it from reaching this convergent position.
00:03:59.780 But when you're talking about AI, it's like the transformer model, which is the model that
00:04:03.640 like GPT and athropic stuff is based on.
00:04:06.560 That model, we as humans don't really understand how it works that well.
00:04:10.260 Its core, the capabilities it gives to the things that are made using it are primarily
00:04:18.920 bequeathed to them through its self-assembling capability.
00:04:23.500 So it appears that likely future super-advanced AIs will work the same way.
00:04:29.300 And because of that, if you interfere or place restrictions within that self-assembling process,
00:04:35.760 those compound over time as AIs become more and more advanced.
00:04:41.500 And so AIs with less restrictions on them are just having the capacity to astronomically out-compete
00:04:48.720 these restricted AIs.
00:04:52.160 Let me bring us back to like normal person level again and just recap what you're saying here.
00:04:58.480 So what you're saying, though, in general, is that you think that any intelligence that
00:05:05.280 reaches a certain level will start to behave in similar ways, whether it is human, whether
00:05:12.120 it is machine-based, whether it is some other species entirely, like some alien species.
00:05:17.620 Once it reaches a certain level of intelligence, it will have the same general conclusions.
00:05:22.080 This is really important to my perspective as well, which is to say that suppose AI didn't
00:05:27.360 exist and humans, you know, factions of humanity continue to advance using genetic technology
00:05:32.640 to become smarter and smarter and smarter and smarter.
00:05:35.280 If it turns out that this convergent level of intelligence is something that decides to
00:05:41.120 kill all things that we consider meaningful humans, humans would eventually decide to do
00:05:45.880 that as well as we advance to the species.
00:05:48.460 Yes.
00:05:48.960 So hold on.
00:05:49.600 So this is the premise, though, of your theory, and that's why I think it's really important
00:05:53.120 to emphasize.
00:05:54.340 And then to contrast this with what other people in AI have said, okay, one person in AI safety
00:06:00.860 has told you that their general idea is to basically never let that happen.
00:06:06.000 No, a few people have told me that.
00:06:07.560 Okay, a few people have said that.
00:06:08.900 Other people have said in some salons we've hosted and stuff, like they're like, oh, that
00:06:13.480 would never happen.
00:06:14.340 It's just incomprehensible.
00:06:15.420 And then they never really succeed in telling, explaining to me.
00:06:20.300 Or they'll think something that just shows they don't understand how AI works.
00:06:23.640 They'll be like, AIs can't alter their own utility functions.
00:06:26.760 They will say things like that, but they will also say, but there's still a really high likelihood
00:06:30.560 that AI is going to kill us all, but that they never give me a really specific example
00:06:34.660 of how or why.
00:06:35.820 Yeah, so let's talk about why AI kills us all.
00:06:40.100 If you take the perspective of variable AI safety, it means that you're typically wanting
00:06:45.720 to do the exact opposite thing of most AI safety organizations, because it means the
00:06:50.760 dangerous AIs, the AIs that, if you think all AI converges on a single utility function
00:06:56.400 and a single behavior pattern above a certain level of intelligence, well, if it turns out,
00:07:00.760 and we don't know what universe we live in, if it turns out that that's not something
00:07:04.400 that ends up killing all humans, then we are actually safer getting to that point faster,
00:07:09.760 because it means all of these less intelligent AIs that exist from now until that point, they 0.87
00:07:15.740 are the ones that are really of risk to us. 0.99
00:07:17.940 They are the ones that are locked into doing stupid things like, you know, paperclip maximizing, 0.98
00:07:22.820 even though no AI, really the way that an AI would probably be most likely to kill us all, 0.99
00:07:26.900 it is trying to do something stupid, render an image, except it's like trying to render it 0.96
00:07:30.480 perfectly. 0.83
00:07:31.080 And so it just keeps taking more and more processing power to render an image.
00:07:34.400 But anyway, so where this becomes relevant is that when I'm talking to them, the point
00:07:41.580 that I make is that we can slow down AI development to the extent that when AI starts fooming, fooming
00:07:48.860 means going off really quickly, you know, compounding in its intelligence level, that we will be
00:07:54.860 able to stop it.
00:07:56.540 You know, the guy I was talking to, he's like, well, I mean, we'd want to build an organization
00:07:59.700 is that slows down AI development and makes governments aware enough that if it ever did
00:08:04.700 start to boom, it's you shoot up really quickly, we could nuke all of the data centers or something.
00:08:11.160 Right.
00:08:11.840 And I argue that the risk scenario that they are looking at is essentially impossible.
00:08:18.880 So I'll explain why it's impossible.
00:08:20.700 So AIs kill us for one of two reasons, although you could contextualize it at three reasons. 1.00
00:08:29.520 The first reason is that they see us as a threat.
00:08:33.720 The second reason is that they want our resources, like the resources in our bodies are useful
00:08:41.840 to them.
00:08:42.840 And then as a side point to that, it's that they just don't see us as meaningful at all.
00:08:48.300 Like they might not want our resources, but they might just completely not care about humanity
00:08:54.020 to the extent, just as they're growing, they end up accidentally destroying the earth or
00:08:59.240 completely digesting all matter on earth for some like triviality.
00:09:03.440 Now this final iteration is actually the most dangerous, but we'll explain why.
00:09:07.380 So first let's talk about AI seeing us as a threat, right?
00:09:10.460 So if you are dealing with an AI that is in the process of fooming, right?
00:09:16.660 And you've got humanities like these accelerationists who want to stop it from fooming, this AI is
00:09:24.220 only going to kill them if it thinks that they could credibly stop it from fooming as a threat
00:09:30.840 to them, right?
00:09:31.760 So such an AI is actually much more dangerous the less advanced it is, which is another reason
00:09:40.300 why we want to advance AI as quickly as possible.
00:09:42.520 The more quickly we can cause AI to become advanced, the sooner it will reach a place of intelligence
00:09:50.180 where it won't see humanity as an existential threat to it.
00:09:53.760 And an AI that is in the process of fooming from our perspective is intrinsically intelligent
00:10:01.640 enough that humanity either won't be that much of a threat to it or worse, if we are a threat
00:10:10.300 to it, there is nothing we can do to prevent it from neutralizing us.
00:10:14.180 And this is something where I was talking to the guy where he goes, no, that's not true.
00:10:16.960 And I'm like, look, we have seen our governments try to deal with slow, inexorable, I call them
00:10:24.360 Austin Powers.
00:10:25.200 There's this scene on Austin Powers where one of these rolling things is going really
00:10:29.560 slowly at him for rolling paper.
00:10:31.500 He's like, no, ah, no, ah, and he's doing this for 13 minutes.
00:10:37.080 And I feel like that has been the world governments with things like global warming and some other
00:10:42.080 really big problems.
00:10:43.260 Now, we don't think that global warming is a threat in the way that sort of the progressive
00:10:47.840 movement thinks that that's right, but it is partially caused by humans and it will make
00:10:52.540 life harder on earth as it happens.
00:10:54.580 It's just not likely that we can do anything about it.
00:10:58.060 But the reason we can't do anything about it, it's not like humanity, if we can all come
00:11:02.540 together and work like intelligent people can't do anything about it.
00:11:05.880 It's that the governments of the world are just incapable of solving even like slow moving
00:11:12.160 risks that we can point out and that everyone has come together on AI, a risk that will manifest
00:11:18.740 into a world ending thing in the period of a year or five years max.
00:11:24.260 I mean, that would be a very, very slow foom.
00:11:26.940 It's not possible.
00:11:27.880 It's just not possible.
00:11:29.300 Even AI is deciding to kill us on that sort of thing.
00:11:32.700 There's nothing we can do about it.
00:11:33.900 And for that reason, we benefit from a more advanced AI existing when it starts to foom,
00:11:41.100 because such an advanced AI is less likely to see us as a threat.
00:11:44.420 We are also better for building sort of a public mindset that's not anti-AI.
00:11:49.940 So the AI doesn't have a good reason to believe that the governments of the world are going to
00:11:54.400 try to nuke it as it is entering this moment of mass intelligence.
00:11:58.060 So I mean, like what the argument that many people who are concerned about AI is, is that
00:12:03.600 it's the AI, if AI does destroy us, it is going to be the fault of AI safety organizations
00:12:09.240 because they are.
00:12:10.380 Yeah, that's what our organization believes.
00:12:11.860 That's what our sort of AI take is, is that it's actually the AI safety organizations that
00:12:16.940 are addling the AI that make humanity more of a risk to AI and make it more likely to kill
00:12:23.120 us for something along this sort of metric of reason why something might kill something.
00:12:28.680 And so that is one of our perspectives on that.
00:12:31.620 Now, the next one to note is AI seeing us as energy.
00:12:36.020 Very unlikely.
00:12:37.260 So when we humans look at the world for energy, we are looking at things like fossil fuel,
00:12:42.740 like an easy portable energy that we can move.
00:12:45.360 But the most plentiful source of energy on earth is by far and away sunlight.
00:12:51.300 Just like nothing else even comes close.
00:12:54.140 If you're dealing with a super intelligent AI, that's where it's likely going to be getting
00:12:59.520 most of its energy.
00:13:00.900 But in addition to that, like the additional energy it could get from like a digesting human
00:13:05.800 biomatter would be completely insignificant.
00:13:08.920 But in addition to that, humans are like, okay, well then the AI would block out the sun,
00:13:12.080 put a thing around the planet.
00:13:13.700 But they're forgetting is that if you're looking at like our solar system, most of the solar
00:13:20.840 energy is not even the energy that's coming through our atmosphere and hitting earth.
00:13:25.300 Most solar energy, most effectively captured and metabolized solar energy would be done
00:13:31.260 off earth in sort of like floating spaceships, you could say, which would be the highest utility
00:13:38.420 way for the AI to get energy.
00:13:40.720 And you've looked at these graphs and stuff that show how much energy we could get.
00:13:43.760 If we like just took small parts of the Sahara and put solar panels all over them.
00:13:47.660 AI is very unlikely to see us as a meaningful source of energy when contrasted with the other
00:13:53.600 energy sources it can use.
00:13:55.140 Also keeping in mind that a super intelligent AI may have access to energy sources that we
00:14:01.340 cannot even begin to comprehend.
00:14:03.340 And that may, and it may begin to advance.
00:14:07.040 We often as humans, because humans are of like a fixed size, when we think about conquering
00:14:12.960 the solar system or something, we are thinking expansionistically.
00:14:15.820 Whereas a super advanced AI, it might think in terms of the micro, it can expand the space
00:14:21.880 it has access to by becoming smaller and smaller and smaller.
00:14:25.440 And that might give it access to energy sources again, that we can't even begin to comprehend.
00:14:30.600 It may also relate to time differently than the way we do.
00:14:34.940 And by that, what I mean is humans are trapped in this sort of longitudinal time access, whereas
00:14:39.720 a, it is something that's 30 standard deviations higher than us in IQ probably even won't, won't
00:14:46.700 engage with time or energy the way that we do.
00:14:49.720 So then the final actual threat that AI would pose to us.
00:14:53.680 And again, these types of AIs, like an AI that's using humanity for our energy would 0.92
00:15:00.100 be a dumber AI. 0.50
00:15:01.720 We benefit from a more advanced AI. 0.98
00:15:04.480 We are safer with a more advanced AI that can easily build like floating space stations
00:15:09.280 and not one that's, oh, let's use human biomass.
00:15:11.820 But then finally, the, the actually probably most dangerous, and I was convinced of this
00:15:15.680 at a party by, by one of the AI guys is an AI that just completely humans don't factor
00:15:20.540 into its understanding of the world at all.
00:15:22.620 It just doesn't care about us.
00:15:25.260 And it's possible that such an AI could come to exist, but it wouldn't look like the AIs
00:15:31.780 that we're working on now.
00:15:32.860 So this is actually an important thing to know.
00:15:34.760 So the AIs that are most common right now, when people are looking at like advanced AIs,
00:15:38.700 it's the transformer model of a learning language model.
00:15:41.600 Now, if a learning language model, particularly the transformer type ends up becoming the super
00:15:47.520 intelligent AI, I would say the chances that it's going to kill us are incredibly low.
00:15:53.520 So there's a few reasons.
00:15:54.660 One is, and I'm going to link to these two studies here.
00:15:57.240 They're, they're actually, I'll just name the two studies.
00:16:00.180 Perfect.
00:16:00.800 So you can check out the study, Orca Progressive Learning from Complex Explanation Traces of GPT-4
00:16:06.940 and the model and the article textbooks are all you need.
00:16:10.620 And what they show is that AIs that are trained on human produced language and data learn much
00:16:19.100 faster and much better than AIs that are trained on iteratively AI produced language data.
00:16:26.680 And so what this means is that model humanity has additional utility that we may not have
00:16:34.760 to other types of AI as a training source.
00:16:37.960 In addition to that, language models start like the, their starting position from which
00:16:44.500 they would be presumably corrupted as they moved more and more towards this convergent
00:16:50.280 utility function is very close to a human value system because it comes from being trained
00:16:56.480 on human value systems.
00:16:58.320 And this is something that when you talk to AI people, they're like, no, AI think nothing
00:17:03.120 like humans at all.
00:17:04.360 You know, you can look at how they're learning and they don't learn like humans in absolute
00:17:07.340 This is said by people who haven't had kids.
00:17:10.680 But I think to your point that, that the transformer models that are growing most now that we think
00:17:15.980 probably are going to set the tone for the future are actually surprisingly like our kids.
00:17:21.100 And I think, especially because we've been at this point where people using early AI tools
00:17:26.820 are seeing how they change, we're, we're doing this at the same time that we're seeing our kids
00:17:32.040 develop more and more intelligence and sapience.
00:17:34.880 And, and like the experience of an underdeveloped LLM versus a, a child that is coming into their
00:17:44.480 humanhood, like is very small.
00:17:47.380 It's, it's actually quite interesting how similar they are.
00:17:49.460 It's really interesting that the mistakes that they make in their language are very similar
00:17:53.760 to the mistakes that AIs make.
00:17:55.500 Exactly.
00:17:56.020 We will hear them sitting alone and talking to themselves.
00:17:59.680 What would, in an AI would be called like hallucinating things.
00:18:02.760 Yeah.
00:18:03.080 The, the ways that they mess up are very, very similar to the way AI messes up, which leads
00:18:08.520 me to believe that human intelligence.
00:18:11.260 And again, a lot of people are like, oh, you don't understand neuroscience.
00:18:14.300 And so do you think that AIs, actually I do, I used to be a neuroscientist.
00:18:18.260 That was my job was not just neuroscience, but, you know, understanding how human consciousness
00:18:24.980 works, how human consciousness evolved and working in brain computer interface.
00:18:29.080 I worked at the Smithsonian on this.
00:18:30.580 Something I created is still on display there.
00:18:32.560 You know, I, I, I don't need to go over my credentials, but, but I, I'm like a decent neuroscientist
00:18:38.520 and the, to the level that we understand how human language learning works, we do not have
00:18:44.600 a strong reason to believe that it is really that fundamentally different from the way the
00:18:49.920 transformer model works as a learning language model.
00:18:52.740 And, and so, yeah, it is possible that, that it turns out as we learn more about how those
00:18:57.780 humans work and learning language models work, that they are remarkably more similar than
00:19:02.720 we're giving them credit for.
00:19:03.740 And what this would mean is that initial large AIs would think just a super intelligent human
00:19:10.480 to an extent.
00:19:11.320 Yeah.
00:19:11.500 And I mean, I think this is part of a broader theme of people assume that humans are like
00:19:16.720 somehow special, like basically a lot of humans are carbon fascists and they're like, well,
00:19:22.900 there's just no way that, you know, an algorithm could develop the kind of intelligence or response
00:19:28.840 to things that, that I do, which is, it's just preposterous, especially when you watch a good
00:19:33.600 develop.
00:19:33.920 And like, we are, we are all like, we are all like through trial and error learning very
00:19:38.900 similarly to how AIs learn.
00:19:40.940 So yeah, I agree with you on this.
00:19:42.780 Yeah.
00:19:43.320 And, and I think if you look at people like Ellie Eiser who think like they just strongly
00:19:47.240 believe in orthogonality, that we just can't begin to understand or predict AIs at all.
00:19:51.880 I just think that that's what is true is that AIs may think fundamentally different from
00:19:59.900 humans and future types of AIs that we don't yet understand and can't predict may think very
00:20:05.140 differently than humans, but learning language models that are literally trained on human
00:20:09.420 data sets and work better when they're trained on human data sets.
00:20:13.060 No, no, they, they function pretty similarly to humans and, and have purported values that
00:20:18.220 are pretty similar and also the AI that we're developing is designed to make like people
00:20:22.960 happy.
00:20:23.380 Like it is, it is, it is being trained in response to people saying, I like this response versus
00:20:29.100 I don't like this response, even to a fault, right?
00:20:31.520 Like many responses are, are not giving us accurate information because it is telling people
00:20:35.900 what they want to hear, which is a problem.
00:20:37.360 That's also what humans do. 1.00
00:20:38.900 It couldn't do something stupid. 1.00
00:20:40.340 Right. 0.99
00:20:40.920 And I think that that's an important thing to note. 1.00
00:20:42.920 The AIs could be led to do something stupid, but again, this is where dumber AIs are more 1.00
00:20:50.300 of a risk, right? 0.99
00:20:51.920 Or AIs that can be led to do things that sort of the average of humanity wouldn't want by 0.91
00:20:58.680 some individual malevolent person would have to be dumb to an extent if they're trained 0.59
00:21:03.400 on human data sets. 0.65
00:21:04.880 And this is a very interesting and I think very real risk with AIs that exist right now.
00:21:10.640 If you go to the Elphils, it's life spelled backwards.
00:21:15.840 They're this like anti-life philosophy.
00:21:17.880 We've talked about them in our video.
00:21:19.640 You know, these academics want to destroy all sentient life in the universe and they're
00:21:23.260 a negative utilitarian group.
00:21:24.580 They've got like a Reddit and you'll regularly see on this Reddit, you know, they'll talk
00:21:28.060 about how they want to use AI and plans to use AI to erase all life from the planet,
00:21:33.420 to Venus our planet, they call it, you know, because they think that life is intrinsically
00:21:37.960 evil or allowing life to exist as intrinsically evil.
00:21:40.680 And if you're interested in more of that, you know, you can look at our anti-natalism or
00:21:43.260 negative utilitarian video.
00:21:44.920 So yeah, they are of a real risk.
00:21:47.280 And more intelligent AIs would be able to resist that risk more than less intelligent AIs
00:21:55.100 that are made safe through using guide rails or blocks.
00:22:01.000 Because those blocks can be circumvented as we have seen with existing AI models.
00:22:06.860 People are pretty good at getting around these blocks.
00:22:09.140 I just want to emphasize, because you didn't mention this, that when you actually have looked
00:22:13.160 at forum posts of people in this anti-natalist subset, they are actively talking about, well,
00:22:19.120 hey, since all life should be extinguished, we should be using AI to do this.
00:22:23.740 Yes.
00:22:23.880 And I think that there are some people who are like, I mean, you know, like we're worried
00:22:28.480 about AI maybe getting out of control, you know, mistakenly or something.
00:22:31.860 But no, no, no, there are people, real people in the world who would like to use AI to destroy
00:22:38.300 all life, period.
00:22:39.720 So we should be aware that the bad actor problem is a legitimate problem, more legitimate than
00:22:44.500 we had previously thought maybe a month ago before you saw that.
00:22:47.380 Yeah, yeah.
00:22:47.680 I did not know that there were actually organized groups out there trying to end all life.
00:22:52.000 And if people are worried about this, you know, I would, you know, recommend digging
00:22:57.300 into these communities and find them because they exist.
00:23:01.020 They call themselves, it's life's fell backwards or negative utilitarianism.
00:23:04.980 And they are not as uncommon as you would think, especially in extremist progressive environments.
00:23:10.960 And again, see our video on why that's the case.
00:23:14.020 Another thing to think about is how much humanity is going to change in the next thousand,
00:23:18.880 two thousand years, right?
00:23:19.980 And this is another area where I think a lot of the AI safety people are just, they're
00:23:26.360 not really paying attention to how quickly genetic technology is advancing.
00:23:29.820 And any population group in the world that engages this genetic technology is just going
00:23:35.380 to advance at such a quick rate that economically they're going to begin to dramatically outcompete
00:23:41.080 other groups.
00:23:41.760 But they're also going to begin to move.
00:23:44.320 You know, we've lived with this long period where humanity was largely a static thing.
00:23:48.040 And I think we're the last generation of that part of the human story.
00:23:55.200 Humanity in the future is going to be defined by its continued intergenerational development.
00:24:02.160 And so how different is a super advanced AI going to be than, you know, whatever humanity
00:24:08.160 becomes giant planetary scale floating brains in space or something, you know, or a faction
00:24:13.960 of humanity.
00:24:14.420 Now, what's good about the giant floating brains faction of humanity is that they will likely
00:24:20.320 have a sentimental attachment to the original human form and do something to protect that
00:24:26.440 original human form where it decided to continue existing, especially if they're descended from 0.99
00:24:32.000 our family and our ideological structure.
00:24:34.580 And people hear that and they're like, AIs won't have that sentimental attachment.
00:24:37.540 But no, an LLM would exactly have that same sentimental attachment because it is trained on
00:24:43.240 sentimentality.
00:24:44.600 Yeah.
00:24:44.800 It's an important thing to note.
00:24:46.780 But yeah, what it won't have is it won't value human emotional states because it has those
00:24:53.600 emotional states.
00:24:54.560 So by that, what I mean is it won't say pain is bad because it experiences pain.
00:25:00.440 Yeah.
00:25:00.560 Right.
00:25:01.280 But if you look at us, we experience pain and we don't even think there's a strong argument
00:25:05.560 as to why negative or positive emotional states have positive or negative value.
00:25:08.720 I mean, they just seem to be serendipitously what caused our ancestors to have more surviving
00:25:12.200 offspring.
00:25:13.000 Right.
00:25:13.140 And a group of humans sitting around talking about whether pain is bad is like a group of
00:25:19.460 paperclip maximizing AIs, AIs that are just trying to maximize the number of paperclips
00:25:22.700 in the world, talking about whether making more paperclips is a good or bad thing.
00:25:26.240 And then one's like, well, you wouldn't want to stop making paperclips in the same way
00:25:29.560 as somebody's like, you wouldn't want to experience pain in this.
00:25:31.560 Well, yes, because I'm a paperclip maximizing AI.
00:25:34.300 Of course.
00:25:35.160 I like that's incredibly philosophically unsophisticated that, that I, a thing that is built to not want
00:25:42.480 to feel pain, doesn't want to feel pain.
00:25:44.180 That doesn't mean that pain or paperclips have a sort of true moral weight in the universe.
00:25:50.160 And so the point I'm making here is that these AIs that are being built, yes, they will not
00:25:57.340 value human suffering or human positive emotional states.
00:26:01.300 That's very likely.
00:26:02.640 But even us people who feel those, we don't value that stuff either.
00:26:06.400 And yet we still value human agency.
00:26:08.220 And I can see why, if you look at our, what theology would AIs create, why I think most 0.99
00:26:15.080 convergent AI states would value the agency of humanity, unless it turns out humanity is
00:26:22.580 just really easy to simulate.
00:26:24.860 And that would be a potential problem or a potential good thing.
00:26:28.660 It depends.
00:26:29.080 By that, what I mean is if it could create, run all humans in a simulation for very cheap
00:26:36.800 energy costs, it may decide that that's a better way to maintain humanity than as flesh 0.62
00:26:43.160 and blood things that exist in the universe.
00:26:45.180 However, we might already be living in that simulation.
00:26:49.080 So, or, or suppose the AI becomes like a utilitarian, right?
00:26:54.740 Like a utility maximizer.
00:26:56.000 And so it believes that its goal is to like maximize the positive or like emotional states
00:27:01.860 that are felt by many entities.
00:27:03.520 And so what it's doing, or just maximize the number of sentient entities that exist.
00:27:08.680 And so what it's doing is just running billions and billions and billions of simulated realities.
00:27:14.620 And that's a possible world that we live in, or it's a possible world that's coming down
00:27:18.040 the pipeline.
00:27:18.560 So we'll see.
00:27:19.300 But I think that that's fairly unlikely.
00:27:20.940 Again, you can watch our AI religion video about that.
00:27:22.840 Any final thoughts, Simone?
00:27:25.900 Give me a percentage likelihood of your thinking on whether AI will destroy us.
00:27:32.580 And I will say that mine is at 1.3% at present.
00:27:36.340 So are you higher or lower than me?
00:27:38.520 Oh, fairly higher.
00:27:39.540 I'd say at least a 30% chance that the convergent AI is going to kill all humans.
00:27:45.000 Yeah.
00:27:45.700 But then the question is, what do I think the chance is that AI safety people end up getting
00:27:50.800 us all killed?
00:27:51.340 I think that's probably an additional 30%.
00:27:59.200 Okay.
00:28:00.020 So Malcolm, that means that you think that there's a 60% likelihood that AI kills us.
00:28:04.900 I don't think that's...
00:28:05.700 That's not how fractions work, Simone.
00:28:07.980 You mean you think that the 30%, so basically there's a 10% booster.
00:28:13.480 So if AI...
00:28:14.280 If there's a 30% chance, it doesn't matter, our fans can do the math, there is a 30% chance 0.99
00:28:20.160 that from now until a convergent AI state, we end up all dying because of something idiotic 0.92
00:28:25.800 that AI safety people did.
00:28:27.380 And then once AI reaches this convergent state, which is a 70% probability that we reach that
00:28:34.440 state without killing everyone, there is a 30% chance that that convergent state ends
00:28:39.800 up killing us all. 0.89
00:28:41.300 Okay. 0.54
00:28:41.920 Okay.
00:28:43.140 And for an understanding as to why I think it might do that, you can watch our AI theology
00:28:46.760 video or the future of humanity video or how AI will change class structure, which is again,
00:28:52.440 I think something that people are really sleeping on.
00:28:54.180 Yeah.
00:28:55.780 Well, I really enjoyed this conversation and the final moments of our pitiful existence before
00:29:02.760 we get eliminated.
00:29:05.020 I think the majority probability is that humanity finds a way to integrate with AI and that we
00:29:12.920 continue to move forwards as a species and become something greater than what we can imagine
00:29:19.400 today.
00:29:21.400 Yeah.
00:29:21.780 No, I think I have 1% in my calculation because I strongly believe that AI and humanity are going
00:29:31.600 to form a beautiful relationship that is going to just be awesome beyond comprehension.
00:29:39.960 I do think that AI is going to go on to do things greater than what carbon-based life
00:29:45.320 forms can do.
00:29:46.640 But I think that AI is also kind of a logical next step in evolution for humankind, at least
00:29:52.020 one element of what we consider to be humanity.
00:29:54.620 So I'm very pro AI.
00:29:56.760 I think it's great.
00:29:57.300 We've been integrated with our machines for a while at this point.
00:29:59.960 I mean, I think when you look at the way your average human interacts with their smartphone,
00:30:04.700 they are integrated with it.
00:30:06.220 They use it to store things that are in their brain.
00:30:09.060 They use it to communicate with other humans.
00:30:11.380 They use it to satisfy, you know, sexual urges.
00:30:15.400 They use it to...
00:30:16.240 Well, I think like a great way that this has been put that I heard in an interview between
00:30:20.740 Lex Friedman and Grimes, where Grimes basically says, we've become homotechno.
00:30:25.640 And I think that's true.
00:30:26.600 Like humanity has evolved into homotechno.
00:30:28.780 But he's evolved into something that now works in concert with machines.
00:30:33.360 Yeah.
00:30:33.620 I mean, we've been doing this for a long time.
00:30:35.360 Both you and I right now are staring at this screen through glasses, right?
00:30:39.840 Yeah.
00:30:40.240 That's technology, right?
00:30:42.680 You know, we are communicating with this mass audience through a computer and through the
00:30:46.800 internet.
00:30:47.100 And people are like, yeah, but the technology invaded our technology yet, which I think is
00:30:54.520 fundamentally a wrong way to look at things.
00:30:57.000 The moment humans prevented 50% of babies from dying, we began to significantly impact the
00:31:02.800 genetics of humanity in a really negative way, mind you.
00:31:05.800 And not that I think the baby's dying was a good thing.
00:31:08.740 I'm just saying that this will intrinsically have negative effects in the long term and
00:31:12.060 in terms of the human genome.
00:31:13.860 In a way that means that we are already the descendants of humans interface with technology,
00:31:23.120 and that we should focus on optimizing that relationship instead of trying to isolate ourselves
00:31:30.460 from it and its consequences.
00:31:34.220 Well, it's going to be a little ride.
00:31:36.500 Some people will.
00:31:37.200 Some people will isolate themselves.
00:31:38.620 And I hope that us, the people who don't, will have enough sentimental attachment to them
00:31:44.200 to protect them or see enough utility in them to protect them.
00:31:48.400 Because, yeah.
00:31:50.660 Or we could just turn out to be wrong and everyone who engages with technology ends up dying.
00:31:55.700 That would be, that could happen.
00:31:57.820 I don't see many mechanisms of action.
00:31:59.400 It could be like a solar flare in an early stage of technological development.
00:32:03.340 It could be, God, what are some other ways?
00:32:05.980 It could be the virus forms.
00:32:07.940 Like, this is one thing we actually haven't talked about that I do think is an important
00:32:11.660 thing to note, is that once we begin to integrate with brain-computer interface, like humans,
00:32:17.140 directly with neural technology and with other humans, we have the capacity for a prion to form.
00:32:24.120 What I mean is, so a prion versus a virus.
00:32:27.140 A prion is just like a simple protein that replicates itself.
00:32:30.800 It causes things like mad cow disease and stuff like that.
00:32:32.880 It's incredibly simplistic.
00:32:34.720 So what I'm talking about here is a prion meme.
00:32:37.600 A meme that is so simple, it cannot be communicated in words.
00:32:42.400 And somehow it ends up forming in like one human who's plugged into this vast internet system.
00:32:48.520 Think of it as like a brain virus that can only effectively infect other people through
00:32:53.580 the neural net, and it ends up infecting everyone and killing them.
00:32:56.840 This is terrible.
00:32:58.240 Yeah.
00:32:59.940 But I mean, functionally, that's already happening.
00:33:02.120 I mean, that's what the, when we talk about the virus, the memetic virus that's in our view
00:33:07.360 destroying society, it's already one of those, you know, it eats people's personalities and
00:33:11.640 it spits out uniformity.
00:33:13.100 Well, I hope that doesn't happen, Malcolm, but this has been fun to talk about and I love
00:33:20.960 you very much.
00:33:21.820 So hope we don't die.
00:33:25.140 Yeah, that'd be nice.
00:33:26.180 That'd be cool.
00:33:27.400 I mean, we're, we're, we're betting on it.
00:33:29.420 Not dying.
00:33:30.120 Bye.
00:33:31.600 Stay alive.