Making Sense - Sam Harris - August 28, 2023


#332 — Can We Contain Artificial Intelligence?


Episode Stats

Length

55 minutes

Words per Minute

165.60564

Word Count

9,148

Sentence Count

6

Hate Speech Sentences

1


Summary

Mustafa Suleiman is the co-founder and ceo of Inflection AI and a venture partner at Greylock, a venture capital firm. Before that, he cofounded Deepmind, one of the world s leading artificial intelligence companies now part of Google, and he was Vice President of AI Product Management and AI Policy at Google. He is also the author of a new book, The Coming Wave, Technology, and the 21st Century's Greatest Dilemma: The Problem of Our Time. In this conversation, we talk about the new book and the progress that was made in AI by his company, various landmarks they achieved, the looming possibility of a misinformation apocalypse, the dangers of asymmetric threats, conflict and cooperation with supply chain monopolies, and other topics. In this episode, we discuss the risks of our making progress in AI as a distraction from more pressing problems: the inevitable spread of general purpose technology, the nature of intelligence, productivity, growth and labor disruption, the containment problem, the importance of scale, open source llms, changing norms of work and leisure, the redistribution of value, the looming possibility of The problem of our time? And so much more! He shares many of my concerns, but with different points of emphasis, and as you ll hear, he shares some of them in different ways. You ll get to the point of the conversation by listening to the making sense podcast. Make sure to subscribe to the Making Sense Podcast wherever you get your podcasts. This is making sense. - Sam Harris Thank you for the podcast by Sam Hanaman and the Making sense Podcast by The Making Sense Team? You'll get access to the latest episodes of the podcast that makes sense of what we're doing here? Sam harris - The making sense Podcast? -- The Making sense podcast? -- The podcast is made possible entirely by the podcast is all about making sense, and it's made possible by the team at Making sense by Sam Harris -- and the podcast makes you better than you get to know you, you'll get a chance to be there, and you can post it on social media or you can hear it on the internet, too you'll become a friend, or you're getting a friend like that, and that's not even that, you're gonna get that, that's a good thing, or that's gonna be a good friend, so you'll hear about it?


Transcript

00:00:00.000 welcome to the making sense podcast this is sam harris just a note to say that if you're hearing
00:00:12.500 this you are not currently on our subscriber feed and will only be hearing the first part
00:00:16.900 of this conversation in order to access full episodes of the making sense podcast you'll
00:00:21.800 need to subscribe at sam harris.org there you'll find our private rss feed to add to your favorite
00:00:27.020 podcatcher along with other subscriber only content we don't run ads on the podcast and
00:00:32.500 therefore it's made possible entirely through the support of our subscribers so if you enjoy
00:00:36.540 what we're doing here please consider becoming one welcome to the making sense podcast this is sam
00:00:49.320 harris okay just a reminder that subscribers to the podcast can now share full episodes by going
00:00:58.600 to the episode page on my website and getting the link and you can share one-to-one with friends and
00:01:06.740 family or you can post to social media whatever you like okay today i'm speaking with mustafa
00:01:16.020 suleiman mustafa is the co-founder and ceo of inflection ai and a venture partner at graylock
00:01:23.520 a venture capital firm before that he co-founded deep mind which is one of the world's leading
00:01:29.980 artificial intelligence companies now part of google and he was vice president of ai product management
00:01:36.940 and ai policy at google and he is also the author of a new book the coming wave technology power and
00:01:45.880 the 21st century's greatest dilemma which is the focus of today's conversation we talk about the
00:01:51.820 new book we talk about the progress that was made in ai by his company deep mind various landmarks they
00:02:00.380 achieved atari dqn alpha go alpha zero alpha fold we discuss the amazing fact that we now have
00:02:08.820 technology that can invent new knowledge the risks of our making progress in ai super intelligence as a distraction
00:02:16.840 from more pressing problems the inevitable spread of general purpose technology the nature of intelligence
00:02:23.220 productivity growth and labor disruption the containment problem the importance of scale
00:02:29.920 open source llms changing norms of work and leisure the redistribution of value introducing friction into the deployment
00:02:39.680 of ai regulatory capture the looming possibility of a misinformation apocalypse digital watermarks asymmetric
00:02:49.060 threats conflict and cooperation with china supply chain monopolies and other topics anyway it was great to get
00:02:58.020 mustafa here he's one of the pioneers in this field and as you'll hear he shares many of my concerns
00:03:03.560 but with different points of emphasis and now i bring you mustafa suleiman
00:03:09.360 i am here with mustafa suleiman mustafa thanks for joining me great to be with you sam thanks for having me
00:03:20.640 so you you have a new book which um the world needs because this is the problem of our time the title is
00:03:28.420 the coming wave technology power and the 21st century's greatest dilemma
00:03:34.000 uh and we will we'll get into the book because um it's really um quite a good read and uh we will
00:03:42.200 talk about what that coming wave is but you're especially concerned about ai which is your um your
00:03:49.260 wheelhouse but also you're talking about synthetic biology uh and to i guess a lesser degree robotics
00:03:56.540 and some other technologies that are going to be more and more present if things don't run totally off
00:04:02.560 the rails for us but before we we jump into the book let's talk about your background but how would
00:04:08.060 you describe the bona fides that have have brought you to this conversation yeah i mean i i started life
00:04:15.740 i guess as a entrepreneur when i was 18 i started my first company which was a point of sale system uh sales
00:04:24.140 company and we we were sort of installing these these sort of very early pdas back in 2002 2003 and networking
00:04:33.340 equipment i wasn't successful uh but that was my first attempt i i dropped out of oxford uh the end of my
00:04:40.380 second year where i was reading philosophy to start a charity and i helped two or three other people
00:04:47.100 get a telephone counseling service off the ground it was a secular service for young british muslims i was
00:04:54.220 a just turned an atheist at the time having been to oxford discovered human rights principles and the
00:05:01.740 ideas of universal justice and uh managed to sort of move out of the faith and uh and decided that i i
00:05:10.220 i really wanted to you know dedicate my life to to doing good and and studying philosophy and the theory
00:05:16.700 uh was too too esoteric and and too distant from action i'm a very kind of practical action focused person
00:05:25.260 so i spent a couple years doing that a little bit after that time i spent a year or so working in
00:05:31.580 local government as a human rights policy officer for the mayor of london at the time i think i was
00:05:37.180 21 when i started that job it was very big and exciting but ultimately quite unsatisfying and
00:05:43.660 frustrating who was the mayor was that johnson that was before johnson yeah quite a bit before uh it was
00:05:50.140 it was ken livingston back in 2004 right so quite a while back in london and then from there i wanted to
00:05:58.860 see how i could scale up my impact in the world and um you know i i helped to start a conflict
00:06:06.780 resolution firm i was very lucky at the age of 22 to be able to co-found this this consultancy with
00:06:14.220 a group of some of the most practiced uh negotiation experts in the world some of the people who are
00:06:21.180 involved in the peace and reconciliation process in south africa post-apartheid and there's a big group
00:06:26.220 of us coming together with very different skills and backgrounds and uh i had incredible three years
00:06:31.580 there working all over the world in in cyprus and for the dutch government for you know on the israel
00:06:38.620 palestine question uh you know many different places and it was hugely inspiring and taught me
00:06:44.860 a lot about the world but i sort of fundamentally realized from there that if i if i didn't get back
00:06:50.220 to technology i would miss the most important transition you know wave if you like happening
00:06:55.420 in my lifetime and uh you know i i said about shortly after the climate negotiations that we were working
00:07:04.140 on in 2009 in copenhagen everyone left feeling frustrated and disappointed that we hadn't managed
00:07:11.020 to reach agreement and this was the year that sort of obama was coming over and everyone had a lot of hope
00:07:16.140 and it didn't happen uh it turns out for another 10 or 12 years and i i sort of had this aha moment i
00:07:22.860 was like if i don't get back to technology then i'm going to miss the most important thing happening and
00:07:27.820 so i set about on this quest trying to sort of find anyone that any you know anyone who i knew even
00:07:34.380 tangentially who was working in technology and uh my best friend from when we were teenagers um his older
00:07:42.620 brother was demis asabis and uh we we were playing poker together one night in the victoria casino in
00:07:49.500 london and and we got chatting about the ways that at the time you know we framed it as robots were
00:07:55.100 going to transform the world and deliver enormous productivity boosts and improve efficiency in every
00:08:01.100 respect and we were sort of debating like how do you do good in the world how do you get things done
00:08:06.780 you know what what is the real set of incentives and efforts that really makes a difference
00:08:12.220 and um you know both very passionate about science and technology and having a positive
00:08:16.620 impact in the world and you know one thing led to another and eventually we um ended up starting deep
00:08:22.540 mind and i did that for 10 years yeah along with shane leg right shane is uh is our other co-founder
00:08:31.580 exactly shane was a gatsby computational neuroscience unit in london at the time and he was just finishing
00:08:37.660 uh he had just finished his phd a few years earlier he was doing postdoctoral research
00:08:42.220 and his phd was on definitions of intelligence which was super interesting it was very obscure and
00:08:48.380 really really relevant he was sort of trying to synthesize 60 or so different definitions of
00:08:54.140 intelligence and trying tried to sort of abstract that into an algorithmic construct one that we could use to
00:09:00.300 measure progress towards some defined goal and um his frame was that intelligence is the ability to
00:09:07.900 perform well across a wide range of environments so the core emphasis was that intelligence was about
00:09:14.940 generality right and we you know we can get into this there's lots of different definitions of
00:09:19.340 intelligence which place emphasis on different aspects of our capabilities but generality has become
00:09:25.900 the core concept that sort of dominated the field for the last sort of 12 15 years and and of course
00:09:32.540 the term agi i mean that that predated shane but i think it was very much popularized by our kind of
00:09:40.220 mission you know sort of it was really the first time in a long time that a company had been founded to
00:09:47.420 invent general intelligence or agi and that was our mission to try and build safe and ethical artificial
00:09:54.140 general intelligence so i'm trying to remember where we met i know we were both at the puerto rico
00:09:59.900 conference at the beginning of 2015 that uh i think i don't know if it was the first of these meetings
00:10:07.900 but it was the first that i was aware of that really focused the conversation on ai safety and risk
00:10:13.820 and i know i met demis there i'm not sure i think you and i met in la subsequent to that is that right
00:10:20.460 yeah i think i think we we met i can't remember if we met before or after that but i think we had we
00:10:26.780 had a common interest in our la conversation it might have been just before that talking about
00:10:32.540 extremism and radicalization and terrorism and oh within islam yeah yeah that's right yeah yeah so yeah
00:10:39.340 i can't i don't think we met in puerto rico but that that conference was very formative of my
00:10:44.140 you know it was really like my first impression of of how big a deal this was going to be ultimately
00:10:50.220 and then there was a subsequent conference in 2017 at asilimar where i think we uh we met again and i
00:10:58.860 think i met shane there as well so let's before we jump into again the book and what you're doing
00:11:05.260 currently because you've you've since moved on from deep mind and you have a new company that we'll talk
00:11:09.740 about but um let's talk about deep mind because it really was you know before it was you know it's
00:11:17.980 been eclipsed in the popular consciousness by open ai of late with with the advent of chat gpt and
00:11:26.540 and uh large language models but prior to that really deep mind was the preeminent
00:11:33.820 uh it may in fact still be the preeminent ai company but it's now a branch of google give me
00:11:41.980 give us a little bit of the history there and and tell us what was accomplished because you at deep
00:11:49.020 mind you had several breakthroughs that were just fundamental and you really put ai back on the map and
00:11:57.740 prior to what you did there we were in an ai so-called ai winter where it was just common
00:12:04.460 knowledge that this artificial intelligence thing wasn't really panning out and then all of a sudden
00:12:09.420 everything changed so i think pre-acquisition um which was in 2014 i think our there were probably two
00:12:17.100 principal contributions that we made i think the first is we made a very early bet on deep learning i
00:12:24.140 mean the company was founded in 2010 in the summer of 2010 and it really wasn't for a couple of years
00:12:30.780 that deep learning had even appeared on the on the field even academically with the image net challenge
00:12:36.140 a few years after we founded so that was a very significant bet that we made early and that we got
00:12:41.660 right and the consequence of that was that we were able to hire some of the best phds and postdoctoral
00:12:48.780 researchers in the world you know who at the time were working on this very obscure very uninteresting
00:12:55.820 you know largely not very valuable subject in fact jeff hinton was one of our consultants so was his
00:13:04.300 student at the time ilia satskiva who's now chief scientist and co-founder of open ai along with many
00:13:10.860 others from open ai and elsewhere who you know basically either worked with us full-time or worked with
00:13:16.700 us as as consultants and that was largely you know reflective of the fact that we got the bet right
00:13:23.180 early on deep learning the second contribution i would say was the combination of deep learning and
00:13:29.500 reinforcement learning i mean if deep learning was was obscure reinforcement learning was even more
00:13:35.020 theoretical and you know we were actually quite careful to frame our mission among academics you know
00:13:42.620 less around sort of agi and more around applied machine learning you know there was a certainly
00:13:48.620 in the very early days we're a bit hush-hush about it but you know as we got more traction in 2011 2012
00:13:54.540 it became very attractive to people who were otherwise quite theoretical in their outlook to come and
00:14:00.620 work on problems like reinforcement learning in you know sort of more engineering focused setting albeit
00:14:06.300 still a research lab and it was the combination of deep learning and reinforcement learning that
00:14:11.820 led to our first i think major contribution which was the atari dqn ai dqn so you know dqn was a pretty
00:14:23.660 incredible system i mean it essentially learned to play 50 or so of the old school sort of 80s ataris
00:14:31.100 games atari games to human level performance uh simply from the pixels learning to correlate a set of
00:14:38.700 rewarding moments in the game via score with a set of frames that led to that score in the run-up to
00:14:46.380 that and the actions that were taken there and that was a really significant achievement it was actually
00:14:50.940 that which caught larry page's attention and led him to email us you know and and you know sort of invite
00:14:59.420 us to come and be part of of google and then google acquired you and uh what was the logic there you
00:15:07.340 just it was good to have google's resources to scale or i mean larry made a very simple claim which was
00:15:16.700 you know i've spent you know the last you know 10 years or so building a platform with all the resources
00:15:25.100 necessary to make a really big bet on agi you know why should you guys go through all of that again
00:15:32.140 you know we'll give you the freedom you need to carry on operating as a essentially a you know
00:15:37.500 independent subsidiary even though we were part of google why wouldn't you just come and work with us
00:15:42.780 and have all the resources you need to to scale you know significantly which is what we did and it's it's
00:15:49.020 it was a very compelling proposition because at the time you know monetizing deep learning back
00:15:54.940 in 2014 was going to be really tough so but google had its own ai division as well that was just kind
00:16:03.420 of working in parallel with deep mind did that did you guys at some point you guys merged i don't know
00:16:09.740 if that happened after you left or before but how did that was there a firewall between the two
00:16:16.060 divisions for a time and then that came down or how'd that work yeah so the division you're referring
00:16:22.380 to is google brain which is run by jeff dean and i think that started in 2015 with andrew ung actually
00:16:29.180 as well and you know in some ways that's the kind of beauty of google's scale right that it was able to
00:16:35.740 run multiple huge billion dollar efforts in parallel and the merger which i think has been long coming
00:16:43.100 actually only happened this year so google plus deep mind is now google deep mind and uh most of
00:16:50.540 the kind of open-ended research on ai is now consolidated around google deep mind and all of
00:16:56.780 the sort of more focused applied research that helps google products more directly in the short term is is
00:17:03.740 focused on a on a separate division google research right so you had those the the atari game
00:17:11.100 breakthrough which caught everyone's attention because you either you have these you know if
00:17:16.220 memory serves you managed to build a system that had i mean it achieved human level competence and
00:17:25.740 beyond and also achieved novel strategies that many humans wouldn't come up with but then the the
00:17:33.500 real breakthroughs that got everyone's attention were with alpha go and alpha zero and alpha fold perhaps
00:17:40.860 you can run through those because that that's when at least to my eye things just became unignorable
00:17:47.580 in the ai field yeah that that's exactly right i mean it it's it's it's pretty interesting because sort
00:17:54.060 of after we got acquired it was actually sergey that was sort of insisting that we tackle go i mean his point
00:18:01.660 was you know that go is a massively complex space and you know all the traditional methods that
00:18:09.580 have previously been used for games before dqn which essentially involved handcrafting rule-based
00:18:15.900 features which is is really what drove the work behind deep blue ibm's uh model you know a long time
00:18:22.940 ago 97 i think it was you know go has something like 10 to the power of 170 possible configurations
00:18:30.860 of the board so it's a 19 by 19 board with black and white stones and the rules are very simple it's a
00:18:37.420 turn-based game where each player simply moves one place places one stone on the board and when you
00:18:44.780 surround your opponent's stones you remove them from the board and the goal is to sort of surround your
00:18:49.740 opponent and so it is a very simple rule set but it's a massively complicated possible set of different
00:18:58.460 configurations that can emerge and so you you can't sort of search all possible branches of of of that
00:19:05.580 space because it's so enormous yeah i mean 10 to the 170 is like more atoms than there are in the known
00:19:10.780 universe approximately yeah i think that's i think something like 10 to the 80 that gets you all the
00:19:16.540 protons in the universe so yeah there's it gets bigger still when you're talking about though yeah
00:19:22.540 right so you know and so this needed a new suite of methods and you know i think it was an incredible
00:19:29.660 experience seeing alpha go progressively get better and better i mean we we already had an inkling for
00:19:35.260 this when we saw it play the atari games but this was just seismically more complicated and vast and yet it
00:19:42.540 was using the same basic principle actually the same principle that has subsequently been applied
00:19:47.900 in in protein folding too so you know i think that's what's really interesting about this is that is the
00:19:54.300 generality of the ideas that simply scale with more compute you know because a a couple of years later
00:20:01.420 alpha go became alpha zero which essentially achieved superhuman performance without any
00:20:08.060 learn without any learning from prior games so you know part of the trick with alpha go is that it
00:20:14.860 looked at hundreds of thousands of prior games it's almost like the expert knowledge of existing players
00:20:20.380 that has been handed down for centuries of playing the game whereas alpha zero was able to learn entirely
00:20:26.780 through self-play you know almost like i think the intuition is spawning instances of itself in order to play
00:20:33.260 against itself in simulated environments many many hundreds of millions of billions of times way more
00:20:39.980 it turns out to be way more valuable than bootstrapping itself from the first principles of human
00:20:45.580 knowledge which if you think about the size of the state space represents you know a minor subset of all
00:20:51.100 possible configurations of that board and that that was a kind of remarkable insight and and actually
00:20:56.220 it did the same thing for for other games including chess and shogi and so on yeah that's a really
00:21:03.500 fascinating development where it's it's now uncoupled from the repository of human knowledge it plays itself
00:21:10.540 and over the course of i think it was just a day of self-play it was better than than alpha go and any other
00:21:19.500 system right right that's exactly right and and obviously that's partly a function of compute but the
00:21:25.900 basic principle gives an important intuition which is that because these methods are so general they
00:21:31.820 can be paralyzed and scaled up and that means that you know we can sort of take advantage of all of the
00:21:39.340 you know traditional assets of you know computing infrastructure rather than relying on you know old
00:21:44.780 school methods you know perfect memory paralyzable compute you know moore's law you know daisy chaining
00:21:52.140 compute together just like we do with with gpus these days so you know in some ways that's a that's the
00:21:58.780 key intuition because it means the sort of barrier to application of the quality of the algorithm is lower
00:22:06.380 because it's turbocharged by all these other underlying drivers which are also improving the power and
00:22:12.940 performance of these models and um also alpha zero in when it was playing the world champion
00:22:20.700 came up with a move that you know all go experts thought they immediately recognized as a mistake but
00:22:27.740 then when the game played out it turned out to be this brilliant novel move that no human
00:22:33.580 would have made and it just a big piece of discovered go knowledge yeah i mean i i remember
00:22:41.340 sitting in the commentary room live watching that unfold and uh listening to the commentator who was
00:22:48.460 himself a nine down uh you know expert say that it was a mistake he was like oh no it's we lost and uh
00:22:55.660 it took 15 minutes for you know him to correct that and and sort of come back and reflect on it it was a
00:23:01.740 really remarkable moment and actually you know for me it was a great inspiration you know because
00:23:08.540 this is why we we started the company i mean the quest was to try to invent new knowledge i mean our
00:23:16.060 our goal here is to try to design algorithms that can teach us something that we don't know um
00:23:22.220 not just reproduce existing knowledge and synthesize information in new ways but genuinely discover
00:23:27.980 new strategies or new molecules or you know new compounds new ideas and contribute to the you know
00:23:35.980 the kind of well of human knowledge and capability and you know this was a kind of first well actually
00:23:41.900 it was the second indication because the first instinct i got for that was watching the atari games
00:23:46.140 player learn new strategies from scratch and this this was kind of the second i think and what about
00:23:52.060 alpha fold because this is a very different application of the same technology what what
00:23:57.820 what did you guys do there and what was the the project well protein folding is a long-standing
00:24:04.860 challenge and we actually started working on this as a hackathon which started in my group back in 2016
00:24:13.100 and it was really just an experiment to see if you know some of the alpha go models could could
00:24:18.540 actually make progress here and the basic idea is that if you can sort of generate you know an an
00:24:25.900 example of the way a protein folds this folding structure represents might tell you something about
00:24:32.940 you know the the the value of that molecule in practice what it can do what you know what its
00:24:38.620 strengths and weaknesses are and so on and so you know the the nice thing about it is because it
00:24:43.820 operated in a simulated environment it was it was quite similar to some of the games that we had been
00:24:48.540 playing you know teaching our models to to play and you know previously the experiments had done
00:24:55.420 something like 190 000 proteins um which is about 0.1 percent of the of all the proteins in existence but
00:25:03.260 but in the in alpha fold 2 the team actually open sourced something like 200 million protein structures
00:25:09.660 all in one go which is sort of all all known proteins this is a massive breakthrough that took you know
00:25:16.140 four or five years of of work in development and and i think just gives a an indication of the kinds
00:25:22.380 of things that become possible with these sorts of methods yeah i forget someone gave a what purported
00:25:29.100 to be a kind of a straightforward comparison between what alpha fold did there and the academic years of
00:25:37.500 phd theses and it was something like you know 200 million phd theses got accomplished in a few years
00:25:44.620 there uh in terms of solving those protein folding problems yeah i mean that that that's those kinds
00:25:50.700 of insights those kinds of sort of compressions are similar to you know across the board with many
00:25:56.380 technologies like i another one that's sort of similar to that is that the amount of of sort of
00:26:01.740 labor that once produced 50 minutes of of light in the 18th century you know now produces 50 years worth of
00:26:10.060 light and and that just gives a a a sense for how technology has this massive compressive effect
00:26:18.300 that is hugely leveraging in terms of what we can do yeah there's there's another crazy analogy in your
00:26:23.740 book talking about the size of these uh the parameters of these new large language models which
00:26:31.260 we'll get to but the comparison was something like executing all of these floating point operations
00:26:36.540 it's it's if you know if every operation were a drop of water you know the largest large language
00:26:43.340 models execute as many calculations as would fit in to the entire pacific ocean so it's just the scale is
00:26:51.020 is astounding right so your book it was a bit of a surprise for me because you are more worried than i
00:27:00.940 i realized about how all of this can go wrong and i i got the sense in you and i haven't spoken very much
00:27:07.740 but in talking to you and demis and shane i got the sense that and these conversations are you know for now
00:27:15.500 several years old that you were more sanguine about our solving all of the relevant problems you know alignment
00:27:24.540 being the chief among them but other concerns of bad incentives and arms race conditions and
00:27:31.500 etc that you were you all were putting a fairly brave face on on a problem that was making many of us
00:27:40.540 increasingly shrill and you know not to say hysterical um and so there were you know i guess the most
00:27:48.060 hysterical voice of the moment is someone like eliezer yudkowski and there was obviously nick bostrom and
00:27:54.060 others who who were you know issuing fairly grave warnings about how it was more likely than not
00:28:01.260 that we were going to screw this up and build something that we really can't control ultimately
00:28:06.460 and that would could well destroy us and on the way to the worst possible outcome there are many bad
00:28:13.740 very likely outcomes like you know a misinformation apocalypse and other risks but in your book you're
00:28:22.460 you don't give the risks short shrift i mean you're really you do seem to suggest that and certainly
00:28:30.540 when you add in the the attendant risks of synthetic biology here which we'll talk about you are quite
00:28:37.660 worried and yet there's no as you agree with a point i made uh early on here which is that you know as
00:28:46.060 worried as as we are there really is no brake to pull i mean the incentives are such that we're
00:28:51.820 going to build this and so we have to sort of figure out how to repair the the rocket as it's
00:28:57.980 taking off and align it properly as it's taking off because there's just no there's no getting off this
00:29:03.020 ride at the moment despite the fact that people are calling for a moratorium or some people are so i'm
00:29:09.100 just i guess what before we jump into the book when did you get worried were you were you were
00:29:15.020 you always worried or um are you among the newly worried people like jeff hinton who i mean like
00:29:21.020 so just like jeffrey hinton who you who you mentioned is really the godfather of this technology and he just
00:29:28.060 recently resigned from google so that he could express his worries in public and he seems to have just
00:29:34.460 become worried in the presence of these you know large language models and it's quite inscrutable to
00:29:41.500 me that he suddenly had this change of heart because you know in my view the basis for this concern was
00:29:49.740 always self-evident so give me the the memoir of your concerns here yeah so this is not a new consideration
00:29:58.060 for me i i've been worried about this from you know the the very first days when we founded the
00:30:03.340 company and you know in fact our our strap line on our business plan that we took to silicon valley
00:30:10.140 in 2010 was building artificial general intelligence safely and ethically for the benefit of everyone
00:30:16.780 and that was something that was critical to me all the way through when we when we sold the company we
00:30:22.380 made it a condition of the acquisition that we have an ethics and safety board with some independent
00:30:27.340 members overseeing technology in the public interest that we our technologies wouldn't be used for
00:30:32.300 military purposes like lethal autonomous weapons or surveillance by the state you know and and since
00:30:38.940 then at google i you know went through lots and lots of different efforts to experiment with different
00:30:44.540 kinds of oversight boards and charters and external scrutiny and independent audits and all kinds of
00:30:50.380 things and so i'd say i definitely been top of mind for me all the way through i think where i diverge from
00:30:57.340 the sort of bostrom camp a bit is that i think that the language around super intelligence has actually
00:31:04.700 been a bit of a distraction and i think it was quite obviously a distraction from from fairly early on i think
00:31:12.140 that the focus on this you know sort of intelligence explosion this ai that recursively self improves and
00:31:20.220 suddenly takes over everybody and turns the world to paper clips i think has consumed way more time than
00:31:27.100 the idea justifies and actually i think there's a bunch of more near-term very practical things that we
00:31:34.780 should be concerned about they don't they shouldn't create shrill alarmism or panic but they are real
00:31:41.660 consequences that if we don't take them seriously then they have the potential to cause you know serious
00:31:47.420 harm and if if if we continue down this path of complete openness without any you know sort of
00:31:54.380 checks and balances on how this technology arrives in the world then essentially it has the potential
00:32:00.940 to cause a great deal of chaos and i'm not talking about ai's running out of control and you know and
00:32:06.780 robots and so on i'm i'm really talking about you know massively amplifying the spread of of
00:32:12.300 misinformation and more generally reducing the power reducing the barrier to entry to be able to
00:32:18.620 exercise power that that is fundamentally what this technology is i mean in my book i have a framing
00:32:25.180 which i think is more helpful around a modern turing test one that evaluates capabilities like what
00:32:31.820 can an ai do and i think that we should be much more focused on what it can do rather than what it can say
00:32:39.260 right what it can say is important and has huge influence but increasingly it's going to have
00:32:44.140 capabilities and so an artificial capable intelligence an aci is something that has the potential not just
00:32:53.180 to influence and persuade but also to learn to use apis and initiate actions queries calls in third-party
00:33:02.380 environments it'll be able to use browsers and parse the pixels on the browser to be able to click
00:33:09.180 buttons and take actions in those environments it'll be able to call you know phone up and speak to
00:33:15.660 communicate with other ai's and other humans so you know these technologies are getting smaller and
00:33:22.940 smaller and more and more capable are getting cheaper to build and so if you look out over a 10 to
00:33:28.940 20 year period i think the story is one of a proliferation of power in the conventional sense
00:33:35.660 not so much an intelligence explosion which by the way just for the record i think is an important
00:33:40.940 thing for us to think about and i care very deeply about existential risk and agi safety but i think that
00:33:48.140 the more practical risks are not getting enough consideration and that's actually a big part of the
00:33:53.500 book in no way does that make me a pessimist i mean i'm absolutely an optimist i'm hopeful and positive
00:34:00.700 about technology i want to build things to make you know people's lives better and to help us create
00:34:05.580 more value in the world and and reduce suffering and i think that's the true upside of these technologies
00:34:10.860 and we will be able to deliver them on that upside but no technology comes without risk and and we have
00:34:16.700 to consciously and proactively you know attend to the downsides you know otherwise you know we haven't
00:34:24.860 really achieved our full objective and that's the purpose of speaking up about it well before we get
00:34:30.700 into details about the downsides let's talk about how this might go well i guess before we talk about
00:34:39.260 the upside let's just define the terms in the title of your book the title is the coming wave what is
00:34:45.740 the coming wave so when you look back over the millennia there have been waves of general purpose
00:34:53.740 technologies from fire to the invention of the wheel to electricity and each of these waves to the
00:35:01.500 extent that they have been lasting and valuable are general purpose technologies which enable other
00:35:07.020 technologies and that's what makes them a wave they're enablers of other activity their general purpose
00:35:12.460 in nature and as they get more useful naturally people experiment with them they iterate they invent
00:35:20.540 they adapt them and they get cheaper and easier to use and that's how they proliferate so in the
00:35:27.740 history of technologies all technologies that have been useful that are real general purpose technologies
00:35:33.420 have spread far and wide and got cheaper and almost universally that is an incredibly good thing it has
00:35:41.180 transformed our world and i think that that's an important but very simple concept to grasp because
00:35:50.940 if that is a law of technology if it is a fundamental property of the evolution of technology which i'm
00:35:57.980 arguing it is then that has real consequences for the next wave because the next wave is a wave of
00:36:05.660 intelligence and of life itself right so intelligence is the ability to take actions it is the ability to
00:36:16.460 synthesize information make predictions and affect the world around you so it's almost the definition of of
00:36:24.300 power and everything that is in our visual sphere everything in our world if you look around you at this very
00:36:30.700 minute today has been affected in a very material way by intelligence it is the thing that has produced
00:36:37.980 all of the value and all of the products and all of the you know affected the landscape that you can see
00:36:43.260 around you in a huge way and so the prospect of being able to distill what makes us unique as a species
00:36:52.220 into an algorithmic construct that can benefit from being scaled up and paralyzed that can benefit from
00:36:59.740 perfect memory and compute and consuming vast amounts of of data trillions of words of data is is enormous i
00:37:06.940 mean that that in itself is is almost like you know gold it's it's like being able it's like alchemy it's it's
00:37:14.380 like being able to capture the essence of what has made us capable and add more knowledge and you know essentially
00:37:22.300 the science and technology into the the human ecosystem so imagine that everybody will now in
00:37:29.020 the future in 10 years 15 years have access to the very best you know doctor in the world the very best
00:37:37.820 educator you know the very best personal assistant and chief of staff and any one of these roles
00:37:43.900 i think is going to be very very widely available to billions of people you know people often say to
00:37:49.660 me well you know isn't aren't the rich going to benefit first or is it going to be unfair in
00:37:54.540 terms of access you know yes for a period of time that's true but we're actually living in one of the
00:38:00.380 most meritocratic moments in the history of our species every single one of us no matter how wealthy
00:38:06.860 you are every one of us in the western world really the top two billion people on the planet
00:38:12.860 have access to the same smartphone right no matter how much you earn you cannot buy a smartphone or a
00:38:19.740 laptop that is better than the very richest that's an unbelievably meritocratic moment that is worth really
00:38:27.020 meditating on and that is largely a function of these exponentials you know the cost of chips has
00:38:34.140 exponentially declined over the last 70 years and that's driven mass proliferation and if intelligence
00:38:41.980 and life are subject to those same exponentials which i think they are over the next two to three decades
00:38:49.260 then the primary trend that we have to cope with in terms of our culture and our politics and commerce
00:38:56.220 is this idea that intelligence the ability to get stuff done
00:38:59.980 is about to proliferate and that's going to produce a a cambrian explosion of productivity
00:39:07.660 everybody is going to get access to a tool that enables them to pursue their agenda to make us all
00:39:13.180 smarter and more productive and more capable so i i think it might be one of the most productive
00:39:18.860 periods in in the history of of humanity and and i think of course the challenge there is that it may
00:39:25.500 also be one of the most unstable over the next 20 years yeah so that um cornucopia image immediately
00:39:35.660 begets the downside concern of massive labor disruption which um many people doubt in principle they just
00:39:44.780 think that we've we've learned over the course of the last 200 years of technological advancement and
00:39:50.860 economic thinking that there is no such thing as a true canceling of a need for human labor
00:39:59.580 and so people draw the obvious analogies from agriculture and other you know previous periods of
00:40:05.340 labor disruption and conclude that this time is no different and we will while there might be a few
00:40:14.140 hiccups what's going to happen here is that all of these productivity gains and job canceling innovations born of ai
00:40:23.660 will just open new lanes for a human creativity and there'll be better jobs and you know we're just as
00:40:32.860 we were happy to get rid of jobs in agriculture and coal mines and open them up in in the service sector
00:40:40.540 we're going to do the same with ai i remain quite skeptical of uh that this time is the same given
00:40:49.420 the nature of the technology and this is this is the first as you just said this is the first moment
00:40:54.220 where we are envisioning a technology which is a true replacement for
00:41:02.300 human intelligence at every move if we're talking about general intelligence and we're talking about
00:41:08.220 the competence that you just described the ability to do things in addition to saying things
00:41:15.580 where we are talking about the cancellation of uh human work at least in principle and you know
00:41:24.540 strangely i mean this is not a not a terrible surprise now but it would have been a surprise
00:41:29.580 probably 20 years ago this is coming for the higher cognitive higher status white-collar jobs before
00:41:36.940 it's coming for blue-collar jobs how do you view the prospect of labor disruption here and how confident
00:41:46.700 are you that everyone can be um retrained with their um nearly omniscient ai assistants and chiefs of
00:41:55.740 staffs and uh find something worth doing that other people will pay them to do yeah i mean i i'm with
00:42:03.020 you i've i've long been skeptical of people who've said that you know this will be just like the
00:42:09.100 agricultural revolution or you know this will be like the horse and cart and and and cars you know
00:42:15.260 people will have more wealth the productivity will will drive wealth creation and then that wealth
00:42:21.820 creation will drive demand for new products and we couldn't possibly imagine you know what people are
00:42:26.620 going to want to consume and and what people are going to create with this new wealth and new time and
00:42:31.580 and that that's typically how the argument goes and i i've never found that compelling i mean i i think
00:42:38.860 that if you if you look at it it's been quite predictable the last decade i mean these models
00:42:44.860 are deliberately trying to replace human cognitive abilities in fact they have been slowly climbing the
00:42:52.860 ladder of of of human cognitive abilities for many years i mean we started with image recognition
00:42:58.940 um and audio recognition and then moved on to you know audio generation image generation and then text
00:43:07.660 you know understanding text recognition and then now text generation and you know it was kind of
00:43:14.620 interesting because if you think even just two or three years ago people would have said well ai's will
00:43:21.260 will will never be creative that's not achievable you know that that creativity will always be the
00:43:27.420 preserve of of humans and and and judgment is somehow unique and special to what it means to
00:43:32.540 be human or like you know ais will never have empathy will always be able to do care work and you know
00:43:37.980 emotional care is is something that's special you can never replace that connection i mean both of
00:43:42.700 those are now self evidently not not true and i think have been quite predictable so i think that the
00:43:48.620 the way to the honest way to look at this is that these are only temporarily augmenting of human
00:43:53.820 intelligence if you think about the trajectory over 30 years i mean let's not quibble over whether it's
00:43:58.700 five years 10 years or 15 years just think about it long term i think we can all agree long term if
00:44:03.100 these exponential trajectories continue then you know they're they're clearly only temporarily
00:44:10.140 going to turbocharge an existing human and so we have to really think okay long term what does it mean
00:44:16.860 to have systems that are this powerful this cheap this widely proliferated and that's where i think
00:44:23.820 the the broad concept i have in the book of containment comes in because you can start to
00:44:27.580 get an intuition for you know the the massive consequences of the spread of this kind of power
00:44:33.740 and then start to think about what are the what are the sorts of things we would want to do about
00:44:37.340 it because on the face of it like you said earlier the incentives are absolutely overwhelming i mean
00:44:42.780 technology has always been a machine of of of statecraft it's been used by militaries and used
00:44:48.940 by nation states to serve citizens and drive us forward and now it is the fundamental driving force
00:44:56.540 of nation states you know being commercially competitive having the best companies having the
00:45:01.180 best labor market that drives our competitive edge you know so from a state perspective a nation state
00:45:07.420 perspective you know from an individual scientific perspective the huge drive to explore and invent
00:45:13.420 and discover and of course from a commercial perspective the you know the the profit incentive
00:45:19.020 is is phenomenal and all of these are good things provided they can be well managed and provided we can
00:45:25.260 mitigate the downsides and i think we have to be focused on those downsides and not be afraid to talk about
00:45:31.020 them i mean you know so i definitely experience when i bring up these topics over over the years this
00:45:37.980 kind of what i describe in the book as a pessimism aversion you know there's there's people who are
00:45:42.700 just just sort of constitutionally unable to have a dark conversation about how things may go wrong
00:45:49.740 and i'll get accused of like not being an optimist or something as though that's like a you know a sin or
00:45:55.260 something or or or that being a pessimist or an optimist is somehow you know a good way of framing
00:46:01.260 things to me both are biased i'm just observing you know the the kind of facts as i see them and i
00:46:07.020 think that's an important sort of misconception and unhelpful framing of pessimism and optimism
00:46:12.380 because we have to start with our best assessment of the facts and try to reject those facts if they're
00:46:17.980 you know inaccurate in some way and then then try to collectively predict what the consequences are
00:46:22.700 going to be like and i think it's sort of another trend over the last sort of decade or so people
00:46:27.260 you know post-financial crisis i feel like people public intellectuals and elites in general and
00:46:32.860 everyone in general has sort of just like got a bit allergic to predictions right we've got a bit
00:46:37.740 scared of being wrong and i think that that's another thing that we've got to shed so we've got to focus
00:46:42.860 on trying to make some of these predictions they may be wrong you know i may have got this completely wrong
00:46:47.340 but it's important to lay out a case for what might happen and start taking steps towards you know
00:46:54.780 mitigation and adaptation well you invoke the concept of containment which does a lot of work
00:47:01.260 in the book and you have this phrase the containment problem that you use throughout what is the
00:47:08.460 containment problem in its most basic form the idea of containment is that we should be able to
00:47:15.260 demonstrate to ourselves that technologies that we invent should always be accountable to humans and
00:47:23.420 within our control so it's the ability to close down or constrain or limit a new technology at any
00:47:30.540 stage of its development or deployment and you know that's that's a grand claim but actually put in
00:47:37.180 the most simple terms it basically says we shouldn't allow technologies to run out of our control
00:47:43.580 right if we can't say what destiny we want for how a technology impacts our species then we're at the
00:47:50.700 mercy of it right and i think i think the the net the the idea is if if we if we don't have mechanisms to
00:47:57.420 shape that and restrict its capabilities then it potentially leads us into some some quite catastrophic
00:48:04.780 outcomes over a 30 or 30 year period do you think we've lost the moment already i mean it seems like
00:48:13.740 the the digital genie is is more or less out of the bottle i mean this is something that i mean if
00:48:18.940 anything surprised me and i know certainly surprised the people who are were more focused on on ai safety
00:48:26.140 and again people like utkowski in recent developments around these llms was that we missed a moment that
00:48:33.580 many of us more or less expected or more or less sure was coming which was there'd be a breakthrough
00:48:39.900 at some company like deep mind where we would you know the people building the technology would
00:48:46.300 recognize that they had finally gotten into the end zone or close enough to it so that they're
00:48:52.380 now in the presence of something that's fundamentally different than anything that's come before and
00:48:57.980 there'd be this question okay is this safe to work with is it is this safe to release into the wild is
00:49:05.820 this safe to create an api for is to say and you know so this the the idea was that you'd have this
00:49:12.620 you know digital oracle you know in in a box that would be um you know would already have been air
00:49:18.860 gapped from the internet and incapable of doing anything until we let it out and then the question would be
00:49:24.620 have we have we done enough safety testing to let it out but now it's pretty clear that everything is
00:49:30.620 already more or less out and we're building our most powerful models already in the wild right and
00:49:38.300 they're already hooked up to things and they're they already have millions of people playing with
00:49:42.540 them and they're they're open source versions of the next best model and so is containment even a
00:49:49.260 dream at this point so it's definitely not too late we're a long long way away this is really
00:49:55.180 just the beginning uh you know we we have plenty of time to address this and the more that these models
00:50:03.100 and these ideas happen in the open the more they can be scrutinized and they can be pressure tested
00:50:09.660 and held accountable so i think it's great that they're happening in open source at the moment so
00:50:14.620 you like sam altman's this is what sam has always said that the the philosophy behind open ai is do
00:50:21.820 this stuff out in the open let people play with it and we will learn a lot as we get closer and
00:50:29.100 closer to building something that we have to worry about i i think that we have to be humble about the
00:50:35.580 practical reality about how these things emerge right so the initial framing that it was going to be
00:50:42.620 possible to invent this oracle ai that stays in a box and we'll just probe it and poke it and test
00:50:48.300 it until we can prove that it's you know going to be safe and that we'll stay in the bunker and keep
00:50:54.140 it hidden from everybody i mean this is a complete nonsense and it's attached to the super intelligence
00:50:58.940 framing it was just a completely wrong metaphor that totally ignores the history of all technologies
00:51:04.780 and actually this is one of the core motivations for me in the book is that i had time during the
00:51:08.540 pandemic to really like you know sleep and reflect and really deeply think okay what is actually
00:51:14.140 happening here on a multi-century scale and what are the patterns of history uh around how inventions
00:51:21.260 end up proliferating and it's really stating the obvious it's almost like ridiculously simplistic but
00:51:27.180 it needed to be said that actually as soon as something as an idea is invented millions of other people
00:51:34.380 people have approximately the same idea within just weeks months years especially in our modern digitized
00:51:41.900 world and so we should expect and as we do see the open source movement to be right hot on the heels
00:51:49.420 of the absolute frontier and so i mean just one small example of that to give an intuition gpt3
00:51:57.420 it was launched in the summer of 2020 so three years ago 175 billion parameters and is now regularly being
00:52:05.500 trained at two billion parameters and so that is a massive reduction in serving cost you know that now
00:52:14.060 means that people can have open source versions of gpt3 that have broadly the same capabilities right but
00:52:21.820 are actually extremely cheap to serve and indeed to train so if that trajectory continues then we should
00:52:29.740 expect that what is cutting edge today frontier models like arza inflection and and like gpt4 gpt3.5
00:52:39.100 even will be open source in the next two to three years and so what does it mean that those capabilities
00:52:45.260 are available to everybody right and i think that is a great thing for where we are today but if the
00:52:50.300 trajectory of exponentially increasing compute and size of models continues for another three four
00:52:57.660 five generations which we all expect it to then that's a different question we have to step back
00:53:03.020 and honestly ask ourselves what does it mean that this kind of power is going to proliferate proliferate
00:53:07.500 in open source number one and number two how do we hold accountable those who are developing these mega
00:53:13.180 models even if they are centralized and closed myself included open ai deep mind etc and if you just look at
00:53:19.660 the amount of compute it's predictable and breathtaking and i think people forget how predictable this is
00:53:26.060 so going back to atari dqn we developed that model in 2013 and it used two petaflops of computation
00:53:36.300 right so a petaflop is a billion million operations right so imagine a billion people each holding one million
00:53:45.820 calculators each and doing a complex calculation all at the same time pressing equals right so that's
00:53:53.100 that would be one petaflop and atari used two petaflops over several weeks of computation a decade later the
00:54:01.180 cutting-edge models that we develop at inflection for pi our ai use five billion times the compute that was used to
00:54:09.820 play atari dqn so 10 billion billion million um it's just like now you're sounding like la g exactly
00:54:22.060 that's that's that's basically 10 orders of magnitude more compute in a decade so one order of magnitude
00:54:28.620 every year so 10 x every year for 10 years which is way more than moore's law everyone's familiar with
00:54:36.140 moore's law 70 years of doubling doubling every 18 months or whatever i mean that's that is minuscule
00:54:42.060 by comparison now of course there's a very good hand if you'd like to continue listening to this
00:54:47.180 conversation you'll need to subscribe at sam harris.org once you do you'll get access to all
00:54:52.540 full-length episodes of the making sense podcast along with other subscriber only content including
00:54:57.740 bonus episodes and amas and the conversations i've been having on the waking up app the making sense
00:55:03.500 podcast is ad free and relies entirely on listener support and you can subscribe now at sam harris.org