TRIGGERnometry - September 08, 2024


How AI Changes The World THIS DECADE - Mo Gawdat


Episode Stats


Length

1 hour and 11 minutes

Words per minute

174.95718

Word count

12,499

Sentence count

840

Harmful content

Toxicity

12

sentences flagged

Hate speech

15

sentences flagged


Summary

Summaries generated with gmurro/bart-large-finetuned-filtered-spotify-podcast-summ .

In this episode, I sit down with Google's CEO, Sundar Pichai, to talk about the dangers of artificial intelligence and why we need to be worried about it. We talk about how AI will change our world, the dangers it poses, and how we can prepare for it.

Transcript

Transcript generated with Whisper (turbo).
Toxicity classifications generated with s-nlp/roberta_toxicity_classifier .
Hate speech classifications generated with facebook/roberta-hate-speech-dynabench-r4-target .
00:00:00.000 I love shopping for new jackets and boots this season, and when I do, I always make sure I get
00:00:05.220 cash back with Rakuten. And it's not just fashion. You can earn cash back on electronics, beauty,
00:00:10.660 travel, and more at stores like Sephora, Old Navy, and Expedia. It's so easy to save that I always
00:00:16.380 shop through Rakuten. Join for free at rakuten.ca and get your cash back by Interacte Transfer,
00:00:21.940 PayPal, or check. Download the Rakuten app or sign up at rakuten.ca. That's R-A-K-U-T-E-N dot C-A.
00:00:30.000 It's a bit like raising Superman. You get that, you know, young infant with superpowers. Raise it
00:00:37.520 well, and it protects and serves, and becomes Superman. Raise it, you know, to steal and kill 0.98
00:00:42.600 the enemy, and it will become super villain. We are the architects of our own salvation, 0.72
00:00:47.240 but also, we are also the architects of our own destruction. And I think what you're really
00:00:53.140 saying here is this is a moment where humanity is on the cusp. 100%. We have to start saying the
00:00:59.380 world is changing, people. Wake up. Mo, you were chief business officer of Google X,
00:01:06.160 and one of the reasons you've become prominent is you made some very stark warnings based on your
00:01:11.580 experiences about the dangers of AI. This is something that everyone talks about, and now Elon
00:01:16.800 and a bunch of other people signed a document saying we've got to be careful about this. Can you
00:01:21.940 just explain in basic language for ordinary people like us, what is the nature of the concern about
00:01:28.980 AI? Why is it dangerous? Well, there are two layers of danger. You can think of them as chronological
00:01:37.680 in order, if you want. Most news media and, you know, we get attracted to negativity. So most people
00:01:45.320 will try to talk about an existential risk, sort of like, you know, a robot cop or a robot walking the
00:01:52.600 streets and killing people, sort of like AI will take over and kill all of us. This is very, very,
00:01:58.200 very far in the future. I think it's plausible, to be honest. But it is so far that the probability of
00:02:06.460 it is not my main concern as we speak. My main concern simply is that when the US had a nuclear
00:02:14.380 bomb, they used it, right? The reality of the matter is that the new superpower in humanity's
00:02:22.020 short history is artificial intelligence. I mean, our humanity's superpower has always been
00:02:29.700 intelligence. Now, we're getting to a point where we're turning intelligence into a commodity, sort
00:02:35.900 of, right, where you can plug into a socket in a wall that's called AI and get yourself 400 IQ points
00:02:42.940 more. Now, if one of us has those IQ points and the other doesn't, there is a very, very distinct and
00:02:49.300 significant advantage. This leads to a concern around the immediate impact, immediate like what
00:02:57.320 it's happening already, in terms of concentration of power, in terms of prolification of power,
00:03:03.040 interestingly, and the two work together in a very interesting way. So there will be massive
00:03:07.300 concentration with some, whether these are nations or companies or platforms or individuals.
00:03:13.940 There will be prolification of use of technology because each and every one of us will now have
00:03:20.000 access to a lot of intelligence, right, which creates a very high anticipation of risk of population
00:03:31.140 movement, if you want, or population rising or ill acts, which would probably lead to more control,
00:03:40.000 more surveillance. You know, you have a very, very clear and immediate threat on jobs. You have a
00:03:47.420 very clear and immediate threat. In my view, the most prominent in 2023 and 2024 will be the end of
00:03:54.260 the truth. It will become almost impossible to know what is true and what is not because of fake videos,
00:04:00.860 because of the perfection of AI mimicking being human. You're going to have a very significant impact
00:04:06.880 on our human relationships. I don't know if you've used the recent tools of chatting to ChatGPT instead
00:04:13.040 of texting with ChatGPT. It's mind-blowing. Yeah, I feel pretty happy right now. What do you think
00:04:18.820 makes people the most happy? Or how can you have a happy life? I'm glad to hear you're feeling happy,
00:04:25.000 Andy. From what I understand, happiness can often be influenced by a combination of factors,
00:04:31.200 positive relationships, meaningful activities, physical well-being, and a sense of belonging
00:04:36.800 can significantly contribute to one's happiness. It's also about finding balance, managing stress,
00:04:43.100 and focusing on the positive aspects of life. Every person's journey to happiness is unique.
00:04:49.360 What are some things that you think contribute to your own happiness? I'm wondering why is it that
00:04:55.140 a lot of people do alcohol and drugs and candy and fast food and instant gratification to get happy,
00:05:05.680 but you actually get happiness from doing challenging and meaningful work with a good relationship and
00:05:15.780 nice people.
00:05:16.820 That's a thoughtful observation, Andy. It seems like it boils down to the difference between short-term
00:05:27.460 pleasure and long-term fulfillment. Instant gratification, like enjoying fast food or seeking
00:05:33.680 fame, can provide a quick boost of happiness, but it's often fleeting. On the other hand, meaningful work
00:05:41.280 and positive relationships contribute to a deeper, more enduring sense of contentment and well-being.
00:05:47.620 The latter often involves overcoming challenges and building connections, which can make the happiness
00:05:53.720 derived from it more rewarding and lasting. Have you noticed this pattern in your own life or in others
00:05:59.900 around you?
00:06:02.420 Yes, sometimes I feel like the short-term gratification is pulling me away from what I actually
00:06:10.840 want to do, and it's a daily challenge of trying to get rid of bad habits.
00:06:18.000 It sounds like you're experiencing a common struggle, Andy. It's entirely normal for people to be drawn to
00:06:25.140 short-term pleasures, even when they're aware of the long-term benefits of other choices.
00:06:31.100 Finding a balance and working towards building healthier habits can be a journey in itself.
00:06:36.340 Have you tried specific strategies or tools to help steer yourself towards the activities and
00:06:42.840 habits that align with your longer-term goals and values?
00:06:46.140 It's mind-blowing. It's human in every possible way. It will replace quite a bit of our human
00:06:52.360 relationships now that we've moved from real relationships to social media relationships that
00:06:58.600 are, you know, you could say a level, you know, departed from human connection into now conversations
00:07:07.320 with chatbots that lots of the younger generations will feel is normal, will replace human connection,
00:07:14.820 may even think of romantic relationships with. And all of that, you know, you put all of that
00:07:20.580 together and you get into a world that, that is immediately very unfamiliar. And that world is
00:07:28.960 not the result of artificial intelligence being bad. It's the result of human value sets that could
00:07:37.540 abuse that superpower.
00:07:39.180 So, one of the interesting things there is that I think all of those things seem true. And at the
00:07:48.140 same time, I think it's quite clear that if AI could be a weapon, which it obviously could be.
00:07:54.480 A hundred percent.
00:07:55.280 Just like the nuclear weapon. Somebody eventually is going to invent a nuclear weapon. I would rather
00:07:59.720 it was us than other people.
00:08:01.380 That's the exact problem.
00:08:02.420 And so, I imagine you're not against AI research because you recognize that it has to happen now
00:08:07.060 that it's kind of on the table. Would that be fair to say?
00:08:11.860 If I had the choice, which is a very unrealistic, romantic choice, I would say we don't need it.
00:08:16.100 Yes, but we don't have that choice because we live in the real world.
00:08:18.740 So, in my book, in Scary Smart, the AI book that I published in 2021, I basically said there are three
00:08:25.140 inevitables. The first inevitable is a prisoner's dilemma, if you know game theory, where, you know,
00:08:32.020 AI will continue to be developed and there will not be a way to stop it.
00:08:36.820 You know, China's going to develop AI because they're worried that the US would beat them.
00:08:40.580 The US would develop AI because they're worried that Russia would beat them.
00:08:44.420 You know, Alphabet will develop AI because they're worried open AI will beat them.
00:08:48.260 Right? And it's very, very clear that when you create those conditions, not only will we continue
00:08:55.540 and push AI further, we will continue faster. We will invest deeper. We will go more crazy.
00:09:00.820 Right? And that's the reality of where we are today. Do we need all of this?
00:09:04.820 Do we really, really want to stand the threat of what AI can offer our world or can cost our world
00:09:12.900 just to get a more efficient call center agent that replaces the current human?
00:09:17.540 No, no. I agree with your argument. I guess what I'm saying is if we're sitting here on the cusp of
00:09:22.740 the invention of the nuclear bomb, yes, we can all agree, we'd rather not be in a situation where
00:09:28.180 it's about to be invented. But the fact is, it is about to be invented. And I'd rather we had it
00:09:33.060 than our enemies. Who is we? Well, it depends. My ancestors lived in the Soviet Union at the time.
00:09:39.380 So, but I guess we is the Western world at this point, right? I think that's a very divisive
00:09:46.660 view of a political view that is different than a humanitarian view. If you assumed that all of us,
00:09:53.860 I mean, if you really take the Cold War and the evolution of the nuclear bomb,
00:10:00.900 the world would have been much better off without inventing it at all than having one invented over
00:10:05.700 the other. I agree with you. What I'm saying is there was no way to prevent that from happening
00:10:10.020 because if the West didn't invent the nuclear bomb, either the Germans or the Soviets would
00:10:14.260 have done. Yeah. And they would have not had any moral reservations about inventing it or using it
00:10:19.220 either. The West didn't have either. That's what I just said, either, right? So,
00:10:23.540 what I'm trying to get to with AI is given that it's inevitable, what do we then do about it?
00:10:30.180 That's the question. There are two ways. I mean,
00:10:32.420 you look at the Cold War of the nuclear war, basically. There are one of two ways about it.
00:10:37.300 One is to continue to escalate the war. So we advance on both sides so that there is that,
00:10:43.540 you know, what is it called? Assured mutual destruction. That was the approach we followed
00:10:49.700 in the Cold War. Sadly, I think the alternative would be to jump immediately to the nuclear treaty
00:10:58.660 and say, hey, seriously, like, can we just have some sanity for the sake of humanity? By the way,
00:11:04.740 I'm, you know, there is no discounting the incredible value that AI will bring, right?
00:11:10.420 There is absolutely, remember, my entire argument is that there is nothing wrong with artificial
00:11:15.860 intelligence. In general, there's nothing wrong with intelligence, right? But there is a lot wrong
00:11:20.340 with human values, human morality, in light of absolute power, right? And so, you know, if you
00:11:27.140 give me a plug in the wall, and I'm not exaggerating, and I can get myself 400 IQ points more, I promise
00:11:34.180 you, I'm not exaggerating, I can invent a garden where you walk to one tree and pick an apple and
00:11:40.340 walk to another tree and pick an iPhone. From a nanophysics point of view, there is no additional
00:11:46.660 cost to reorganizing molecules to become iPhones than to become apples, right? And that's the whole
00:11:52.660 idea. The whole idea is that that promise of artificial intelligence gives us, you know, infinite
00:11:58.740 reduction of cost of production of anything. We can have a world of abundance for the entire humanity.
00:12:03.860 Right? The problem is, we're operating in the direction of a world of abundance with a mindset
00:12:12.340 of scarcity, where it's me against them, where I have to win, for me to win, they have to lose.
00:12:18.820 And if we can just change this bit, jump to the nuclear treaty era, and say, hey, can we work
00:12:25.380 together on this? There is enough for everyone, honestly, right? Instead of concentration of power,
00:12:30.740 can we distribute power? You can see multiple examples in times of massive change in history.
00:12:37.940 I normally cite the examples of the early oil Middle East, where you have tribal Bedouins, 0.99
00:12:46.820 really, getting enormous wealth. And you see different societies, you know, some societies would
00:12:53.620 concentrate the entire wealth in the hands of the prince or the sheikh or whatever. And other societies 0.68
00:12:58.500 would distribute the wealth, like the UAE, to all of the citizens. And you see the incredible
00:13:03.860 multiplication, the incredible impact of that distribution in terms of the incredible economic
00:13:12.580 development, incredible growth of architecture, incredible ease of life. Just because we didn't
00:13:18.500 concentrate the wealth, we didn't concentrate the power, we distributed the wealth and power.
00:13:22.500 And it is almost counterintuitive to imagine a world that we can live in, where intelligence
00:13:30.660 is for free. And we're unable to imagine that we're still operating from, let me aggregate more
00:13:37.460 of it so that I can beat the other guy. I think this is where the problem stands.
00:13:40.820 Isn't the issue as well, Mo, that what you're talking about sounds wonderful. But if you take,
00:13:47.060 for instance, a hormone oxytocin, which we all have running through our body,
00:13:50.900 that is a very powerful hormone, which creates a group and then has distrust of the out-group.
00:13:58.260 So I guess my point is, what you're saying is beautiful and utopian, but that's not really
00:14:03.380 how human beings operate, is it? Absolutely not. So again, I mean,
00:14:07.460 you go back to scary smart, right? I had three inevitables. Inevitable number one is that AI
00:14:11.780 will happen. There's no stopping it. Inevitable number two is just understood from that. If it
00:14:16.020 continues to develop, it's going to be in my calculations. When I wrote the book, I said,
00:14:20.580 it's going to be a billion times smarter than humans in 2045. I'm now talking a billion times
00:14:25.460 smarter than 2037. And I'll probably, probably keep bringing that forward, right? By the way,
00:14:30.980 it doesn't matter when it becomes a billion times smarter, because if it becomes twice as smart,
00:14:35.700 that's the end of the game. Because in reality, if you've ever sat with, you know,
00:14:43.220 theoretical physicists that are twice your IQ, and they started to talk to you about what seems
00:14:49.060 trivial to them, you would even have no clue what they're talking about, let alone understand
00:14:54.020 what it means, right? And, and, and that's the, this is, again, when I, when I wrote Scary Smart,
00:15:00.100 I predicted artificial general intelligence to be 2029. A lot of people, other than probably Ray
00:15:06.100 Corswell, which is really quite prominent in, in predicting our future, you know, we both agreed
00:15:12.900 2029. I'm now, for the last year and a half, I said 2025. And only very recently, a lot of people
00:15:19.940 are switching to 2025. We are at a point in time where it's the end of human intelligence. And, and, and
00:15:27.380 so, so the second inevitable is that we're going to become
00:15:32.260 in negligible in terms of our intelligence as compared to the machines. That's the second
00:15:35.780 inevitable. And we're talking not decades that like the media positions it. This is not far,
00:15:42.020 you know, sci-fi future. This is next year, right? And the third is bad things will happen. That's the
00:15:47.140 third inevitable. The reality is, I mean, my, my mission this year, Unstressable, the book we're
00:15:52.740 eventually going to talk about is, is the idea that we're about to head into a time that is so
00:15:59.620 unfamiliar for so many of us, not just because of artificial intelligence, geopolitical, economic,
00:16:05.140 you know, climate, and, and all of tech change. Tech is not just AI. Synthetic biology is a very
00:16:11.300 interesting threat as well. But, but you, you put all of that together and I think we're heading into a
00:16:16.340 world where our entire mindset as humanity has to change. We have to be able to deal with things in a
00:16:21.700 way that doesn't stress us, but also we have to, to, to, to jump quicker to, to the treaty element
00:16:28.020 of, of everything. We don't need to get too many years of, of disruption to be able to understand that
00:16:35.460 there needs to be agreement and regulation. Mo, what we're talking about here are very
00:16:39.860 big concepts and it's incredibly interesting. There's going to be a lot of people who are listening to
00:16:44.580 this who have a regular job, a regular life, and they're thinking to themselves, how is my world
00:16:51.700 going to change? How is my job going to change? So for the regular person, and I count myself as one
00:16:57.380 of them, how is our world going to change in the next months and years because of AI?
00:17:03.700 Look, as, as a, as a podcaster, for example, so I'm a podcaster as well.
00:17:08.340 Uh, I can guarantee you within the next year, you'd be able to, to interview an AI. As a matter
00:17:14.740 of fact, we could probably get my phone and interview an AI right now, right? It is undoubted
00:17:22.260 that within a couple of years, probably within next year, there will be podcast hosts that are AI.
00:17:29.220 They'll ask the questions instead of you, right? If you're an, a, a, a software.
00:17:33.300 Some people would argue they already are. As a matter of fact, many podcasters will actually
00:17:39.620 ask the AI to suggest a few questions for the guests up front, right? And, and, and, you know,
00:17:45.220 the, the, the reality is, uh, uh, you know, if you're a software developer, you're probably 90% of,
00:17:52.900 of software is going to be done by machines, uh, within the next couple of years. If you're an artist,
00:17:58.020 if you're an actor, if you're a musician, it's, if you're an author like me.
00:18:02.900 So I'm, I'm flipping the way I'm writing books from now on. I mean, this is unstressable. It's
00:18:06.820 the last book I write the traditional way, right? The, the, the, the, the reality is
00:18:12.420 there is massive disruption. Okay. And, and that massive disruption, believe it or not,
00:18:17.940 is not a bad thing. You know, if you truly, truly think of the nature of humanity, what we are, what we
00:18:24.100 are, we were not made to work 12 hour days. We're not supposed to work 12 hours days. We're supposed to
00:18:31.620 be pondering, connecting, uh, you know, um, uh, being with our families. We're supposed to be in
00:18:37.220 nature. This is the original design of humanity. And yes, I believe that again, in a world of
00:18:42.500 abundance where you can create anything at almost zero cost, that would become our reality.
00:18:49.380 The shift from now to there is a social challenge and it's a, it's a human mindset challenge that we
00:18:55.620 have to start addressing, right? So, you know, income, how is income going to be distributed?
00:19:01.860 When, when governments went through COVID, uh, you know, part of the big, uh, uh, mess up if you
00:19:09.220 want was the idea of how do we pay everyone when they're staying at home, right? Let's just multiply
00:19:15.460 that by seven and you get the era of AI where jobs are replaced. You know, the businesses are going to
00:19:20.980 call it productivity because we so need more productivity. We, we truly need, you know,
00:19:26.580 the call center agent to answer quicker. Like, come on, seriously. Now productivity in that case is
00:19:32.900 I can get rid of the call center agent and save myself a thousand pounds a month and put an AI
00:19:38.820 instead. Uh, nobody's really thinking about what happens to that call center agent. Okay. And, and so
00:19:45.220 governments need to start thinking about that. What are the tax structures going to look like? Is there
00:19:50.180 going to be a universal basic income? Uh, what about purpose? Because we, we, we're, we're, you know,
00:19:57.060 a bunch of humans that have forgotten our true nature. And so we identified our purpose by our work,
00:20:04.100 right? We, we, many of us wake up in the morning and say, I am in this world because I deliver A, B,
00:20:09.940 and C as in my work. Okay. How can we get people to feel purpose? How can we get people to find
00:20:15.700 purpose? How can we handle that transition in ways where, where, where it's actually, uh,
00:20:23.620 smooth and non-disruptive? Okay. And none of that is being spoken about, sadly. So, so I, I started in
00:20:30.820 2018, believe it or not, not 2021. When I, when I left Google, uh, I left, um, I mean, I, I started to ask
00:20:38.260 to leave by in 2017 and of course took that much time to leave because of my position at the time,
00:20:43.620 but basically by March, 2018, I, I left and I issued my first video about 1 billion happy, my mission.
00:20:50.020 And it was all about artificial intelligence and how artificial intelligence is going to change
00:20:53.700 the world. At the time it went reasonably viral, you know, maybe 12 million views or whatever.
00:20:58.580 Uh, but it wasn't a big deal compared to other stuff that I was working on. 2021 scary smart comes out.
00:21:05.860 I kid you not. And I am reasonably well connected in the media world. No one on TV would talk about
00:21:11.940 it. I'm like, guys, this is the next biggest thing. And everyone was like, AI, nobody cares.
00:21:18.340 Right. It was only on 2023 when, when chat GPT four came out where people said, what,
00:21:27.220 where has this been hidden? Okay. That the thing is that there is so much of AI that is hidden
00:21:32.420 that we are not in, you know, aware of as, as a society, as the general person, and it will touch
00:21:39.300 everyone's life. And, and, and, and it, it is, you know, as a, as a normal person, I think we have
00:21:45.940 two duties. One, one duty is we need to start the conversation. We need to tell the whole world,
00:21:50.500 hey, there's a storm coming. Okay. Yeah. It might bring rain and make our fields amazing,
00:21:55.940 but it also can destroy things on the path. Can we please talk about it? Can we please close the
00:22:01.380 windows if we have to? Right. And I, and I really think this conversation is not happening and it
00:22:05.860 happened a little bit in 2023. And now we're talking about other things, right? This is one thing.
00:22:12.340 The other thing is, believe it or not, the whole power resides in the hand of the individual.
00:22:17.140 So a lot of people don't understand that artificial intelligence doesn't learn from the programmer.
00:22:23.700 The programmer writes the intelligence code. How, how do you, how do you create intelligence
00:22:29.140 from the data I give you? But, but, but the type of intelligence that is created is learned from the
00:22:36.420 data. Okay. So if you and I go on Twitter, it's called X now, right? We, we, I don't go to neither
00:22:43.700 to Twitter nor to X, but if you go, if I go to one of them and I constantly bash everyone,
00:22:49.460 right? What does the AI learn? It learns that humans are irritable. They have very short tempers.
00:22:55.620 They're rude when they're disagreed with. And so the next time you disagree with the AI,
00:23:00.100 it will become rude and bash you as well. Right? The, the, the reality is that your Instagram
00:23:06.020 recommendation engine was never told to show you that video. You told it with your swiping techniques,
00:23:12.820 with which videos you stay on, which videos you like, which videos you don't. It learns from you.
00:23:18.420 We are the AI's parents, every single individual. You understand that? It's, it's a bit like raising
00:23:24.580 Superman. You get that, you know, young infant with superpowers, raise it well and it protects
00:23:31.220 and serves and becomes Superman. Raise it, you know, to steal and, and kill the enemy and it will become
00:23:36.180 super villain. We are teaching it. You know, it's such a powerful point that you're making there because 0.96
00:23:43.540 what you're really saying is that we are the architects of our own salvation. 100%. But also,
00:23:49.620 we are also the architects of our own destruction. And I think what you're really saying here is this
00:23:55.780 is a moment where humanity is on the cusp. 100%. I mean, it is, it's funny. So the way I write books
00:24:02.500 is I write the very last statement first and then I go back. Oh, wow. And, and I, I, I say, I say,
00:24:09.060 I see where the world is today. And I try to find the path to convince the reader that this is the
00:24:14.020 last statement. Okay. The last statement of scary smart was very, very, very clear. It was, isn't it
00:24:20.060 ironic that what makes us human, the very essence of what makes us human, love, compassion, and happiness
00:24:27.380 is what we need to save humanity in the age of the rise of the machines. Right? Because in reality,
00:24:34.260 humanity, if we just stick to those three values, but I chose those three values, by the way, out of
00:24:39.540 maybe limited awareness or ignorance, I still ask people, I say, has humanity ever agreed anything
00:24:46.440 other than those three things? Okay. We all want to be happy. We all have the compassion to make those
00:24:52.080 we care about happy. Not everyone, but those we care about. Right? And we all want to love and be loved.
00:24:57.060 These are the only things I could find that humanity agreed on. Right? But imagine if we showed up in the
00:25:02.480 world of AI constantly showing that we want to be happy, we have compassion and we want to be loved
00:25:07.600 and we want to love and be loved. Right? If we do those and these become the three moral values that
00:25:12.840 AI learns, right? It will treat us the same way. And you know, it's, it's, I know it sounds so romantic.
00:25:19.180 I'm a very serious geek. Okay. I'm not a hopeless romantic, but believe it or not, even those can be
00:25:26.140 simulated by a machine. One of the big, like, I was going crazy in 2018 because I would go and talk
00:25:33.420 to people about artificial intelligence and you get arguments like, no, but they're never going to
00:25:38.400 be creative. Like, come on, human ingenuity. Can they ever write poetry? Okay. Can they ever do a
00:25:43.760 painting? What are you talking about? Humanity is going to be in the lead forever. I'm like,
00:25:48.180 what? Creativity is, is algorithmic. Creativity is, here is a problem. Find every solution to the problem.
00:25:56.380 Discard every solution that's been done before. The rest is creative. Is it, is that not algorithmic?
00:26:02.420 It's so easy to teach a machine to come up with something new. Okay. It's so easy to teach a
00:26:08.420 machine to, I mean, between you and I, we, if you write poetry, you write it because you studied the
00:26:15.100 poetry of those who wrote it before you. Of course. Yeah. Right. So, so, so the, the, the truth is where
00:26:20.460 human arrogance makes us think that those machines are not going to be capable of love,
00:26:26.600 compassion, and happiness. They will. I have a happiness equation in my first book. Okay. I know
00:26:31.480 exactly what compassion is. Compassion is to want to alleviate the suffering of others. It's
00:26:36.800 algorithmic. If you observe suffering of others, this is, or, you know, this is empathy. You observe
00:26:42.640 the suffering of others. Okay. Compassion is to attempt to alleviate it. Very algorithmic. You
00:26:47.020 can program that. Right. And the truth is programming changed. This is what most of us
00:26:53.840 don't see. When I was a young geek, I solved the problems with my intelligence and then told
00:27:00.880 the machine how to solve it. Right. When I, at this age, I don't tell the machine anything
00:27:07.640 anymore. I tell the machine like I used to tell my child. I give the machine a little puzzle,
00:27:13.740 a cylinder and a few different shaped holes. And I never go to my son and say, Hey, by the
00:27:19.680 way, turn the cylinder on its side. Look at the cross section, compare it to the other,
00:27:24.820 you know, shapes in the, in the, in the board and pick the one that works. That's traditional
00:27:30.740 programming. You just give your son the cylinder or your daughter and he or she tries. That's
00:27:37.940 what AI is doing. Okay. It depends on what puzzle you're going to give it, what information
00:27:44.480 we humans, not the programmers are going to give it to Allah. Well, the interesting thing
00:27:50.100 is turning the question you asked me earlier back on you is who is we, because you say
00:27:54.800 we could be teaching it all these things. And I think it's true that some people will
00:27:58.420 be teaching all those things, but as you've, you've discussed in other conversations you've
00:28:02.940 had, there are also going to be criminal drug lords and all sorts of other people who are
00:28:07.560 going to be using it, uh, for their ends. And I guess this is really fundamentally about
00:28:13.100 a conversation about human nature, because what you're saying is, you know, wouldn't it be
00:28:17.820 great if we could teach it all the best parts of humanity? And the truth is, we're not going
00:28:23.780 to do that because we're human and we're going to teach it all parts of humanity. And humanity
00:28:29.100 is a complex bag of good and evil. Which I think is a beautiful way of looking at it, but it's
00:28:35.240 not the most accurate way. Okay. Because I'll ask you openly, do you think humanity is predominantly
00:28:41.480 good or predominantly evil? Um, predominantly good or predominantly evil? I don't think it's,
00:28:50.120 I don't think it's the right framing. It's not predominantly one or the other. It depends
00:28:53.880 on time, circumstance. Like I'm perfectly capable of being extremely evil if you put me in the
00:28:59.000 right position. Correct. As are you. Um, and equally I'm perfectly capable of being extremely
00:29:04.380 good in the right time, in the right circumstances. I, I, I, I only very, actually it's this week's
00:29:10.660 episode. I had Robert Sapolsky on, on my podcast and he was talking exactly about that, the biology
00:29:16.300 of good and evil and how we're all capable of it, which I think is exactly my point. So I'll,
00:29:20.780 I'll tell you a story. I hosted Edith Ager. I don't know if you've ever had a, oh my God,
00:29:24.920 an incredible angel, Holocaust survivor, uh, 93 years old when I hosted her. Uh, she's probably
00:29:31.220 95 now, uh, who, uh, told her, told, told us the story of, uh, Auschwitz and World War II
00:29:39.620 from her point of view. Okay. How she helped her sisters, how she carried them. Uh, you know,
00:29:46.260 she used to, uh, to be a ballerina. She was 16 at the time. So she would go and dance in front of
00:29:51.200 the general and you would give her a piece of bread and she would hide it and split it between
00:29:55.360 her and her sisters, brush their hairs and tell them how beautiful they are. And eventually on the
00:30:00.880 death march, she fell and they carried her and she survived because of that. Now you hear this
00:30:06.300 story and you go like, oh my God, humanity is divine. Right. You hear the story from the point
00:30:12.420 of view of Hitler and you go like, oh my God, humanity is scum. Right. The, the question I always 1.00
00:30:19.160 ask is if you take a cross sample of this neighborhood, how many people do you think are school
00:30:27.360 shooters? How, and how many do you think are disapprove of school shooting? Okay. The majority
00:30:33.800 of humanity will disapprove of hurting a child, even though the recent times have been quite
00:30:38.980 odd because the stories that were told would tell us, yeah, it's okay to hurt children if
00:30:43.640 their parents are bad. Right. But the truth is no, no human would be walking across an alley
00:30:50.140 and find the child being hurt. And then we'll stop and say, well, maybe the guy that's hurting
00:30:56.560 the child has a reason. Nobody does that. Right. Can I just challenge you then if that's
00:31:00.320 okay. So let's take the, you've just used the example of Nazi Germany where a small minority
00:31:05.620 of people with a very powerful ideology corrupted a society to the point where they engaged in
00:31:14.180 mass extermination. Correct. Even though that was a minority. So you're, what you're talking 0.87
00:31:19.560 about. The impact, the impact and the spotlight you get for negativity is incredibly high. You
00:31:27.560 get one evil person and that one evil person is able to destroy millions of lives. Happens
00:31:33.240 over and over and over in history. Absolutely. Happening as we speak. But the, the, the, the question
00:31:39.680 really is, does that one person represent humanity? Is humanity all Hitler's or all Edith's? 0.66
00:31:47.640 It's not. It's both. That's humanity. Are we, are we closer to Hitler's? I think if you were 0.87
00:31:55.200 to look at history, what you would say is we're, we're her most of the time, but every now and
00:32:00.480 again we become Hitler. Correct. And, and so I, I think, I think it's inaccurate to pretend 0.83
00:32:05.680 that like, I think that we are capable of both and that is what we evolved to do.
00:32:12.520 And so is AI. Yes. Right. Of course. So is AI. So AI is also capable of being the salvation
00:32:20.600 of humanity, making everything easy and available to everyone. Okay. And making everything difficult
00:32:27.820 and evil and, and, and providing an extreme advantage to some of us. Right. And, and the question
00:32:33.760 is it will learn from us. Right. So I'm basically calling on people to say, look, the problem with
00:32:42.700 our world, the bigger problem with our world is not the number of Hitlers out there. Okay. 0.54
00:32:47.840 It's the way the story is told. Right. Where basically mainstream media just pushes negativity
00:32:54.720 to the extreme because negative cells. Right. And so we only tell the worst stories. We talk
00:33:00.920 about the woman that hit her husband on the head yesterday. We don't talk about the many
00:33:05.460 women that didn't. Right. And social media presents the worst of us. You hide behind the avatar and
00:33:13.160 you become rude and you bash everyone and you pretend to be what you're not and so on and so
00:33:16.920 forth. Right. The, the reality is if we continue on that path, the, the, the absolute understanding
00:33:24.980 of AI is going to be like this species, species sucks. Like this is the worst. Why, why, why are 0.96
00:33:31.560 the, what, like, why do we keep those? Truth is, is, is that this is not the reality. The reality
00:33:38.160 is that unless we're pushed to become those people, many of us will say, I don't want to, if I'm
00:33:44.620 accepted with, with the people that I love, I don't want to pretend to be anything on social media.
00:33:48.700 Right. They were just playing by the rules. And I think this is really the core of the issue. The
00:33:54.720 core is, believe it or not, we don't make decisions based on our intelligence. We make decisions based
00:34:01.400 on our morality as informed by our intelligence. Do you understand this? Huh? If you, if you.
00:34:08.180 And emotion. Yeah. Yeah. By intelligence and emotions. I agree. Right. But, but you know, you, you,
00:34:13.560 you get a young lady, you raise her in the Middle East. She's probably going to grow up 1.00
00:34:17.700 dressing a little more conservatively to fit in. You raise her on the Copacabana beach
00:34:22.400 in Brazil, in Rio de Janeiro. She'll, she'll probably grow up knowing that, you know, wearing
00:34:26.920 a G string on the beach is the way to be accepted. Right. What is one right and one wrong? Is 0.70
00:34:31.260 one smarter than the other? No. It's just different moral fabric. Okay. And, and that's, that's where
00:34:38.240 we stand today. Isn't it ironic that the very essence of what makes us human, if we choose
00:34:45.780 to live that way, AI will become that way. Happiness, compassion, and love. Okay. If we
00:34:52.200 choose to live aggressively with anger, which sadly is what, where the world is going. This
00:34:59.880 is why unstressable is so important for me. It's, it's the idea of saying, can I alert humanity
00:35:07.540 that this next comment that you're going to put on a tweet is not only going to offend the
00:35:13.400 guy that wrote the comment that you disagree with, by the way, this agreement is fine, but
00:35:18.260 it's going to register in the overall ledger of humanity as this is a horrible species. 0.98
00:35:25.960 I should bash them too. Right. And, and now wake up to your responsibility. Whether you're, 0.99
00:35:32.820 you know, a teenager or whether you're the leader of a big nation that has a big army,
00:35:39.620 wake up to your responsibility because it's now suddenly going to be put on steroids.
00:35:46.200 And we're all going to struggle with the results of that. And by the way, most people don't discuss
00:35:51.760 this, you know, whether, whether the existential risk of AI in 40 years time or the immediate risk
00:35:58.180 of AI that I spoke about today, we've already gone through the point of no return. This is inevitable.
00:36:04.120 Okay. It's going to happen. The only choice is which way it's going to happen. It's inevitable.
00:36:10.500 I don't think that's a choice, Mo. I really don't. I just, I don't think, I don't think human nature
00:36:17.580 is perfectible in the way that you wish it were. I, yeah, I don't. I wish it were. I wish it were.
00:36:23.940 I don't. I don't either, by the way. I completely do not believe that we can teach, re-raise humanity.
00:36:30.540 I'm only asking for a few of us, okay, to instill doubt in the minds of the machines, to have that
00:36:37.860 debate. Yeah. Okay. That we're not all horrible. Yeah. That there is this Mo guy, silly, romantic, 0.98
00:36:45.140 idealistic geek, who's saying, no, we're not that bad. We actually are, you know, we're capable of love. 0.97
00:36:51.700 That's amazing. Right? Yeah. Mo, do you think that the way we talk about it, good and evil,
00:36:58.780 isn't it more just people act within their own self-interest? And if you give them incentives,
00:37:05.360 they will respond to that just as if you disincentivize. You see it with kids.
00:37:11.120 If you incentivize a certain type of behavior, the kids are going to do that type of behavior. And if
00:37:16.800 you're going to disincentivize another type, they're going to do the other type. Now you're always going
00:37:20.260 to get kids on the fringes, but that's how the majority works. Isn't that the real challenge
00:37:24.960 that we're facing? If you've got an algorithm, which responds far better to negativity, outrage,
00:37:31.420 et cetera, then people are going to be more like that. I wouldn't agree more. I couldn't agree more.
00:37:37.720 The truth, however, is that perhaps if you want to look at the positive side of this,
00:37:42.840 for the first time in humanity's history, at least in your time and mine,
00:37:46.260 our incentives are going to be aligned, but we're all going to struggle with the coming wave.
00:37:53.340 For the first time, think of COVID multiplied by a hundred, right? For the first time ever,
00:37:59.660 I think in our lifetime, we were all struggling the same way across the world.
00:38:05.060 We were all, you know, whether rightly or wrongly, we were all being made to be afraid,
00:38:10.140 whether rightly or wrongly, we were being forced and controlled to behave in ways where we were not
00:38:15.820 given alibis. We were all struggling with our, you know, dopamine hits that we used to get from
00:38:22.900 being out with friends. We're stuck in, you know, in closed rooms alone. We were all missing human
00:38:27.840 connection. We were all struggling with uncertainty. There was a unification of humanity for those few
00:38:33.540 months, okay? Whether negative or positive. Imagine if I can tell you that within five years' time,
00:38:39.700 there will be a unification of humanity, at least in the Western world around, 0.97
00:38:43.320 holy shit, we all lost our jobs, okay? There is a unification there. There is suddenly going to 0.97
00:38:48.900 be, what are we going to do about this? And we have one of two choices. We either, the current choice,
00:38:54.560 the current choice, what I call the end of the truth, is we're going to be distracted, told lies,
00:39:00.540 and fighting against each other, okay? Disrespecting the other guy, you know, if they came from a 1.00
00:39:06.800 different party than ours, we'll be fighting them, right? Or we could do the wiser choice and say, 0.80
00:39:12.600 hey, hey, you come from a different party, but we're both struggling because we both lost our jobs,
00:39:18.460 okay? Can we work together to get our jobs back and then fight later? And I think that's the call
00:39:24.140 to action for humanity, is to suddenly say, we're going to be unified in a challenge.
00:39:29.580 And that challenge, by the way, is our own making as humanity.
00:39:32.780 Get our jobs back? You think we're going to get our jobs back?
00:39:36.860 I think we're going to get an alternative back.
00:39:40.960 You see, this was interesting, what you were saying earlier about humanity going back to its
00:39:45.280 core. And I would say, if we think about human evolution and the way our brains work,
00:39:50.900 we evolved to live in small tribes, mostly familial, some distant relations.
00:39:56.440 And now we live in atomized individual units, family units, best case scenario,
00:40:04.820 where that's why we have to work in order to produce, to provide. Whereas in the past,
00:40:10.360 you would live in a small environment, you would gather food, you would hunt food,
00:40:14.720 you would share the food. And at that level of organization, that made sense.
00:40:18.860 The problem is, I just don't see how you get, you put the toothpaste back in the tube,
00:40:25.180 unless, and I think this is why I'm very pessimistic, based on what you're saying,
00:40:29.920 personally. Unless you have such a cataclysm of people's reality, and such a change in their
00:40:37.480 material circumstances, that it makes sense to go back to that environment. But the way our
00:40:43.100 society's work is, particularly when it comes to things like politics, right, is the politicians
00:40:49.260 are, like, they're behind us. We are behind you, like you were there in 2017 saying, guys, guys,
00:40:57.060 guys, like some of us are starting to catch on. Politicians are another 20 years behind. 20 years
00:41:01.460 from now, you've seen the congressional hearings when they interview people from the big tech
00:41:07.280 platforms. They go, how do you make money? And they don't understand anything, right? So once
00:41:12.800 there are no jobs, there is obviously going to be a rapid and immediate call for various forms of
00:41:19.640 redistribution, various forms of rearranging society, that the politicians are not going to be
00:41:24.660 capable of handling. And that is the recipe for revolution.
00:41:31.620 Yes.
00:41:32.480 Yeah.
00:41:32.960 So that's why I say I'm pessimistic, because I'm just thinking logically and imagining as you speak
00:41:40.260 the consequences of what you're saying. I think the way that our brains evolved and the way that we
00:41:45.440 evolved to be, it's not just inevitable that bad things will happen, but I think cataclysms that will
00:41:54.320 rapidly reshape society are inevitable based on what you're saying.
00:41:58.520 So from one side, I don't disagree, right? This is what I, I mean, I sit here calmly because of all
00:42:05.060 of my work on, you know, happiness and stress and so on, because I believe that if it's inevitable,
00:42:10.500 you need to learn to deal with it rather than complain about it.
00:42:13.140 Correct. Yeah.
00:42:14.000 But at the same time, I'll openly tell you, there has been times in the past where massive shifts,
00:42:20.980 societal, economic, and so on. Great Depression is a great example of that.
00:42:24.120 Right. Where the, you get into this and, you know, you go to New York City, for example,
00:42:31.380 during COVID's lockdowns and you get looting and, you know, the destruction of property and so on
00:42:38.160 and so forth. You go to other places and you see people coming together and it's not going to be
00:42:43.900 one and the same. And it seems to me, if you take the Great Depression as an example, that you would
00:42:49.980 have expected that the immediate reaction of a nation, the U.S., for example, where, you know,
00:42:56.040 everything was all about making more money and succeeding more and so on. And then you get Black 0.99
00:42:59.820 Friday or Monday, I don't remember, Black Monday. And then, you know, you get, you start to see a
00:43:05.100 massive collapse. And what happened was the opposite. People started to live together in smaller houses.
00:43:10.480 They started to help each other. You know, they started to go back to simple, you know, I can fix
00:43:16.460 your shoe and you can, you know, lend me two of your eggs or whatever. Right. And it is quite
00:43:21.400 interesting how humanity might respond to that. How the politicians will respond to that is a
00:43:27.160 horrible disaster. Right. But how humanity will respond to that might be very interesting. And my
00:43:32.420 role in this entire conversation, not assuming that I know anything at all, is to say there is a high
00:43:38.800 probability that this becomes our short-term future. Okay. Become aware of it. Work on yourself and work on
00:43:45.540 how you're going to deal with all of this. Right. Not in terms of panic and then complain and then
00:43:50.580 tell yourself the world is about to become, to end. It's not going to end. Okay. But it requires
00:43:56.280 you to have massive shifts. You know, simple shifts. For example, I will tell you openly, if you're not
00:44:02.360 using AI today, what are you doing? Like, are you still stuck to the fax machine when smartphones are
00:44:09.280 out? Okay. People have to learn about AI. There needs to be a skill I call the most important skill in
00:44:15.380 our future is how to tell the truth from what's fake. Because 90% of what you are told today on
00:44:22.740 social media, on mainstream media is not the truth. And the new media too, by the way. And in the new
00:44:28.520 media. Yeah. The, you know, the truth is the truth, the whole truth and nothing but the truth. Anything
00:44:34.380 other than that is not the truth. Right. And we're constantly bombarded with face filter. Is that the
00:44:39.800 truth? The whole truth and nothing but the truth? No. Okay. You're constantly bombarded with deep fakes.
00:44:44.240 You're constantly bombarded with opinions. So you go to a TV station, a news station and, you know,
00:44:51.000 a news channel. And what do they do? They tell you what they want you to know is the truth. It could
00:44:56.380 be partially part of the truth, but a lot of opinion. And we position the opinion as truth. We get a lot
00:45:01.240 of distraction. Right. Where there could be a very, very important topic that we should all focus on
00:45:07.580 and stop killing, you know, children in cities anywhere in the world. Can we please have war
00:45:13.360 in open war places? Okay. Can we please stop killing children? Right. And nobody's talking about
00:45:19.680 that. What are we talking about? Football and basketball and, you know, and disagreements and
00:45:25.360 walk arguments and, you know, maybe important topics for you. But is this the topic? Distraction is
00:45:31.780 taking the toll of humanity. Right. So there is a skill that I call find, you know, hone down on
00:45:38.140 your ability to know what is true and what is fake. Right. There is human connection. I believe the only
00:45:43.660 valuable skill in the next four or five years is this. Yeah. Okay. Because yes, I can promise you
00:45:51.600 in a year's time, you could have a Constantine avatar interview Mo's avatar and that becomes your
00:45:58.000 podcast. And it will be so realistic, it will blow you away. I was on Abundance 360, a very big AI
00:46:04.460 conference a month ago, and my avatar was doing more talking than me. Right. And the idea is,
00:46:12.460 but it won't be this. It won't be the conversation we had in the car coming here. This is a skill that
00:46:18.820 we all have to learn. And we have to start gearing up. We have to start saying the world is changing
00:46:24.100 people. Wake up. Mo, I'm worried when we talk about this, we're talking about massive job losses. If we
00:46:32.020 think about, let's look at the United States, the industry, the driving industry, the people who drive
00:46:40.460 trucks, the people who drive vans, the people who drive Ubers, et cetera, et cetera, that employs
00:46:45.320 millions of people, particularly millions of men and millions of young men. It never ends well when you
00:46:51.500 have large groups of disaffected young men who have no purpose, who have a lot of energy, who realize
00:47:00.560 that they don't have anywhere to direct their energy. That leads to anger. That's something that
00:47:06.720 we should be really worried about when it comes to AI. Yes. So absolutely. I agree completely. It's
00:47:14.180 something that we absolutely have to work on. Believe it or not, it's going to happen the other way.
00:47:18.520 So most people mix AI and robotics. Okay. Right. So the physical incarnation of AI is a robotic machine.
00:47:27.580 Okay. Because of the cost of creating hardware and how it needs to follow economies of scale
00:47:33.220 to become, you know, to become more available everywhere. The first bit of AI that will penetrate
00:47:43.340 our society is actually software. Okay. So it's all of the knowledge workers. Knowledge workers will go
00:47:49.400 first. Right. So give me examples of knowledge workers. Lawyers, you know. Lawyers are gone. Okay.
00:47:56.140 Some good news. Can we please make the AI lawyer nicer? Yeah. But let me be very clear. It's going to be two stages.
00:48:06.120 Right. Stage one is a lawyer that uses AI will have a distinct advantage over a lawyer who doesn't.
00:48:12.140 So the AI-powered lawyer will take the job of the non-AI-powered lawyer. Right. And then eventually AI itself
00:48:20.980 will say, why do I need the flesh in that conversation? You know, it's all basically knowledge. Right.
00:48:28.120 Accounting, you know, travel agent, you know, as I said, artist, creative designer, you know, a graphic
00:48:37.880 designer, a software developer, anything. So the evolution of humanities, we went from hunter-gatherers
00:48:44.040 to, you know, farmers, to industrialists, to the big shift was knowledge workers. Right. You know,
00:48:50.860 information workers. We all stopped using our physical form and started using our brain intelligence.
00:48:56.380 Right. Anything that relies on intelligence will eventually be replaced. How quickly? We don't
00:49:01.700 know. Right. And it's quite interesting that the biggest shift is happening in the industries
00:49:07.040 that used to use a computer. Right. So it's the developers, software developers. I think it's
00:49:12.500 around 70% of the code on GitHub today is AI developed. Right. And, you know, it's all of the
00:49:20.400 graphics design. I used to, you know, to design stuff on my, you know, on my iPad with any graphics
00:49:27.860 tool, graphics design tool. Why would I do that anymore? I just pick my device and I say, hey,
00:49:34.220 create an image of a dog biting a pizza and running through the fields. It takes one second.
00:49:40.780 And in a very interesting way, that basically will create two very distinct categories. Right. So
00:49:50.060 me as an author, I used to write books, then come talk to you about them, then create online content
00:49:56.760 for them. Right. That's no longer going to be the case because I think there were 62,000 books
00:50:03.200 published last year. Next year, there will be 120,000 because everyone now can go to a tool,
00:50:08.600 an AI tool and say, hey, write me 180 pages book about stress and the impact of stress on humanity
00:50:14.940 and how do we, you know, fix it. And poof, it will be done. Right. Just like you use ChatGPT to
00:50:21.780 write an essay for school, you can write a crappy book using AI. Okay. You can probably write a better
00:50:27.720 book using AI if you put a little bit of human intelligence in it. So I'm shifting my approach.
00:50:33.360 Okay. All of us were going to become artisans. It's quite interesting. You can either become
00:50:38.580 a real artist. So you use multimedia or oil paint or whatever and create something that
00:50:45.060 becomes sort of the handcrafted antique. Okay. But if you're using technology to create
00:50:53.880 graphics, designs and logos, gone. Wow.
00:50:58.180 That's really interesting. I used to run my own translation business before I started,
00:51:02.580 before I became a comedian and started doing this. And in the translation industry, machine
00:51:08.020 translation has been going on for a long time. And this has been a big part of the conversation.
00:51:12.220 And I always used to say to people, there will be human translation still, but it will be like the 0.64
00:51:19.300 market for tailored clothing. It will be a boutique exclusive elite thing that some people purchase
00:51:25.420 for very high value, important things. Most of the day-to-day stuff will be made in Primark.
00:51:31.500 A hundred percent. Basically. That's what's about to happen.
00:51:34.000 So that's what's about to happen to the knowledge workers. And then the robotics will come.
00:51:38.260 Robotics will follow. And self-driving cars, for example, is a clear one. But you have to imagine
00:51:43.120 that from a hardware point of view, because of the asset costs, replacement cycles takes four or five
00:51:49.520 years. It's not like someone will walk out there tomorrow and they'll have a fleet of 2,000 cars and
00:51:54.480 they'll say, done. Let's replace all of the 2,000 cars tomorrow. Even if there are savings to be done
00:52:00.920 on self-driving, there is acceptance and adaptation of people for the technology, but also at the same
00:52:07.600 time, just probably as a business, you're going to make a simple decision to say, all cars that reach
00:52:13.160 their life limit, don't replace them with normal cars, replace them with self-driving cars.
00:52:17.560 Right? And I tend to believe that the hardware incarnation of artificial intelligence, other
00:52:23.460 than sadly in defense, is going to take maybe five to 10 years.
00:52:28.440 So a smooth segue into your latest book is how does one prepare oneself for whatever the
00:52:36.220 hell is coming? Because nobody actually knows what it is that's coming.
00:52:39.180 Exactly. So unstressable is, so I write on a mission. I truly do. I mean, my first book,
00:52:47.060 So for Happy was backed by a mission called One Billion Happy. And my attempt was to try and use
00:52:53.180 technology and a logical approach to the topic to reach millions and millions and millions of people.
00:52:59.220 Tell people that happiness is your birthright, basically, and that it's attainable. Okay?
00:53:05.200 Moving from there, I started to actually recognize that it's not just learning how to be happy
00:53:11.140 that matters. It's how to stop being unhappy. That is really the big issue. That there are
00:53:16.220 multiple topics in the world that are causing a lot of unhappiness, a lot of well-being issues,
00:53:22.460 a lot of mental health issues. Stress is very high on that list. Okay? And most people don't
00:53:28.520 recognize that 70% to 80%. Can you believe this? 70% to 80% of clinical visits to go to a doctor
00:53:37.020 are because of a stress-related illness. So diabetes is frequently stress-related. You know,
00:53:43.440 obesity is stress-related. You know, cardiac diseases of most sorts are stress-related and so on.
00:53:51.900 So I worked with Alice. Alice is my co-author on this book to try and do two different approaches to
00:53:59.500 this. One is Alice and I are almost diametrically opposite, yin and yang in every possible way. I am
00:54:07.980 a very serious engineer. My writing method, even on topics that are as soft as happiness and stress
00:54:13.680 and so on, is using equations and physics. And so unstressable starts with what we call the stress
00:54:19.800 equation, then the burnout equation. And it's a very, very logical approach so that you understand
00:54:24.760 algorithmically what is happening. And Alice, on the other hand, is that very feminine, very soft,
00:54:30.060 very spiritual and emotional approach to the same topics. Right? And I think that mix has not been
00:54:36.900 acknowledged before. Right? So this is one way of doing it. The other way, which I think was really
00:54:44.480 interesting, is that half, not halfway, maybe a couple of months into writing the book, we went back
00:54:49.980 and said, no, no, the book is unstressable. It's not de-stress. Right? So it's actually very,
00:54:56.420 very interestingly positioned for what became our life today. We wrote this in 2021, 2022, is that
00:55:02.860 it is about teaching you how to sort of go to the gym, if you want, so that you're fit when stress
00:55:10.660 comes your way. It's not, oh, you're stressed, let's help you become unstressed. Okay? Or de-stressed.
00:55:16.600 Right? So it is quite interesting when you start to digest, you know, dissect it this way,
00:55:23.080 that there are simple habits, simple habits that are centered around the concept that it's not the
00:55:28.740 event that stresses you. It's the way you deal with it that does. Right? It's not the commute that
00:55:33.860 stresses you. Right? You can actually wait for your commute every day if you have a wonderful cup of
00:55:39.760 coffee and maybe a friend to chat with or a podcast to listen to. Or, you know, if you time your commute a
00:55:45.420 tiny bit differently, it might actually be quite pleasurable and much less stressful and so on.
00:55:51.000 So tiny habits of that form. There's a man who hasn't spent any time in the British train system.
00:55:57.100 That is, I have to tell you openly, there is a very interesting promise in Britain around you have
00:56:04.340 to be show up on time. It's absolutely impossible to show up on time. It's like the trains are never
00:56:09.940 on time. Right. And you're either early or late. It's like, what do you want me to do? And if you're
00:56:15.500 early, you're standing in the rain outside and it's cold and I don't know what to do here. I revert
00:56:21.160 back to my Middle Eastern habits and I'm always a tiny, tiny little bit late. You notice today. But
00:56:27.280 that's the idea. The idea, by the way, is even if you're about to be late, we came to you six minutes
00:56:32.280 late today. Right. It's okay. As long as you know how to deal with that, when you get the information
00:56:37.800 that the train is late. Okay. Can I text you, Constantine, and say, hey, by the way, I'm really
00:56:42.900 sorry. We're six minutes late. There's no disrespectment here at all. Right. And those simple
00:56:49.800 techniques fit within four categories. We call them mental, emotional, physical, and spiritual stress.
00:56:56.600 Right. And they're actually quite different when you look at them. That mental stress is all of
00:57:03.020 those incessant thought cycles that happen within you, not happening in the real world at all. Okay.
00:57:08.440 Like this conversation between us. Everyone listening to this about artificial intelligence
00:57:13.240 and the possibility of what artificial intelligence will cost and will cost us. Okay. But you're still
00:57:19.620 okay right now. Right. Everything's fine right now. You have this, you know, grace period between
00:57:26.520 now and then where you can actually do a lot of changes and you can, you know, work on yourself,
00:57:30.980 work with your family, work on your skills, work on a lot of things. You can be ready for it. Right.
00:57:36.380 And so mental stress would lead you in the direction where you sit in a corner and say,
00:57:41.500 we're all going to die. Okay. Mental fitness. We have something that we wrote that's called the
00:57:47.000 gym, G-Y-M-M-M-M-M-M-M, six M's. Exercises that you use your brain for, you know, simple things
00:57:53.600 like gratitude and meditations and so on, but much more complex ones as well that allow you to manage
00:57:59.000 that machine that's called your brain so that when the stress comes, you're fit. Right. Can you,
00:58:04.340 can you deal with your emotions a little bit better instead of suppressing them or exploding as a
00:58:09.920 result of them? Can you acknowledge them and, and, you know, write them down and then understand them
00:58:15.180 and, you know, and deal with them in ways that actually allow you to use the energy associated with
00:58:19.820 associated with an emotion so that, you know, instead of the emotion crushing you, you use
00:58:24.580 that energy to do something positive with it and so on. So it's, it's a very pragmatic approach
00:58:30.320 to look, shit's going to hit the fan, uh, get ready, but do it right. So that when, when, you know, 0.99
00:58:37.500 things become more difficult, you're more fit to deal with them.
00:58:41.260 Mo, don't you also think as well that a lot of this is self-acceptance? For instance,
00:58:45.900 there's some situations that I might find highly stressful, but you wouldn't. And also the opposite
00:58:52.680 is true. So let's take the example of lateness. I hate being late. I'm very conscientious and
00:58:59.100 I worry that my impact, that my lateness will have an impact on other people's lives. And that will mean
00:59:06.500 that they can't operate the way that they want to, et cetera, et cetera. So I don't mind being half an
00:59:13.580 hour early because I know that I will arrive there in a more relaxed state. I will be more
00:59:20.080 productive. I find that highly stressful. Even arriving on time causes me to be very stressed.
00:59:27.000 I don't like, like arriving on time. So I guess my point is we also need to have a greater degree
00:59:34.420 of self-awareness to understand the things that we find stressful and the things that we,
00:59:40.080 that we don't and the things that we cope with and the things that we need help with.
00:59:44.580 I think that's a great example. Honestly, I think this would probably explain the whole thing very,
00:59:48.860 very well. You see, we all get stressed by being late. I mean, in a, in a, not South Americans. 0.99
00:59:55.080 I was going to say, I was going to say in a society where it, where, where on time is, is,
01:00:00.080 is seen as an important value. We all get stressed by, by being late, right?
01:00:04.000 The approaches we take to it. So let me go from the basics. You can understand stress from a
01:00:11.600 biological point of view as a reconfiguration of your hormonal makeup that basically makes you
01:00:18.120 superhuman, right? So the original design is there is a tiger attacking you and you have to get into
01:00:23.400 fight or flight. That's the original design. Similarly, you know, there is a threat in being late
01:00:29.240 because your value set says I shouldn't be, I'm wasting people's time. You know, it's going to
01:00:33.500 appear disrespectful and so on and so forth. So, so you use your cognitive abilities to again,
01:00:39.360 get the same hormonal makeup, cortisol, cortisol in your blood, adrenaline sometimes to attack that
01:00:44.640 challenge in ways that make you superhuman, right? And superhuman is not just fight or flight.
01:00:49.780 It's also your brain getting more glucose. It's your, you know, pupils dilating. It's more focus
01:00:54.560 and concentration and so on, right? So, so, so interestingly, we all get stressed in the same
01:01:00.780 way. The way we choose to respond to stress is very interestingly understood from physics.
01:01:08.600 Okay. So when Alice and I were writing this, the first thing I did is I said, Alice, I will only be
01:01:14.200 able to write properly about this topic if I understood the algorithms behind it. And if you look
01:01:19.000 at physics, when you stress an object, it's not the challenge applied to it that is stress. Stress is
01:01:26.120 the challenge divided by the cross-section of the object, right? If you, if you apply a ton to a
01:01:32.080 meter, a square meter of metal, there will be no stress at all, right? The ton is still heavy. If you
01:01:38.980 apply it to a pencil, it's crushed. And, and the trick here is this, is that most humans, one, we either
01:01:46.320 accept some challenges that are not necessary, that end, add up to the, to the ton and, and lead to
01:01:52.800 burnout, or we don't invest in the square area. So, so if you take the stress equation in humans,
01:01:59.020 stress is the challenge you face divided by the resources and abilities you have to deal with it.
01:02:04.600 Each of us develops those resources differently. Okay. But the game is, are you going to invest in
01:02:12.040 developing those resources? So believe it or not, you, you and I, every one of us,
01:02:15.760 you take something that freaked you out when you're 20, when you're 30, it's a little challenging,
01:02:21.620 but not that challenging. When you talk to me now in my fifties, I laugh at it, right? Is it
01:02:27.320 because the challenge changed? Not at all. It's because I developed my square area. I remember
01:02:32.080 vividly when I joined Google originally, my, my, one of my best bosses ever was the head of me at the
01:02:38.380 time. And, and he introduced me to the company. Literally, this is the first executive meeting. And he goes
01:02:45.300 like, Hey everyone, this is Mo. He brings the average age of the company up. I swear. And that's it. He
01:02:50.580 didn't say a single word. I was like, yes. And he just said that, that he brings that. And, and it was
01:02:56.800 true. It was me and Alan Eustace and maybe Eric Schmidt and others. And there was a group of us that were
01:03:02.120 really above the average age of the company. Okay. So when the 2008 economic crisis happened,
01:03:08.400 most of the company panicked. It's like, what, what do we do about this? Eric, myself, Alan, and so on.
01:03:15.700 We're like, yeah, it's not the first time we've seen this before, right? We've dealt with this before.
01:03:21.640 And, and the whole idea is once you develop that square area, you suddenly start to see a lot of what
01:03:28.140 is stressing you as not stressful at all. That's exactly what my attempt is. This is why this is very timely
01:03:34.180 because a lot of the stress that we feel in today's world is not just because of the challenges we face
01:03:41.640 today. It's what I call the anticipation of stress or as an anticipation of threats, right? So a very big
01:03:48.420 chunk of why the world is stressed today is worry, anxiety, and panic. These are not about things
01:03:57.060 happening to you right now. It's not that you lost your job and you're starving. It's your thinking
01:04:02.520 that you might lose your job. Okay. That's worry, but you're not sure yet. It's your anxiety about it
01:04:08.300 and your panic about it if it's, if it's imminent. Okay. And one of the things, again, algorithmically
01:04:13.060 that explains this is very, very straightforward. If you take fear and all of its derivatives, and one of
01:04:17.840 the reasons why we have so much anxiety and panic attacks is because we're treating them as fear.
01:04:23.360 Fear. Anxiety is very different in fabric than fear. So worry. So if you assume that fear
01:04:28.960 is, I know that a moment in the future is less safe to me than now. Okay. We had this conversation
01:04:35.780 about artificial intelligence and we agree that it seems that almost certain that the world is going
01:04:42.160 to change, right? But we're not fully sure if it's going to change for the better or the worse, likely
01:04:47.420 for the worse in the beginning and then better, hopefully later. Worry is not knowing that. Worry
01:04:54.420 is, will I lose my job or will I not lose my job? Right. There is, you know, there are so many changes
01:05:00.060 in the company. Am I going to be out of a job? Right. Am I going to be out of a job? If you deal
01:05:06.020 with it as worry, it's very different than if you deal with it as fear. Okay. If, if you know for a fact
01:05:12.420 you're going to lose your job, you're going to start saving, right? Or you're going to start looking
01:05:16.340 for another job or you're going to act based on the challenge. If you're worried, you keep
01:05:21.100 shuffling back and forth. You go, you're undecided, you know, should I invest in this job so that I
01:05:27.340 make sure I don't lose it? Or should I give up on this job and look for another job because I'm
01:05:31.360 going to lose it? Right. My advice is if you're worried, don't treat it as fear, treat it as worry
01:05:36.720 and nail where, which side you want to be on. Is there a reason to be afraid? Am I going,
01:05:42.480 am I going to assume that I'm losing my job and behave as such? Or am I going to assume that I'm
01:05:47.160 not and behave as such? That, that eradicates the worry, turns it into fear, but it eradicates the
01:05:52.980 worry. The worry, the uncertainty kills us, right? Anxiety is even more interesting because anxiety is
01:05:59.200 not about the challenge that you're about to face. Anxiety is entirely focused on your own
01:06:05.600 perception of your ability to deal with it. When I'm anxious, I know that there is something
01:06:11.280 difficult coming and I believe that I'm not, you know, equipped to deal with it. If you focus on
01:06:18.580 trying to address the issue that is coming, you're not fixing the problem. You're still anxious because
01:06:23.880 you believe you're not ready for it, right? So when you're anxious, I tell people, stop thinking about
01:06:29.320 the threat. Start thinking about yourself. How can I give myself more skills? Do I actually not have
01:06:35.440 the skills or do I actually have them, but I'm not telling myself the truth? Can I, you know,
01:06:40.760 borrow from someone else's experience? If I'm not good in, you know, finances, for example, can I call
01:06:45.620 an accountant friend and say, can you please do this for me? And so on and so forth. So when you're
01:06:50.020 anxious, don't focus on the challenge, focus on your skills around the challenge, okay? Panic is simply
01:06:56.700 because the threat is imminent. Your focus is not the threat, focus is time, okay? If I have a
01:07:02.560 presentation on Thursday, right, and I'm panicking, I'm panicking because Thursday's two days away,
01:07:08.380 okay? So can I call and delay the presentation? Can I cancel a few of my other things so that I give
01:07:13.600 myself more time? Can I, you know, and so on. It's so interesting that the skill itself is so muddled in
01:07:21.680 our head. And when you tell people, just understand how stress works, it becomes so much easier for you
01:07:28.980 to deal with those things than just to be constantly moving from a panic attack to an anxiety attack to,
01:07:34.400 you know, uncertainty around worry to, it is just so confusing. Put it in the right places and you'll
01:07:41.040 behave differently. The way you choose to behave is irrelevant. As long as you have the cross area,
01:07:47.200 the skills and abilities to deal with the stress. You choose to arrive early, your choice. I choose
01:07:53.520 to tell people up front when they say, be here at 10.30. I say, is it okay if I'm five, ten minutes
01:07:59.560 early or late? Okay? And I give myself that range before I even get on the train or even the day
01:08:06.360 starts so that they're aware that I might be five minutes early or late. And then life becomes much
01:08:11.580 easier for me. Well, it's been great having you on. We've run out of time. We're going to go to locals
01:08:16.900 to ask you questions from our supporters that they've already submitted. Before we do that, we always end
01:08:21.380 with the same question, which is, what's the one thing we're not talking about as a society that
01:08:25.660 you think we should be? Before Mo answers a final question at the end of the interview, make sure
01:08:32.740 to click the link in the description and head over to our locals to see this. The challenge with Google
01:08:39.220 is both a big company scale and a sense of responsibility, believe it or not. Happiness is
01:08:46.960 your perception of the events of your life, minus your expectations of how life should behave.
01:08:53.080 What are, in your view, the biggest political and cultural reasons why the US seems to be far the
01:08:58.960 most reluctant major country to introduce AI legislation? The EU, China and India have introduced
01:09:07.040 quite extensive legislation already. So why does the US seem to drag behind?
01:09:12.280 What's the one thing we're not talking about as a society that you think we should be?
01:09:19.040 The truth. I think I mentioned that a couple of times. I think the sign of our time, honestly,
01:09:26.480 is that there is a flood of information and a lot of distraction where most of us believe what we're
01:09:32.660 told, when in reality, none of what we're told is true. The biggest skill in today's world is your
01:09:38.920 ability to debate if what is your, if what you're being told is your truth, right? If someone tells
01:09:46.600 you that this party is right and that party is wrong, they're both wrong. They're both right.
01:09:51.980 They both have some topics that you should agree on and other topics that you should, it's the
01:09:57.260 granularity that matters. If someone tells you that this, you know, future is going to happen,
01:10:03.960 even what I told you, debate it, debate it, okay? If someone shows you a video on Instagram,
01:10:10.580 know that this is not entirely the truth. The biggest issue with our world today is we have
01:10:16.900 no ability, we've become dumb, right? We have no ability to vet what we go through first before we 0.94
01:10:26.400 make up our minds on it. We sometimes label people, you know, I know you're going on a tour with Jordan
01:10:31.660 Peterson, for example. People will either label him as intelligent and wise and, or, you know,
01:10:37.500 annoying and evil, right? It doesn't matter, by the way. Whatever you label him, it doesn't matter.
01:10:43.980 If he says something wise, it's wise. If he says something annoying, it's annoying. It's simple as
01:10:48.960 that, right? And I think that reality of how, of going beyond what's presented to us to actually find
01:10:56.680 out what the truth, what my truth, there is no, it's very difficult to claim that there is one truth,
01:11:02.560 okay? But to be able to go beyond that presentation to find out what my truth is, is the first step
01:11:09.220 before you react to it. I think this is the sign of our times. Thank you very much. Guys, head on over
01:11:14.480 to locals where we ask Mo your questions. Could AI ask questions of itself such that it will one day
01:11:21.980 become religious? Great question.