Making Sense - Sam Harris - May 02, 2022


#281 — Western Culture and Its Discontents


Episode Stats

Length

1 hour and 12 minutes

Words per Minute

170.21198

Word Count

12,272

Sentence Count

558

Misogynist Sentences

5

Hate Speech Sentences

11


Summary

Elon Musk has acquired a controlling stake in the social media giant, but what will that mean for the future of the site under his control? And what will it mean for free speech on the site? And what does this mean for our understanding of the First Amendment and the role of free speech as it pertains to social media platforms like Twitter and other social media sites like it. Is this a good or bad thing? And does it change the way we think about what's going on on social media as it relates to free speech and freedom of speech? And is it even possible that Elon will be able to improve the site in the long-term, or will it continue to erode the quality of the service it provides to its users and users' experience? And will it be enough to fix the problem of moderating the site, or is it just more of the same old censorship and hate speech we ve seen in the past? We'll talk about that and much more in this episode of The Making Sense Podcast. Please consider becoming a supporter of the podcast by becoming a patron of the show. Subscribe to the podcast and/or subscribing to our other podcast, Making Sense to get immediate access to all the latest episodes. We don't run ads on the podcast, and therefore, we don't need to pay for your ad-free listening experience! We make the podcast made possible entirely through the support of our subscribers only. by you, the listener! Thanks to our sponsorships, our listeners make the show better, better listening and better listening to our podcasting experience, and we improve our quality of listening to the making sense and making sense of our podcast. -Sam Harris and the podcast becomes more of your listening experience, more of you understand what we're listening to you, your brain gets better and less of you listening to us, and you become more of a place to listen to us listening to more of us making sense, more like you, and more of this podcasting more of what we all of us understand us, more understanding us, less of us, better of you, more listening, and less stuff like that, and so on and more understanding you, you get it, more profound and more like us, you'll get more of that, you're helping us all a better of it, right Thank you, Thanks for listening, friend us, thank you, friendlier, more thoughtful of you're making sense?


Transcript

00:00:00.000 Welcome to the Making Sense Podcast. This is Sam Harris. Just a note to say that if
00:00:12.160 you're hearing this, you are not currently on our subscriber feed and will only be hearing
00:00:16.300 the first part of this conversation. In order to access full episodes of the Making Sense
00:00:20.800 Podcast, you'll need to subscribe at samharris.org. There you'll find our private RSS feed to
00:00:26.400 add to your favorite podcatcher, along with other subscriber-only content. We don't run ads on the
00:00:31.560 podcast, and therefore it's made possible entirely through the support of our subscribers. So if you
00:00:36.340 enjoy what we're doing here, please consider becoming one. Okay. Well, a lot happening out
00:00:50.660 there in the world. I think a brief comment on the immense amount of attention and controversy
00:01:00.360 sparked by Elon Musk planning to buy Twitter. It seems like that is happening. It could still
00:01:06.920 not happen, but it seems like it's more likely than not at this point. I have been fairly astounded by
00:01:16.300 how much of what has been said about this on both sides seems to miss some obvious points. Again,
00:01:25.780 from both sides. On the left, there's been a fair amount of hysteria around a billionaire,
00:01:32.380 and one as outspoken and opinionated as Elon, buying Twitter and therefore controlling such an
00:01:40.960 important media property. Well, billionaires control so much of what's important that there's
00:01:47.900 nothing new there. And from the right, there has been a lot of celebratory nonsense about how much
00:01:57.360 is guaranteed to change under Elon's stewardship. If I was going to summarize my opinion here, I think
00:02:04.260 I'm agnostic as to whether or not Elon can actually do much to improve Twitter. There's some obvious
00:02:11.700 things he could and should do, and I trust will do, like cleaning up a lot of the bots and not doing some of
00:02:21.020 the very stupid things that Twitter has done in the service of its moderation policy in the past. The people on the
00:02:28.180 left that think that Twitter did not have a problem with heavy-handed moderation either weren't paying
00:02:34.900 attention or agreed with that heavy-handed moderation for ideological reasons, right? I mean, literally
00:02:42.600 someone got kicked off for life, I believe, for tweeting, men are not women, right? That was considered
00:02:49.300 hate speech in the context in which she tweeted it. And meanwhile, ISIS and the Chinese Communist Party,
00:02:57.720 I mean, all of these groups have accounts in good standing at that point, right? So that's crazy.
00:03:07.240 And insofar as Elon is going to insist upon a more transparent and ethical moderation policy,
00:03:16.980 that will be to the good. But in truth, if moderation were easy, someone would have figured it out by now.
00:03:26.420 And, you know, I'm not especially close to this problem technically and what algorithms can do to
00:03:33.360 solve it. But it just seems like there are always going to be apes in the loop, at least to adjudicate
00:03:41.540 someone being kicked off and reinstated. You need people at a certain point to process these claims of
00:03:49.520 who should be kicked off and who shouldn't. And what you have in front of you are an endless series of
00:03:56.640 judgment calls, some of which are trivially easy and some of which are really hard, right? And I don't
00:04:02.600 see how that problem ever goes away. So I don't see how you don't always have enormous numbers of
00:04:09.080 dissatisfied people in the wake of even the wisest moderation policy. Now, for the so-called free speech
00:04:17.020 absolutists who seem to not want much of a moderation policy and who are claiming that Twitter's or any other
00:04:25.180 platform's attempts at moderation in the past amounted to censorship, first of all, we already know what an
00:04:32.300 unmoderated or effectively unmoderated platform looks like. You go over to 4chan or 8chan and see what no
00:04:41.660 moderation gets you. Here's where any sane moderation policy parts ways with the First
00:04:48.700 Amendment. Everything happening over at 4chan and 8chan is protected by the First Amendment. I think those
00:04:56.160 platforms should exist. Right now, there are things on there that might be illegal, right? Child pornography
00:05:02.980 and any other video record of a crime that was perpetrated for the purpose of creating the video.
00:05:10.180 There are laws against all of that. People should go to prison for that stuff. Totally understood. But
00:05:16.220 that leaves immense scope for absolutely obnoxious and soul-destroying poison that can be spread on a
00:05:26.480 social network and which most of us want nothing to do with. And there are legitimately hard calls. Like,
00:05:32.640 for instance, I think every platform should have a no-doxing policy. And the people who have been
00:05:39.820 kicked off Twitter, people like Alex Jones and even Donald Trump, I think, should have been kicked
00:05:45.900 off. And they should have been kicked off largely for their knowingly marshalling their crazy followers
00:05:53.200 to dox and harass and effectively ruin the lives of identifiable people on the platform. Right? That is
00:06:02.880 what was happening. Jones and Trump knew that's what was happening whenever they targeted an individual
00:06:08.420 on social media. And it was absolutely despicable. But there is no bright line between malicious doxing
00:06:19.280 and necessary journalism. Right? As users sort of know it when you see it. Do I think that members of
00:06:28.000 ISIS should be doxed? Absolutely. Right? Show me some terrorist atrocity with people caught on cell phone
00:06:37.240 cameras or security cameras? Do I want those people identified? Do I want them caught by the cops? Of
00:06:44.460 course. But do I want somebody who has an opinion that is not shared widely by the woke mob doxed by that
00:06:53.940 mob and hunted as an apostate out in the real world because of what happened to them on Twitter? Of
00:07:00.520 course not. But again, this is a hard problem to solve. And there will be edge cases. And I just don't
00:07:06.860 see how that problem goes away by taking Twitter private or by cleaning up all the bots or by implementing
00:07:15.100 an appropriate algorithm. There are still going to be people at the end of the day trying to figure out
00:07:21.260 where the edge cases are and what to do about them. So I think the right and the left have much of this
00:07:29.040 wrong. I think appeals to the First Amendment are generally misleading. I think we want platforms that have
00:07:38.140 coherent moderation policies that prevent them from becoming like 4chan and 8chan. And I certainly wish
00:07:47.380 Elon the best of luck in developing such a policy, implementing it, and in making Twitter better than
00:07:55.060 it is. I think Elon's claim that Twitter is the town square and that it's absolutely crucial to make
00:08:03.440 it much more in line with the First Amendment is an understandable but I think ultimately dubious one.
00:08:11.880 Twitter isn't the town square. There are many successful influential people who are not on Twitter.
00:08:18.600 The problem is that most people in tech, most people in journalism, most academics, and certainly
00:08:26.220 Elon among them, are addicted to Twitter. And I think it's pretty clear, or it should be, that in almost
00:08:34.540 every case that addiction is counterproductive. It's not to say that Twitter isn't useful. I'm still on
00:08:41.980 it. I still find it a valuable source of news and recommendations. Occasionally it's a great spot to
00:08:49.620 connect with someone who I wouldn't otherwise connect with. But I have pulled back a lot because I witnessed
00:08:56.360 a fair amount of the dysfunction of over-engaging with the platform in my own life. And I certainly
00:09:04.400 see that dysfunction well advertised in the lives of others. So there are many reasons not to be on
00:09:12.440 Twitter or not to be on it much. And there are many people who are thriving who are not on it. So it's not
00:09:20.320 the town square. You have not lost your personhood if for some reason you get deplatformed from
00:09:27.520 Twitter. So I think the analogy to the town square is a false one. And I think the notion that any legal
00:09:37.100 speech must be tolerated on the platform is going to lead to a truly awful place to be. And then people
00:09:45.880 will be free to leave and start a new platform. Anyway, this topic comes up however briefly in my
00:09:52.280 conversation today. And in the end there'll be much more to say about it. But I think creating a social
00:09:58.820 media platform that actually works, that becomes a place where smart, well-intentioned people are wise
00:10:05.220 to spend their time, I think that is a really difficult problem to solve. And I certainly hope someone
00:10:12.140 solves it. Anyway, those are my two cents. And now for today's podcast. Today I'm speaking with
00:10:19.340 Douglas Murray. Douglas is a friend who's been on the podcast before. He's the author of several books,
00:10:25.800 most recently The War on the West, which we talk about in depth. His previous books were The Strange
00:10:33.160 Death of Europe and The Madness of Crowds. He's also an associate editor for The Spectator. He writes for
00:10:39.260 several other publications. He's immensely prolific. And as you'll hear, he is always great to talk to.
00:10:45.460 We get deep into his book, The War on the West. Before we do, I go fishing for some areas where we
00:10:51.500 might disagree. And actually this question of moderation on social media platforms is one of
00:10:57.880 those areas. We talk about the problem of hyper-partisanship on the left and the right,
00:11:02.600 and the primacy of culture. We talk about the problem with Trump and use the Hunter Biden laptop
00:11:12.480 controversy as a lens there. We talk about the deplatforming of Trump and Alex Jones specifically.
00:11:21.100 And then we get into the topic of his book, Proper. We talk about the new religion of anti-racism,
00:11:26.400 the problem of inequality, the 1619 project, the history of slavery, moral panics, the strange case
00:11:36.660 of Michel Foucault, and other topics. Anyway, it's always great to talk to Douglas. I hope you enjoy
00:11:43.600 this conversation as much as I did. And I bring you Douglas Murray.
00:11:47.600 I am here with Douglas Murray. Douglas, thanks for joining me again.
00:11:58.060 It's a huge pleasure to be with you, Sam.
00:12:00.180 So we have a lot to talk about. First, I should apologize to our listeners for canceling the
00:12:07.840 live Zoom event, which had been scheduled for this podcast. But as I told you offline,
00:12:13.320 the house across the street from me was being demolished. And rather than have sounds of the
00:12:20.000 apocalypse intrude upon our recording, I had to forsake my Zoom recording space in order to go
00:12:27.860 just for the pure audio experience. So there we are.
00:12:30.960 I'm in New York, Sam. So it's permanent Armageddon noises in the background here. So we could have
00:12:36.560 canceled each other out.
00:12:37.700 Yeah, yeah. Okay, so you have a new book, which we will definitely talk about. That book is The War
00:12:42.560 on the West. And it is a, in case it's not obvious from the title, it is a passionate
00:12:48.800 defense of Western culture of a sort that only you could muster. And it's a fantastic read. It's
00:12:55.400 actually a doubly fantastic listen. I read some of it and listened to the rest of it. And as I did
00:13:03.360 with your last book, I can't remember if you, I don't think I heard the audio for The Strange Death
00:13:09.300 of Europe.
00:13:09.960 That wasn't done by me, sadly. Yeah. Only this one and Madness of Crowds.
00:13:14.080 Yeah. So both of those, Madness of Crowds and your new one, The War on the West, you read. And
00:13:20.060 it's one of the great pleasures of having ears and a brain to which they're connected is to hear you,
00:13:27.540 reading your own stuff is great, but to hear you reading quotations from people you deem to be
00:13:35.120 either insane or sinister. And giving it the topspin of derision is just amazing. So I recommend that
00:13:46.240 people listen to the audio. That's very kind of you to say so. I actually, I enormously enjoy doing
00:13:50.720 my own audio books, partly because I find it incredibly funny. And with Madness of Crowds,
00:13:56.540 as with The War on the West, I had to apologize repeatedly to the sound engineers and explain to
00:14:00.720 them. I wasn't laughing at my own jokes. I was laughing at the things I quote, because so often
00:14:06.760 they're ridiculous on the page, but they're even more ridiculous when you say them out loud.
00:14:12.180 Yeah. And just some of your own writing also gets the benefit of your reading. And there are lines
00:14:18.480 that really are laugh out loud funny, which I'm not sure everyone would discover on the page quite as
00:14:24.520 readily as when you're reading them.
00:14:26.800 That's very kind. There was one in Madness of Crowds, I remember, that was much better on
00:14:30.620 Audible, which was I quoted somebody referring to something as being literally like Adolf Hitler's
00:14:35.500 Mein Kampf. And I say, not just any old Mein Kampf, but Adolf Hitler's Mein Kampf.
00:14:41.920 It's much better in audio than on the page.
00:14:44.480 Yeah. So you've had a tremendous amount of fun at the expense of the left, and we will get into that.
00:14:50.040 But I want, I mean, one thing I noticed when I announced this conversation, when I announced the
00:14:54.840 Zoom event, I got some of your hate mail on social media, and some of my own, perhaps. And I think
00:15:01.300 many people were expecting that any conversation between the two of us about the derangement of the
00:15:08.500 left would just be an exercise in confirmation bias, right? We're basically going to agree about
00:15:13.820 everything.
00:15:14.740 Yeah, there's something in that.
00:15:15.540 Yeah. And so I think it would be good for us to remain alert to any areas where we actually
00:15:21.900 might disagree. I mean, I think we will fully agree, perhaps with tiny little shadings of gray
00:15:29.740 somewhere when our attention is directed to the left and to the topic of your book. But I think if we
00:15:36.780 talk about the right at all, that we may find some differences of opinion.
00:15:41.280 I mean, one area of difference for me, maybe we can just start here. Because I do see, again, we totally
00:15:50.020 agree about the central problem in its leftist form. But I do see that a similar thing happening on the
00:15:58.500 right, and you don't tend to focus on it. And I guess my, I do have a general question as to why, but let me just
00:16:04.620 spell it out for you. I think the generic problem that we both see is that there's now a concern
00:16:11.140 with identity that seems to supersede any honest engagement with ethics or facts or even a concern
00:16:21.520 about whether one's own beliefs are internally consistent, right? So there's just immense double
00:16:26.680 standards and instances of hypocrisy and just shoddy thinking happening under the aegis of identity
00:16:34.620 politics. But I'm finding this both on the right and the left. And I mean, so on the, you know, there's
00:16:41.600 this obsession with group difference and victimhood. There's the same willingness to destroy institutions
00:16:47.620 without any thought as to what could replace them. There's the right has just grown demented by
00:16:53.760 conspiracy theories and a cult of personality, you know, under Trump. And so on the left, you'll see
00:16:59.860 people deny that, you know, there's anything strange about being told that all white people
00:17:04.700 inherit the original sin of racism, or that there's anything strange about a new book titled
00:17:11.240 Anti-Racist Baby, right? I mean, this is where we'll get deep into this when we hit your book. But
00:17:17.620 on the right, we see people denying that there's anything wrong with Trump or the January 6th attack
00:17:23.800 on the Capitol, or the big lie about the 2020 election. So I guess my question for you in search
00:17:29.960 of possible disagreement here is, is why focus exclusively on the left? Well, the first thing is
00:17:36.920 I don't. And I'm always, I'm sure like you, Sam, I've become aware of quite easily of who doesn't
00:17:45.920 read me. And, you know, actually an interviewer said to me the other day, you know, what, what do
00:17:51.320 you, what do you think about what people think about you sort of thing? And I said, I just, I don't
00:17:55.180 really know what they think. I don't spend that much time trying to absorb it. But I know when people
00:18:00.100 don't read me, and I know that one of the signs is when people say you only talk about X, when
00:18:05.620 actually I write about a pretty wide range of subjects. I write three to four national newspaper
00:18:11.640 columns a week. And I wouldn't be employable if I wrote about only one issue.
00:18:16.900 Well, let me just to claim not to be guilty of not reading you because as I, because I
00:18:22.820 do, perhaps only is too strong. But I mean, so you and I both have several friends and
00:18:29.140 colleagues. And in certain cases, it might be former friends and colleagues who have been
00:18:34.200 fighting from the same trench as the two of us aimed at the left, but they've focused entirely
00:18:40.500 on the left, right? I mean, and some of them appear to have lost their minds, or at least
00:18:45.480 lost certain principles of intellectual honesty. And I won't name names. I know, I know, you
00:18:49.860 know who I'm talking about. And, you know, I certainly don't put you in that category.
00:18:56.180 But there's no question that you, this book you've just written, is entirely focused on the
00:19:03.360 leftist assault on Western culture.
00:19:06.340 Yes, well, because I see the left as providing the assault that I'm trying to push back against,
00:19:14.100 identify, and I think inoculate us against. But I mean, I'm by no means silent on problems
00:19:21.500 on the right. Obviously, I'm more politically aligned with the right than you are. And, you know,
00:19:26.440 I don't particularly mind that, albeit the right that I knew from the UK is rather different from
00:19:31.540 some of the American right. But that's not to dodge matters. It's simply to say, as I say,
00:19:36.020 about people not reading me, I mean, anyone who reads, you know, what I write will know
00:19:41.400 that I've consistently critiqued my own side. I mean, for instance, and let me just rattle off a few
00:19:47.420 that come to the top of my mind. Immediately, January the 6th happened, I wrote in the main
00:19:52.940 conservative newspaper in the UK, this sits solely at the feet of Donald Trump. He led his troops to the
00:19:58.980 top of the hill. And what did he expect them to do? I make no apology for that, got plenty of
00:20:05.240 criticism for it for people, but I still will not regard and do not regard the attack on the US
00:20:10.880 capital as being nothing. And have consistently said that among other things, you know, whatever
00:20:18.000 happens with Donald Trump himself, you cannot claim that what people around him were saying was not
00:20:22.940 essentially up to and past the point of what we call incitement. That seems perfectly clear.
00:20:27.960 And I've, and I've, and I've written about that repeatedly. And let me give you two other quick
00:20:31.580 examples. There are on the American right things, and I've been in America for a year now, there are
00:20:36.400 things which do not exist anything like the proximity to the political center on the American
00:20:43.100 right than what exists in Britain. We'll give you a couple of examples. An obvious one is conspiracy
00:20:49.800 theories. Another one is like very unpleasant forms of prejudice, which, again, would totally knock you
00:20:58.120 out of the race in the UK. Just in fact, I spent New Year's Day this year, not taking a break, because
00:21:06.220 a, I won't name him, but a very ugly, unpleasant right winger in the US had been had spent his new
00:21:13.460 year tweeting about people, one particular person who he described as having a Rothschild physiognomy.
00:21:19.980 And I spoke to Barry Weiss and said, this is, this is where the right goes wrong. This guy is
00:21:27.160 actually affiliated with some conservative institutions in the US. It seems to me totally
00:21:31.660 intolerable that flagrant antisemitism should be anywhere near the center of the American right,
00:21:38.040 and immediately criticized him for this. And got, I have to say, I mean, absolutely no reward in
00:21:44.400 return, only a heap of bile from right-wingers who thought that he was either ignorant and didn't know
00:21:50.180 what he was saying, or that there was nothing wrong with talking about people having a Rothschild
00:21:54.200 physiognomy. And thirdly, I'd say just off the top of my head, at the moment that Russia invaded
00:21:59.940 Ukraine, I saw that a part of the right in America was going very wrong indeed, as was a part of the
00:22:05.700 right in Europe. And I immediately used my column in The Spectator, which is the oldest right of
00:22:10.100 center magazine, the oldest weekly magazine in the English-speaking world. I used my weekly column
00:22:14.240 there to talk about the right that had gone wrong on Russia, how it had been misled, how it was lying,
00:22:21.180 how it was providing counterfactuals, counterinformation, how it was pumping out Russian
00:22:25.360 disinformation, how it had fallen for Vladimir Putin and been taken for a ride. Again, I say this,
00:22:32.300 not just because anyone can go and search this stuff, but because I don't think I ever make,
00:22:38.540 ever have any problem with saying what I think about people who are identified as being on my
00:22:43.320 own side. And there's a reason for that. It's not tactical. It's because I don't want to be a
00:22:47.540 million miles near these people. I wouldn't want to be near these people. So when people say,
00:22:52.960 and they did with that person I identified who was obviously a nasty little anti-Semite,
00:22:56.340 when they said, oh, you've no idea how many people are going to turn on you about this,
00:23:00.380 I don't care. Why would I care? Why would I care? Why would I want to be aligned with people who
00:23:05.320 thought that Vladimir Putin was the savior of Christendom and a devout, honest Christian who
00:23:11.960 must sort of provide the bulwark to the madness of left-wing liberalism? Of course not. I don't
00:23:18.880 want to be anywhere near these people. And as for the Trump point, by the way, I mean, sorry,
00:23:22.840 it sounds like only because of the nature of the question. I don't want to sound too self-defensive,
00:23:27.580 but no less a platform than the National Conservatism Conference in Orlando, Florida
00:23:36.000 last year. I was on a stage with several people you'd know, and the question of Trump came up.
00:23:41.300 I criticized Trump in front of an audience that was mainly supportive of him. And I said,
00:23:45.040 there is something absolutely unsustainable about the fact that in front of an audience like this,
00:23:50.460 I mean, various people like Ted Cruz had spoken as well. And I said, one of the only dissenting
00:23:55.080 notes of the conference, and I said, among other things, it is totally unsustainable that you have
00:23:59.220 this situation where at a conference like this, somebody asks a question about Donald Trump,
00:24:03.080 and everyone on the panel pretends to know less than they know about him. They pretend not to know
00:24:09.180 that he's got a really horrible character, for instance, and pretend that merely his ability to
00:24:15.860 win is what we like, and therefore we'll park everything else. We'll pretend that January the 6th
00:24:21.960 didn't happen, and that it's just the libertards going mad. I said that in front of that audience.
00:24:27.720 Again, I'm not searching for popularity, but I would not want to be on the stage, which included
00:24:34.000 people who simply uncritically praise Donald Trump and join in with it. Why would I want to be anywhere
00:24:37.740 near that?
00:24:38.820 Yeah, yeah. Well, I'm very glad I gave you an opportunity to get that off your chest, because
00:24:45.080 you're often lumped in with the people who do not make those points, which I think are absolutely
00:24:52.580 just necessary concessions to political sanity, and it's a problem. So I just, you know.
00:24:59.180 By the way, sorry, one other whilst I'm at it, which is I said immediately after, I covered the US
00:25:03.400 election for a number of newspapers, and I traveled around about 10 states in the days before the 2020
00:25:07.780 election, in the weeks before the 2020 election. Went everywhere from, you know, sort of across the
00:25:13.460 country, covered a Trump rally in Florida. And the minute that the results came out, the right started
00:25:21.600 to lie about them. I said then, and again in The Spectator, I said, this is going to be a real
00:25:27.100 problem for us, because these people are going to waste our time for years. They are going to waste
00:25:31.240 our time with this conspiracy about this election, and they don't realize that they're not just wasting
00:25:35.260 our time, they're wasting their own, because they will do the crucial mistake that always happens
00:25:39.540 when people fall into this, as some Democrats did after 2016. They will fall into the mistake of
00:25:44.420 thinking that they won, and as a result, they will not do the necessary self-searching that you need
00:25:50.400 to do when you've actually lost an election and work out why you've lost. You know, so it's to their
00:25:56.160 own fault as well. It both demeans their opponents and it demeans themselves. I said that straight away.
00:26:02.480 Yeah, so let's, again, before we dive into the left side of the chaos.
00:26:09.040 No, we should focus on the right, for sure.
00:26:10.780 Yeah, no, I just want to see if I can find the generic essence of our problem first. I mean,
00:26:17.280 I think we both are worried about what appears to be a derangement of our culture, and culture is not
00:26:25.720 this expendable thing. Culture really is the operating system for humanity at this point. I mean,
00:26:32.720 we have, you know, insofar as we surmount, you know, mere nature, you know, red in tooth and claw,
00:26:40.360 we arrive fully in culture, and it's just the basis for every epistemic and emotional and ethical
00:26:51.100 engagement with our shared social reality. And, you know, politics is a strand of that,
00:26:57.960 but there's much more to it than politics. And what we're seeing now is an environment
00:27:04.040 wherein misinformation and moral panics and social contagion are getting made immensely worse
00:27:13.160 by social media and current trends of, you know, loss of trust in institutions and just other forms
00:27:21.380 of fragmentation of society. And again, this is, you know, whatever we're going to say about the
00:27:26.220 left, as crazy as it is, and as easily seen to be in your recent book, you know, on the right,
00:27:33.140 we have QAnon and the other odious exports from Trumpistan. I mean, just the amazing thing on the
00:27:41.400 right. I mean, the moment I can't forget, and really it was the point of no return for us,
00:27:47.560 I thought, politically, was when we had a sitting president repeatedly not commit to a peaceful
00:27:54.700 transfer of power in the event that he lost an election, right? He was given multiple opportunities
00:27:59.160 to do this, and he refused. And the Republican Party was okay with that. And I mean, this precedes
00:28:06.660 January 6th, right? And all the knock-on effects of that. But the, just in the run-up to the election,
00:28:15.240 when we have a president who won't commit to the, arguably, the most important norm politically in
00:28:22.760 our system, you know, upon which everything else that matters is anchored politically, and that the
00:28:29.720 Republican Party just swallowed that without comment, it just, it seemed to me to be, it was a sign that
00:28:36.980 we actually could lose our democracy in the hands of this buffoon. And I mean, I know you were, you
00:28:44.240 objected to that at the time as well, but it's so many people who will delight in the contents of your
00:28:51.960 book, and who want to hear everything we have to say about the craziness on the left, just didn't
00:28:58.040 care about that, you know? And I'm just, yeah. I say several things. I mean, one is, several years
00:29:05.660 ago, our mutual friend Jordan Peterson and I did a discussion on video about where the left goes
00:29:12.160 wrong, which was a discussion which I thought was really very interesting, very generative. And because
00:29:17.840 of this idea that Jordan kicked off, which was, you know, we sort of have a clear idea of where the
00:29:22.240 right goes wrong politically and playing games of racial superiority, for instance, authoritarianism
00:29:26.940 and much more, we don't have an absolutely clear blueprint, by contrast, of where the left goes
00:29:31.200 wrong. And I think that's a totally accurate statement, and I think that it is a big problem.
00:29:35.940 Where is it in collectivism? Where is it within the social justice movement that the left starts,
00:29:41.960 you know, how do you end up with the gulag? And we had a very interesting discussion about this,
00:29:45.800 and one of the things, looking back on it, and I've said this since, including to Jordan,
00:29:49.320 I said this to him, indeed, when we did a discussion, a podcast, on his podcast a couple of days
00:29:53.680 after January the 6th, I said, when we did that discussion several years ago, we did it in the
00:29:58.700 belief that it was clear where the right went wrong and that the right was therefore unlikely to go
00:30:02.920 wrong. And we can no longer make that assumption. We're having to revisit those statements, those basic
00:30:11.880 underpinnings that we thought everybody had. We do actually have to revisit them. And we did,
00:30:17.660 by the way. And again, I don't say this by any means to search for praise, but neither Jordan nor
00:30:24.720 I got any particular love from followers for this. But I said to him, this is a very important thing,
00:30:32.860 that two figures who are more identified as being on the right than the left, certainly,
00:30:36.060 make it plain that this is where the right goes wrong. And the discussion we had included us,
00:30:43.940 I thought, rather helpfully, helping each other to the following realizations. I think the best
00:30:48.640 way I could sum it up is, I said, if you went back five years from where we were then to, say,
00:30:55.420 like 2015, and you said there was going to be a time in 2020 in American politics where a significant
00:31:01.980 amount of the right is going to believe the following, that no media is telling the truth,
00:31:07.520 that no politicians tell the truth, that the law courts are all totally corrupted,
00:31:12.920 that every one of the intelligence agencies is totally corrupted, that the ballot is totally
00:31:18.640 corrupted to the extent that an election is going to be stolen. But you have one great virtue on your
00:31:24.660 side. There is one virtuous man in the republic. And you know who that man is? It's the dude off The
00:31:29.300 Apprentice. Yes, Donald J. Trump. Now, if you just said that to anyone in 2015, they just said,
00:31:36.240 oh, sorry, and also the vice president, Mike Pence, he's also completely corrupted and not a
00:31:43.340 conservative. If you just said that to somebody in 2015, they just said that you're a maniac.
00:31:49.440 How is that going to happen? How am I going to end up in a position where the only man
00:31:53.700 who I'm going to trust and possibly turn up to the Capitol and risk my life for and risk other
00:32:01.440 people's lives for is Donald J. Trump of all the people? Yes. That's a point I've almost made before
00:32:12.340 in the following form. Back in 2015, I would have said that there was literally not a single
00:32:18.660 Fortune 500 company in America that would have ever had the thought, the situation is really
00:32:26.940 grim for us. What we need is a complete rebooting of our organization. We need to bring in a new CEO.
00:32:32.980 We need to find them. And we have found the most competent, most inspiring person for the job. And
00:32:39.000 that man is Donald Trump. Yes. That would have been... We've done a headhunting exercise,
00:32:43.060 guys. And we've come back with a song. A man of high integrity. That's what we need.
00:32:49.460 Never knowingly told none truth. Yes. So this is definitely a problem. And it's a problem of,
00:32:56.220 I would say, particularly of the American right. And the problem is, obviously, that there is
00:33:02.260 something that Trump taps into, which they fear that nobody else can. And I don't know whether they're
00:33:07.740 right or not. I have no electoral crystal ball. What I do know also is that there's one other
00:33:12.520 interesting, which is worth highlighting, which is that for some years, I think, in the cultural realm
00:33:19.680 and others, there was a perception on the conservative side that conservatives had played
00:33:25.420 too nice. That basically what happened was the left advanced incrementally and sometimes actually
00:33:32.140 in bounds. That it enjoyed rubbing the right's noses in its defeats. That conservatives were too
00:33:41.100 gentlemanly to ever do anything other than slightly slow down that progress of the left or to fight
00:33:47.920 another... to fight the next battle they were going to lose. And that this was the sort of trajectory of
00:33:54.160 politics. Now, again, I'm not saying that I agree with it or disagree with it or whether it's true or
00:33:57.700 not. That was a perception on the right. And that the point that Donald Trump came along, as far as it
00:34:04.400 seems to me... And I, by the way, I tried to persuade... I tried for most of his presidency not to write
00:34:08.300 about him because I thought that since everyone on the planet had a thought, it wasn't particularly
00:34:12.720 worth my while adding to the melee. And I thought the same with Brexit, incidentally, after the Brexit
00:34:18.020 vote happened. Not that they're connected. But I just... I tend not to... if everybody on the planet
00:34:22.600 is writing about the same thing, I tend not to want to join in the cacophony. And also because it
00:34:27.600 seemed to me there was relatively little to add. But just to return to this point, that, you know,
00:34:33.800 there was this perception on the right, particularly in America. And they did something
00:34:40.380 which I think is both understandable and reprehensible, which was to essentially choose
00:34:46.400 as a tool of fight, a weapon of fighting, the weapon that they believed would most upset their
00:34:51.640 opponents. It effectively goes to that instinct to hurt your enemy, not to just win, to kick them
00:34:59.420 in the balls. And Donald Trump was that dirty fighter. And the right suddenly, or a section
00:35:05.580 of the right, suddenly got excited about that. They got excited about the fact there was somebody
00:35:09.800 who took the fight to the enemy, who, you know, literally calling them the enemy, who would
00:35:15.100 derange the other side. You know, all that sort of liberal tears sort of thing.
00:35:20.000 It was rejoicing in it, saying, basically, we're so fed up because we've spent years being
00:35:25.300 bullied. And so we're going to have some fun being the bully. And that is, as far as I can
00:35:31.180 see, the dynamic that led to Donald Trump. And because the Republicans don't know whether
00:35:36.620 they can tap into that feeling of resentment without his aid, they're sort of sticking around
00:35:42.660 him. That's why you have this ludicrous dance that's going on at the moment when no one will
00:35:46.440 declare.
00:35:46.840 Yeah. Yeah. Well, there's one, again, in the service of looking for some place where
00:35:54.380 that we might discern some daylight between us, I think there's going to be very little.
00:35:58.960 But there's one instance that I am genuinely undecided about. I mean, in the rubble of our
00:36:07.080 information space, one thing stands out to me recently. I don't know if you've written
00:36:11.340 about it, or I think I've heard you comment about it briefly on a podcast. But the Hunter
00:36:16.600 Biden laptop scandal, right? That is a, I genuinely don't know what I think should have happened
00:36:25.160 there. Because let's just summarize the state of our knowledge currently is that it was treated
00:36:31.100 like a product of Russian disinformation at the time. There were a bunch of former intelligence
00:36:37.740 chiefs signed a letter saying this is, you know, classic fake news out of Russia. And it
00:36:44.120 was treated like, you know, pornography, journalistically, and suppressed by social media, Twitter, I
00:36:52.680 think, delinked New York Post's account. I think you couldn't forward the story any longer.
00:36:58.680 And all of this was done, you know, immediately before the election, this was the some kind of
00:37:05.240 October surprise. And at the time, I didn't know what to think about it. I, you know, I didn't know
00:37:10.740 any more than anyone else knew who was being denied access to the information. Except I did know one
00:37:18.020 thing, which is, I didn't care if Hunter Biden had severed heads in his basement, right? I mean,
00:37:25.420 like, there was literally nothing you could have told me about Hunter Biden, that would have been
00:37:31.340 relevant to me, when the goal was to keep Trump out of office, at that point, right? Because it was
00:37:37.600 just, I did view Trump given, you know, the aforesaid non commitment to the most important principle of
00:37:44.820 the survival of our democracy, I viewed him as an existential threat. And given what had happened
00:37:50.400 in 2016, with Comey, reopening, you know, the email, right, the case into Hillary Clinton's emails,
00:37:58.360 we know that, you know, though her failure to win the presidency was certainly overdetermined,
00:38:02.960 we know that in the in the last 11 days of the campaign, that was the coup de grace, right? With,
00:38:09.780 you know, and this could have proved the same for the election of Biden, because it was it was going
00:38:16.540 to be this bright, shiny object that was going to captivate everyone and suck up all the oxygen.
00:38:21.560 So I honestly don't know what I think should have happened there, because I think you and I will
00:38:26.860 agree that there really is a problem when you have the are preeminent sources of journalism,
00:38:32.680 pretending that a significant story is, in fact, a non story. You know, I guess I should close the
00:38:40.420 loop on this is recently been admitted by the New York Times, in an article to which they gave
00:38:45.640 very little oxygen that, oh, sorry, guys, this really was a story. And it was legitimate. And
00:38:52.460 there are all kinds of heinous things on that laptop. And who knows, to what degree it it suggests
00:38:59.200 the corruption of, of Joe Biden and the Biden family in their engagement overseas. So I don't
00:39:06.080 know, I don't know how you feel about that. But I, I don't know what the counterfactual is what what
00:39:11.440 might have been done differently, that would would have been within bounds ethically,
00:39:16.040 journalistically. But it's just, I just, I don't know what I would change about the past,
00:39:21.220 with respect to that story, given the outcome.
00:39:24.580 I should declare an interest. I do write a weekly column for the New York Post, which is
00:39:28.940 the paper that broke the story, of course. I wasn't actually writing for the Post at the time,
00:39:33.800 more, apart from occasionally. And now I'm a regular. So I just add that as a, as it were,
00:39:39.340 just in case anyone thinks there's a conflict of interest. But I had no, no involvement in the
00:39:43.600 Hunter Biden laptop story, but I know the people who were involved in it. I think that it was a
00:39:49.280 catastrophic mistake to silence the Post, America's oldest newspaper, at that moment,
00:39:55.140 I thought it was a decision by a few big tech companies, who were basically helping Biden out
00:40:02.580 to win the election. The contents of the laptop, there's a good book by a colleague of mine,
00:40:07.640 the New York Post, Miranda Devine, who did a lot of the work on the story,
00:40:11.680 one of the people who had access to everything on the laptop. There's a very good book about it now
00:40:15.440 called Laptop from Hell, which, which, which if you, if you read, or even you read excerpts from,
00:40:19.980 you'll see that the problem is, I mean, I should stress, I'm not that prurient person. And I
00:40:25.280 actually have no, I mean, nobody's not prurient, but I have genuinely have no interest in the rather
00:40:30.220 sad private life of Hunter Biden. And I would have thought that a lot of the story would have
00:40:35.960 got caught up with that was people looking at dick pics and, and falling asleep with smack beside him
00:40:42.740 and this sort of thing. I have no interest in that. And I don't think it would have made any serious
00:40:46.740 change to the election. That wasn't the real story. The story was, as you mentioned, the fact
00:40:53.340 that Hunter Biden had been making money among other pieces, places in Ukraine, to the tunes of
00:40:58.740 hundreds of thousands of dollars a year, to sit on the board of an energy company in a discipline he
00:41:03.400 doesn't have in a country he doesn't know, in a job he wasn't doing. Now, does that matter? You might
00:41:09.100 say no. What matters has come across in some of the emails, which now not only the New York Times,
00:41:13.960 but the Washington Post has said, okay, the emails are true. And by the way, they could have done all of
00:41:18.280 this back then. It wasn't hard to, you could have called up anyone who was on the receiving end of
00:41:23.160 any of the emails, of the many emails that are on the laptop and say, is this actually an email from,
00:41:28.240 from Hunter Biden to you? And they could have confirmed or denied. It would not have been a
00:41:32.920 hard story to have chased up and followed up as the Post did then. But none of the rest of the media
00:41:37.840 came in behind. And the things that are on about the business thing should concern people. I mean,
00:41:43.340 I think the top of American politics is more corrupt than almost any other civilized nation,
00:41:49.280 it has to be said. There is something outrageous about the amount of money that can be accrued at
00:41:54.440 the top of American politics, both during and after office. And that is not exclusive to any one party.
00:42:01.200 I think that it is, I mean, whenever there's a financial scandal in the UK, by comparison,
00:42:05.640 it is laughable, you know.
00:42:07.800 $15 change stands.
00:42:09.500 Yeah, exactly. Somebody was, there was a backbench Labour MP at the time who
00:42:13.740 went into incredible trouble because she expensed a whirlpool bath that cost 800 pounds. I was in
00:42:19.420 America, this would be absolutely nothing, you know, compared to Nancy Pelosi's share deals.
00:42:24.840 But the point is, is that, is that, is that the interesting thing in the laptop was Hunter saying
00:42:30.740 to his daughter, for instance, whatever I do to you in your life, know that, that I will never do what
00:42:36.700 my father did to me and demand half of all the money I earn. Now, that is a very interesting story
00:42:42.520 of true. If it's true that there is, that Hunter Biden makes money and the father hives part of it
00:42:48.200 off, and we know that the uncle takes, look, the problem is, is that nobody on the left, as far as
00:42:53.600 I can see, particularly wants to engage in this. Why? Because they'll say, but Trump. They don't like
00:42:58.620 it, they wish it away, and they'll say, but Trump. Well, that is exactly what the right does with
00:43:03.440 some of the Trump stuff. They say, but Biden, but Democrats, but Hillary. And so they should
00:43:10.200 have published. I don't think the private prurient stuff would have made any difference. But I think
00:43:14.900 that a realization that the top level of American politics is, is wildly corrupt, where there is,
00:43:23.220 that family members of people, again, in both parties, become rich when their relatives enter
00:43:29.660 the White House, Congress, or what's more. I think that is something that's worth confronting.
00:43:34.760 Would it have changed the results of the election? I don't know. Nobody does know. But the New York
00:43:39.320 Post was completely right to run, because this was a hell of a story. And the rest of the media were
00:43:45.440 woefully gave themselves away by not reporting. And the media companies revealed what was revealed
00:43:52.840 after the election, which was, anyone could tell, which was a lot of the tech bosses and others
00:43:57.220 were so desperate to make sure that Trump didn't win the election, that they were willing to suppress
00:44:01.880 news that was negative about Biden. I think that's a scandal. I think it is part of the thing that
00:44:08.420 leads to this ever-increasing distrust in every single entity of power and information.
00:44:13.920 Yeah, no, I agree. I just think that Trump, given his, I would argue, treasonous non-commitment
00:44:23.520 to the most basic principles of our democracy, he's a singular problem that had to be solved at
00:44:31.680 that moment. Well, I would have said that American democracy, let's take the idea of Trump being a
00:44:38.680 kind of stress test of the American Republic. The American Republic survived him. Now, you might
00:44:44.920 think it was closer- By the skin of our teeth.
00:44:46.500 I do. By the skin of our teeth. But it survived. Court survived. The democracy survived.
00:44:52.780 But it was down to a handful of people who would just not accede to his demands. Had Mike Pence
00:44:59.020 done as instructed, had a few Republican election officials done as instructed, we would have had an
00:45:07.700 absolute constitutional crisis, the resolution to which was just non-obvious.
00:45:12.000 Absolutely. But they did stand up. Lindsey Graham did say on the floor of the House on the 6th of
00:45:20.000 January, I have asked repeatedly for evidence of this fraud in the election, and he doesn't provide
00:45:27.260 it to me. So I agree that too many people went along with it. There were mad theories going around.
00:45:34.740 Almost none of it has stood up since. But let me just return to this issue of the laptop,
00:45:41.940 because it's important in terms of this issue of trust in American politics, which disturbs the
00:45:47.300 hell out of me. The problem with the Post story was not just the suppression of the story, but what
00:45:52.060 you described, Sam, the joint letter by intelligence chiefs saying this is classic Russian disinformation.
00:45:58.100 Here's the problem. In my view, every single person who signed that letter should lose their pension,
00:46:02.780 should be cashiered, should be
00:46:04.740 should be disgraced. Why? Because these were people involved in the NSA, the FBI, the CIA,
00:46:11.320 who became political actors in order to support the suppression of a newspaper breaking a story
00:46:17.140 that enabled Joe Biden to be elected president. It was a wildly political intervention.
00:46:23.200 Except, Douglas, we know that there was massive Russian meddling into every aspect of
00:46:31.620 the conversation. I mean, on social media, with the hacking of the DNC, there was a continuous assault
00:46:40.020 upon our democracy with a kind of information warfare campaign from Russia. So it was certainly
00:46:46.080 plausible to think that this might have been a Russian compromise of some kind. You know,
00:46:51.760 I mean, again, it's not a crazy allegation.
00:46:54.660 On the contrary, I think it is. I think that
00:46:57.340 both sides have wound
00:47:00.240 themselves up in American politics in recent years
00:47:03.040 and
00:47:03.900 politicized institutions that should never have been politicized
00:47:07.540 and have
00:47:08.560 overemphasized this allegation that the democracy has been hacked.
00:47:12.220 The Democrats did it immediately after the 2016
00:47:14.460 election. Again, it's not a popular
00:47:16.740 point to make to some Democrat listeners. But what Donald Trump did in 2020 was unforgivable.
00:47:23.400 But part of his ability to get away with it, I believe, came from the fact that there were so
00:47:28.820 many Democrats who were not willing to believe that he had been legitimately elected in 2016 either.
00:47:33.580 In other words, what I'm saying is, you might say it's a 1% injection of falsehood or a 5% injection.
00:47:39.300 But the point is, is that it was already up for grabs in America that the ballot was not secure,
00:47:45.320 that the vote was not secure, that you could be hacked by Russia and actually it didn't matter.
00:47:50.840 Now, here's the thing. You're now at this stage, and I wrote this some time ago, and if you translate this into the British context,
00:47:57.260 in Britain,
00:47:58.780 if you had a situation where
00:48:00.800 conservatives, never mind, we can put the left to the side for the second, but I can do the same exercise on the left,
00:48:05.760 conservatives didn't believe that any of the following institutions were on their sides.
00:48:11.800 The court, the ballot,
00:48:14.360 MI5, MI6,
00:48:16.080 the police,
00:48:17.540 the GCHQ, if they believed that all of these institutions and more were against them,
00:48:22.840 these people would no longer be conservatives.
00:48:25.740 They would be something else, but they would not be conservatives.
00:48:28.840 You cannot be a conservative if you believe that there are no
00:48:32.380 institutions in the state that are trustworthy.
00:48:35.760 Since what's so strange about the Republican Party at the moment, you could argue it's not at all conservative.
00:48:41.500 I think there's a lot of truth in that.
00:48:44.220 And again, the problem with this is that there is an element in everything that they believe on this that is true.
00:48:49.900 It is true that the intelligence services, for instance, in the U.S., have massively politicized themselves unnecessarily.
00:48:57.560 They have therefore ended up losing the trust of even the political side that would be most likely to be nascently supportive of them as an institution of state.
00:49:08.680 Again, I mean, to give one other example,
00:49:12.180 when in our lifetimes before could you have imagined a situation,
00:49:15.540 not when the left derided and dismissed the heads of the armed forces,
00:49:19.420 including the chairman of the Joint Chiefs of Staff,
00:49:21.120 but where the right did, you know, that's the extraordinary terrain that we're now in.
00:49:26.980 Yeah, no, it was, and we were in there quite early on when, just during the campaign,
00:49:33.040 when Trump derided John McCain and his service as a, you know, a war hero and prisoner of war
00:49:39.780 and suffered absolutely no political penalty for it.
00:49:43.820 Oh, I remember very clearly sitting in a friend's house in America the day that that story broke on the front of the post,
00:49:51.580 and I remember this friend who'd been in politics all her life saying,
00:49:53.860 that's him done, you know, and of course it wasn't.
00:49:57.500 Well, I mean, it happened that on the right there wasn't that much love for John McCain,
00:50:00.540 it turned out there was a certain amount of respect and not much love for him.
00:50:02.980 But still, for somebody who had skipped the draft to be deriding somebody who'd spent years in a prison of war camp
00:50:12.880 and who had refused to leave until his men had left was a very, I agree, a very strange and sinister turn of things.
00:50:22.540 So where do you sit, this is an adjacent lurid topic,
00:50:25.900 where do you sit on the subject of de-platforming people like, well, Trump is one case from Twitter,
00:50:33.660 but maybe a clearer case is someone like Alex Jones.
00:50:37.100 Are you a free speech absolutist of the sort that you think that,
00:50:43.400 forgive me for leading the witness quite this hard,
00:50:46.580 think that private companies should be forced to give a megaphone to someone like Alex Jones,
00:50:52.500 who is, you know, with every tweet is ruining the lives of identifiable people?
00:50:59.980 In general, I am, yes.
00:51:01.780 I mean, I certainly think Trump should be on Twitter.
00:51:04.960 But why wouldn't you take the company's eye view of that?
00:51:10.120 You know, if I start a social media platform tomorrow,
00:51:14.000 why should I be forced to put any particular person on who I want to exclude?
00:51:18.240 Well, effectively, it's a thing that Elon Musk pointed out the other week when he started his bid for Twitter,
00:51:23.220 which is whether we like it or not, Twitter is the public square.
00:51:27.560 And this dance between private company and that is a tricky one.
00:51:32.920 It is a private company.
00:51:34.800 They can make their own decisions.
00:51:36.340 However, it is true that if the tech platforms decide to down-regulate you,
00:51:42.220 dampen you, or let alone chuck you off,
00:51:44.400 you are left essentially voiceless.
00:51:46.040 Yeah, but clearly, but all of these companies have terms of service,
00:51:50.740 which if you violate them by declaring a change,
00:51:53.980 but in any plausible terms of service,
00:51:57.680 you would think that ramping up the risk of nuclear war
00:52:01.680 or singling out private individuals who you know,
00:52:06.880 based on the insane multitude following you,
00:52:09.720 will be doxed immediately and have their lives ruined.
00:52:12.980 And then, in fact, that's why you target them, right?
00:52:16.900 But people do that all the time.
00:52:18.760 A Washington Post reporter just did that to someone.
00:52:21.460 You know, a Washington Post reporter, Taylor Lorenz,
00:52:24.520 just doxed this woman who,
00:52:26.620 the private individual who runs this account called Libs of TikTok,
00:52:29.240 that's pretty revealing, quite funny,
00:52:31.320 and has a big following.
00:52:32.340 And she had only a couple of weeks beforehand
00:52:36.580 been complaining about what it's like when a Twitter mob comes for you
00:52:40.040 and indeed cried on air talking about this.
00:52:42.020 A couple of weeks later, she doxed this person who runs this account.
00:52:44.140 I mean, this goes in all directions.
00:52:45.660 I'm not saying it makes Alex Jones right.
00:52:47.180 And by the way, in the Alex Jones case,
00:52:48.640 I mean, I think it's less clear with him
00:52:50.260 because, I mean, it's just so obvious.
00:52:53.140 But, I mean, the courts are taking care of him,
00:52:55.360 the families of the Sandy Hook victims
00:52:57.600 who he defamed and lied about.
00:53:00.040 He's being looked after in the courts.
00:53:02.320 I mean, it's a really, it's, I mean,
00:53:05.640 and just one other thing, it's an obvious point to make.
00:53:08.540 It doesn't quite solve the deeper point you're trying to get to,
00:53:10.740 but of who gets the right to the dart gun,
00:53:13.720 which is, of course, what Twitter is.
00:53:15.460 But it's nevertheless crucial to say
00:53:18.260 that if a platform like Twitter actually cared about,
00:53:21.660 you know, threatening entities on the site,
00:53:24.960 they wouldn't have kept the accounts
00:53:26.480 of the Russian government open all this time.
00:53:28.400 They wouldn't have the supreme leader of Iran.
00:53:31.080 Al-Akhshari Taibur carried out the Mumbai massacre,
00:53:33.760 remained on Twitter until a couple of years ago,
00:53:35.420 and I actually alerted one of the heads of the company
00:53:37.200 to the fact that I thought this was a bit too close to home
00:53:39.840 for most Indian citizens, and much more.
00:53:43.680 They're not fit for purpose.
00:53:45.380 Companies like Twitter grew,
00:53:47.680 they're a small thing that grew far too fast,
00:53:51.280 have ended up having to understand free speech,
00:53:53.680 and seem apparently not to have thought about the subject
00:53:56.000 until yesterday, and they're incredibly inept,
00:53:59.780 and they get inept people to ineptly police these platforms.
00:54:02.800 Yeah, well, I agree there,
00:54:04.900 and it may actually be an impossible task, right,
00:54:07.780 to actually moderate billions upon billions of posts effectively,
00:54:13.160 but it seems to me that if you, for instance,
00:54:15.880 had a no-doxing policy,
00:54:18.020 you know, doxing is an unrecoverable error on this platform,
00:54:23.100 well, then the question is,
00:54:25.700 you should be free to have,
00:54:26.840 I mean, in my view,
00:54:27.660 you should be free to have that policy,
00:54:29.440 and then do your best to enforce it,
00:54:31.820 and if you see irregularities in its enforcement,
00:54:34.360 well, then those are worthy of criticism.
00:54:36.120 So, a Washington Post writer is probably up for defenestration also,
00:54:42.880 if she doxes people,
00:54:45.380 but it just seemed clear that the most prominent examples of people,
00:54:50.420 I mean, in the case of Alex Jones,
00:54:52.220 you have parents whose six-year-olds were murdered,
00:54:57.660 and he was monetizing their agony
00:55:00.740 by claiming that they weren't murdered,
00:55:03.240 and that they were just crisis actors,
00:55:04.960 and there are some of these families
00:55:07.860 that have had to move,
00:55:09.580 literally change homes 10 times
00:55:12.280 since their kids were murdered
00:55:13.580 because of his insane cult that is following them.
00:55:16.300 Yeah, I know, it's wicked.
00:55:17.640 It's wicked, but the,
00:55:19.040 I mean, we come back to this thing of,
00:55:21.360 you know, where, how on earth you run this,
00:55:24.580 and clearly nobody exactly knows.
00:55:28.100 I mean, if you and I were on the board of Twitter,
00:55:29.840 I think we would struggle with it,
00:55:31.700 you know, as well.
00:55:32.560 Well, I don't think there's any obvious solution.
00:55:35.460 I know that,
00:55:36.160 I know there are some things
00:55:38.020 that are also, I think,
00:55:40.120 incumbent upon people not to do themselves
00:55:42.780 to make the situation worse.
00:55:45.120 Can I give a quick example?
00:55:46.500 Sure.
00:55:47.020 Which is not even in the realm of laws,
00:55:48.920 but in the realm of manners.
00:55:50.460 I am consistently horrified
00:55:52.440 by the number of people of,
00:55:54.180 particularly young people
00:55:55.780 who are willing to put out on social media things.
00:55:57.880 I simply think they should not put out.
00:56:00.460 I mean, my rule on this is
00:56:02.320 never, ever say anything
00:56:04.340 that you don't want to be used back at you,
00:56:06.580 because you can just bet your life it will be,
00:56:09.760 you know?
00:56:10.420 Send out a photo of yourself
00:56:12.420 and someone will say you don't look great, fine.
00:56:16.420 Send out a photo of yourself with your wife.
00:56:18.920 Somebody will say something about your wife.
00:56:20.820 That's the name of the game.
00:56:21.980 Send out a photograph of your children.
00:56:23.420 Not everyone's going to love your children,
00:56:24.860 and so on and so on.
00:56:26.840 Somebody I know a little bit recently divorced
00:56:29.160 announced the news on Twitter,
00:56:30.580 and a load of people get into it
00:56:32.860 and celebrate it and are laughing at him and so on.
00:56:35.000 And I just look at this and I think,
00:56:36.300 why on earth would you announce stuff
00:56:37.780 about your private life on this bloody platform?
00:56:40.520 So I do think part of it,
00:56:42.100 and this obviously isn't the case
00:56:43.180 with the Sandy Hook parents
00:56:44.320 or anyone who just were thrown into this situation,
00:56:47.920 but a lot of what people complain about
00:56:49.600 on social media of what they get back
00:56:51.440 is a result of them feeding the beast themselves
00:56:54.220 in the first place.
00:56:55.560 And there are things that if you put out there,
00:56:57.100 you're just not going to get 100% positive likes back.
00:57:01.860 It's an ugly medium.
00:57:03.260 It's an ugly platform.
00:57:04.860 And I have infinite compassion
00:57:06.160 for the people who suffer from it
00:57:07.380 and sort of what's happened
00:57:08.520 with the victims of Alex Jones,
00:57:10.300 but not when it comes to,
00:57:12.080 you know, the I said this thing
00:57:14.540 and now I've been criticized
00:57:15.840 and I'm upset
00:57:17.300 and now I've got PTSD sort of thing.
00:57:19.060 And I'm afraid that is so commonplace now,
00:57:21.840 the cry bully thing,
00:57:23.320 where people behave one way on social media
00:57:25.660 and can't take it in response
00:57:28.460 and complain then.
00:57:30.680 You know, there are plenty of cry bullies
00:57:32.440 on these platforms
00:57:33.640 and I don't have much sympathy for that.
00:57:36.420 Yeah, well, I think that it's just a natural fact
00:57:39.600 that we're new to this situation
00:57:42.400 that evolution,
00:57:44.100 neither evolution nor previous culture
00:57:47.680 has prepared us,
00:57:48.820 which is you can join a mob,
00:57:51.860 a virtual mob,
00:57:52.980 and perform a reputational murder on someone
00:57:56.900 and you can be the object of a mob like that
00:57:59.760 and you never quite know what it's like
00:58:02.860 until you're on the receiving end.
00:58:05.620 Yes, but also, I mean,
00:58:06.880 people do have to try to find a way
00:58:08.580 to live their lives
00:58:09.440 without this having maximal impact.
00:58:11.280 I mean, you know,
00:58:13.660 I do think there's sort of,
00:58:15.880 as I say,
00:58:16.340 the realm of manners in this,
00:58:17.740 the realms of customs,
00:58:19.380 that we should also try to come
00:58:20.960 towards a better type of custom
00:58:22.440 with these platforms, you know.
00:58:24.560 And in the same way
00:58:25.340 that we did with email early on,
00:58:26.640 do you remember
00:58:26.880 sort of the beginning of email,
00:58:28.220 you know,
00:58:28.420 people would sort of pass around
00:58:29.540 crazy stories
00:58:30.900 about how if you eat tomatoes,
00:58:32.780 you know,
00:58:33.180 you will never get cancer,
00:58:35.280 that sort of thing.
00:58:36.120 And just quite early on,
00:58:37.280 those sorts of people
00:58:38.160 who would send those things around
00:58:39.420 learned that people didn't want
00:58:41.140 to get them and stopped.
00:58:43.160 At least that was my experience.
00:58:44.500 You know,
00:58:44.640 please don't send me this shit.
00:58:46.600 Thank you.
00:58:47.740 And they stopped.
00:58:48.800 I mean,
00:58:48.980 we're just not quite there yet
00:58:50.900 or not remotely there yet
00:58:52.660 with a platform as furious
00:58:54.240 and as fast as Twitter.
00:58:56.120 So, I mean,
00:58:56.820 you know,
00:58:57.220 we have to assist our own behavior
00:59:00.360 as well as hoping
00:59:01.700 that Twitter can solve
00:59:03.280 its side of the problem.
00:59:05.380 Yeah, yeah.
00:59:06.440 Okay, well,
00:59:06.940 I'm not sure
00:59:07.660 how much we disagreed in there.
00:59:09.520 I think we,
00:59:10.680 if we were on the Twitter board,
00:59:12.500 we might disagree
00:59:13.120 about who to de-platform.
00:59:15.160 But would you acknowledge that,
00:59:17.080 I mean,
00:59:17.340 do you just think
00:59:18.140 Twitter should be declared
00:59:20.300 essentially,
00:59:21.400 effectively,
00:59:21.880 no longer
00:59:22.580 a private company
00:59:24.500 able to function
00:59:26.320 by its own
00:59:27.120 policies,
00:59:28.420 but more like
00:59:29.800 the town square
00:59:31.580 that just has to function
00:59:32.620 in deference
00:59:33.400 to the First Amendment?
00:59:34.780 Yeah,
00:59:34.960 I think it's basically
00:59:35.840 the Wild West
00:59:36.700 and there's not much
00:59:38.720 you can do about it
00:59:39.480 and you have to decide
00:59:40.560 whether you want to go
00:59:41.260 into the saloon.
00:59:42.820 Right.
00:59:43.120 It just seems strange
00:59:44.540 that,
00:59:45.100 because again,
00:59:45.540 I take the company's eye view
00:59:47.000 of this.
00:59:47.420 You start a company,
00:59:48.160 for instance,
00:59:49.060 I think,
00:59:50.060 you know,
00:59:50.420 actually,
00:59:51.100 I recommended that Jack Dorsey
00:59:52.920 pull the plug on Twitter
00:59:54.020 at some point
00:59:54.580 and he would have been given
00:59:55.320 the Nobel Prize for peace.
00:59:56.560 Yes,
00:59:57.640 I suggested Elon Musk
00:59:58.740 send it into outer space.
01:00:00.380 Yeah,
01:00:00.640 so Elon's buying Twitter
01:00:02.660 as of the hour
01:00:04.260 we're recording this,
01:00:05.280 it seems that's happening
01:00:06.100 and taking it private.
01:00:09.460 If he's doing that
01:00:10.480 all with his own money,
01:00:11.980 I'm not sure he is,
01:00:13.120 but let's say he were
01:00:14.520 to do that,
01:00:15.660 couldn't he just destroy
01:00:18.380 the entire thing
01:00:19.300 and say,
01:00:20.360 I'm doing you all a favor?
01:00:21.940 He could do.
01:00:22.960 He could do.
01:00:23.940 I mean,
01:00:24.160 the interesting thing
01:00:24.880 about it,
01:00:25.740 as you know,
01:00:26.420 is that the people
01:00:26.900 who've tried to set up
01:00:27.740 rivals to it,
01:00:29.060 it doesn't actually work.
01:00:30.480 I mean,
01:00:30.640 somebody said to me
01:00:31.260 the other day,
01:00:31.740 you know,
01:00:31.900 assume that Elon's
01:00:32.540 a rather smart guy.
01:00:34.280 There must be a reason
01:00:35.060 why he hasn't tried
01:00:35.880 to start his own Twitter.
01:00:37.560 Right.
01:00:38.100 Well,
01:00:38.300 I mean,
01:00:38.460 it's just hard to get to,
01:00:39.960 once you get the kind
01:00:41.980 of traction people have
01:00:43.700 on Twitter,
01:00:44.200 I mean,
01:00:44.440 he's got 80 million
01:00:45.780 plus people following him.
01:00:47.100 It's hard to imagine
01:00:48.020 starting that on a new platform,
01:00:50.580 but,
01:00:50.880 but,
01:00:51.080 yeah,
01:00:51.900 but just to close the loop
01:00:53.720 on where I was going,
01:00:54.880 there,
01:00:55.620 if it would be
01:00:56.320 within the bounds
01:00:57.140 of propriety
01:00:58.620 to actually just
01:01:00.780 pull the plug on it,
01:01:02.860 you know,
01:01:03.280 effectively cancelling
01:01:05.800 or deplatforming everyone,
01:01:08.240 why can't you
01:01:09.340 deplatform Alex Jones
01:01:10.960 for his misbehavior?
01:01:13.240 Well,
01:01:13.520 as I say,
01:01:14.100 I mean,
01:01:14.260 I would regard him
01:01:15.000 as being a borderline case.
01:01:16.820 I don't know.
01:01:17.660 I mean,
01:01:17.860 I think basically
01:01:19.140 once you get into
01:01:19.700 the realms of harassment,
01:01:20.740 which is where he,
01:01:21.980 he was,
01:01:22.980 that's,
01:01:24.000 that's a viable case
01:01:26.520 for taking somebody off.
01:01:27.680 I haven't really thought
01:01:28.360 enough about his particular case
01:01:29.960 because I don't follow him
01:01:30.840 very closely.
01:01:31.920 But I know that,
01:01:33.020 I know that it's basically
01:01:34.040 unsustainable
01:01:34.900 that an American company
01:01:36.240 deplatforms
01:01:37.280 the U.S. president
01:01:38.160 and doesn't deplatform
01:01:40.020 the Ayatollah of Iran.
01:01:41.640 It's,
01:01:41.960 yeah,
01:01:42.580 but I just hear that
01:01:43.560 as an argument
01:01:44.080 for deplatforming
01:01:45.320 both of them
01:01:46.040 given who they are.
01:01:48.280 that could be the case,
01:01:50.660 although I wouldn't agree
01:01:51.940 with the moral equivalency,
01:01:52.940 but yeah,
01:01:53.520 it could be the case.
01:01:54.720 I mean,
01:01:55.780 as I say,
01:01:56.440 all of these things
01:01:57.300 set themselves up
01:01:58.320 and end up having to run
01:02:00.100 the town square
01:02:01.120 and they're just clearly
01:02:01.960 not suitable for the task.
01:02:03.320 I don't,
01:02:03.740 I don't know
01:02:04.200 what all the answers are
01:02:05.320 to it,
01:02:06.120 but as I say,
01:02:06.820 that,
01:02:06.980 that's,
01:02:07.440 most people don't like
01:02:08.760 the unfairness thing.
01:02:09.920 I mean,
01:02:10.760 personally,
01:02:11.220 you know,
01:02:11.400 I don't,
01:02:12.320 I think it was wrong
01:02:12.920 to throw Trump off Twitter,
01:02:15.400 but,
01:02:15.920 you know,
01:02:16.280 I mean,
01:02:16.520 there are obviously upsides.
01:02:17.780 I mean,
01:02:17.900 not least that every day's news
01:02:19.120 is now not about Donald Trump
01:02:20.520 and what he tweeted today.
01:02:22.180 I mean,
01:02:22.320 that's quite a relief.
01:02:23.220 Yeah.
01:02:23.700 No,
01:02:23.920 it definitely
01:02:24.740 had the desired effect,
01:02:27.400 right?
01:02:27.640 I mean,
01:02:27.840 he's not,
01:02:28.720 he's not gone,
01:02:29.920 but he,
01:02:30.460 he,
01:02:31.180 he is in a kind of oblivion
01:02:33.360 with respect to the rest of culture
01:02:35.160 and the news cycle,
01:02:36.780 and I think that,
01:02:38.260 that was a good thing.
01:02:38.980 I mean,
01:02:39.140 I do view,
01:02:39.700 you know,
01:02:39.920 again,
01:02:40.360 it's only,
01:02:42.020 it is a historical fact
01:02:43.920 that he was president
01:02:45.420 of the United States,
01:02:46.500 but I actually think
01:02:48.120 it's more accurate
01:02:48.960 to describe him now
01:02:50.220 as the most dangerous
01:02:51.840 cult leader on earth,
01:02:53.860 right?
01:02:54.060 I mean,
01:02:54.180 I just,
01:02:54.600 I just think he's,
01:02:55.280 there's just no telling
01:02:56.880 what harm he's capable
01:02:58.180 of creating
01:02:58.860 if he manages
01:03:00.500 to continue
01:03:02.020 to hold
01:03:02.620 half of American society
01:03:04.120 or a third
01:03:04.660 of American society
01:03:05.640 in his thrall,
01:03:07.600 and it's just,
01:03:09.320 it's the most deranging
01:03:10.360 thing to happen
01:03:11.160 in our lifetime,
01:03:12.460 including a global pandemic.
01:03:14.440 I think it's one of them.
01:03:16.500 I don't think it's the most deranging.
01:03:18.080 I think it's one of the most deranging.
01:03:19.940 But we've lived in several very,
01:03:22.260 very long years
01:03:23.520 where every day
01:03:26.660 has enough information
01:03:28.020 to derange some people.
01:03:29.640 And whereas I've often said
01:03:32.000 that the range
01:03:34.000 of the things
01:03:34.560 that have come across us
01:03:35.660 and afflicted us
01:03:36.660 from Trump
01:03:37.860 to pandemic
01:03:39.180 to Afghanistan
01:03:41.800 to Ukraine
01:03:42.780 to all of this stuff,
01:03:44.240 the range of things
01:03:45.480 means that almost nobody
01:03:47.060 is ending up
01:03:47.820 in exactly the same place
01:03:49.060 as their erstwhile bedfellows.
01:03:50.700 Except the derangement
01:03:52.380 of all of those other things,
01:03:54.200 so much of the onus
01:03:55.120 of that falls on Trump.
01:03:56.700 I mean, so explain to me
01:03:59.540 why you have Republicans,
01:04:02.280 you know,
01:04:02.560 otherwise sane,
01:04:04.120 well-intentioned human beings,
01:04:06.660 one must presume,
01:04:08.220 lionizing Vladimir Putin
01:04:10.240 at this point in history.
01:04:12.480 I mean, that would have been unthinkable,
01:04:13.900 I think,
01:04:15.060 but for Trump, right?
01:04:17.340 I mean, is there some other mechanism
01:04:18.400 that got that meme
01:04:19.440 into their heads?
01:04:21.000 Yes, I think so.
01:04:22.000 I think that there is,
01:04:22.700 as I say,
01:04:23.080 I've written repeatedly
01:04:24.020 against these people,
01:04:24.800 but I think that there is a section
01:04:28.660 of the right
01:04:29.340 that was misled
01:04:30.880 by Vladimir Putin
01:04:32.260 and whether it's stupidity
01:04:34.660 or ignorance
01:04:35.360 or generational loss of memory,
01:04:37.900 I don't know.
01:04:38.740 I think it's a combination
01:04:39.780 of all of these things,
01:04:40.920 but I mean,
01:04:41.660 there was certainly an element
01:04:42.560 of the right
01:04:43.040 that in recent years
01:04:43.880 has said
01:04:44.220 that American liberalism
01:04:45.900 has gone so wrong
01:04:47.260 that we need a bulwark
01:04:49.100 against it
01:04:49.800 and the bulwark
01:04:50.560 is,
01:04:52.620 and they looked around
01:04:53.420 for people,
01:04:53.980 but one of them
01:04:55.160 that some of them
01:04:55.660 landed on
01:04:56.040 was Vladimir Putin
01:04:56.880 and I always said
01:04:58.460 this is so monumentally stupid
01:05:01.160 among other things
01:05:02.400 because you had to take
01:05:03.220 Vladimir Putin
01:05:03.880 at his own word.
01:05:05.000 I mean,
01:05:05.400 you had to actually pretend
01:05:06.540 that he was this devout,
01:05:08.700 pious ex-KGB man
01:05:10.480 who said his prayers
01:05:11.700 and was going to lead
01:05:12.580 Christendom to revival.
01:05:14.660 You know,
01:05:14.880 you had to actually believe
01:05:16.280 he was sincere in that
01:05:17.840 and if you believe that,
01:05:18.700 you're a damn fool.
01:05:20.040 You know,
01:05:20.280 you had to believe
01:05:20.880 that somebody
01:05:21.380 who's used
01:05:22.520 jihadi mercenaries
01:05:24.100 from Chechnya
01:05:24.960 to go and slaughter
01:05:25.840 Ukrainian orthodox Christians,
01:05:28.100 not that one should need
01:05:29.240 to talk in these terms,
01:05:30.200 but let's talk in those terms
01:05:31.380 for a moment,
01:05:31.980 that that person
01:05:32.700 is somehow the defender
01:05:33.900 of the Christian faith.
01:05:35.060 I mean,
01:05:35.240 it's so unbelievably stupid,
01:05:38.280 but there is,
01:05:39.080 yes,
01:05:39.300 there was an element
01:05:40.400 of the right
01:05:40.840 in recent years
01:05:41.480 that fell for that,
01:05:42.120 plus some
01:05:42.720 that just had no history
01:05:43.920 and memory
01:05:44.340 of the Cold War,
01:05:45.780 no memory
01:05:46.660 of what the KGB
01:05:48.680 or the Kremlin,
01:05:50.620 let's just say,
01:05:51.440 does,
01:05:52.200 and then there were
01:05:52.760 the people
01:05:53.120 who were so fed up
01:05:54.040 with false claims
01:05:55.120 of overstatements
01:05:56.960 of what Russia
01:05:57.720 had done in recent years
01:05:58.800 that they believed
01:05:59.700 that the Kremlin
01:06:00.180 was a sort of
01:06:00.800 quiescent pacifist-like institution
01:06:03.220 that never did anything.
01:06:04.880 You know,
01:06:05.280 I mean,
01:06:05.760 it's maddening in itself,
01:06:07.340 but there were so many people
01:06:08.840 who fell for this
01:06:09.600 and I thought
01:06:10.040 that it was
01:06:10.440 an unbelievable error
01:06:12.420 and what has happened
01:06:13.680 in Ukraine
01:06:14.360 is one of the fastest
01:06:17.220 demonstrations
01:06:18.880 of a moral error
01:06:20.280 that I've ever seen.
01:06:22.280 Okay,
01:06:22.820 well,
01:06:22.960 let's take the turn
01:06:23.620 toward the left
01:06:24.980 and its assault
01:06:26.340 on Western civilization
01:06:27.360 where you and I
01:06:29.240 will be singing
01:06:30.520 from the same hymn book
01:06:31.700 because you wrote
01:06:32.540 the hymn book
01:06:33.040 and I just read it,
01:06:35.420 but I think,
01:06:36.500 so I think we,
01:06:37.860 just as you thought
01:06:39.080 it was,
01:06:39.440 unseemly
01:06:41.060 to have to spell out
01:06:42.380 the problem
01:06:43.200 with recruiting
01:06:44.880 jihadi mercenaries
01:06:46.180 to attack
01:06:47.160 Orthodox Christians
01:06:48.180 as though
01:06:49.100 the identities
01:06:50.000 would be
01:06:52.500 especially relevant
01:06:53.880 to the moral calculus
01:06:54.720 there.
01:06:55.380 I'm going to lead us
01:06:56.600 in a brief exercise
01:06:57.880 of masochism
01:06:59.960 to just
01:07:01.280 inoculate
01:07:02.420 some people
01:07:03.220 in our audience
01:07:03.820 against
01:07:04.280 the rest
01:07:05.480 of our conversation.
01:07:06.360 I think,
01:07:06.560 I think we need
01:07:07.700 to start
01:07:08.180 with the obligatory
01:07:09.080 acknowledgement
01:07:09.580 that we are
01:07:10.340 two white guys
01:07:11.800 about to express
01:07:13.160 our opinions
01:07:13.780 on many combustible
01:07:15.680 topics
01:07:16.220 and I know,
01:07:18.500 I mean,
01:07:18.840 this is,
01:07:20.280 I guess,
01:07:21.220 half tongue-in-cheek
01:07:22.280 but not entirely.
01:07:24.260 First,
01:07:25.520 let me just spell out,
01:07:26.580 I consider it
01:07:27.240 morally
01:07:27.920 and intellectually
01:07:28.840 obscene
01:07:30.160 to have to take note
01:07:31.780 of our skin color
01:07:32.640 as a preface
01:07:33.680 to this conversation
01:07:34.560 but the truth is
01:07:36.600 that there are only
01:07:37.100 so many hills
01:07:37.820 I'm willing to die on
01:07:38.880 and I do think
01:07:40.300 it's prudent
01:07:40.940 for us to acknowledge
01:07:42.480 what any sane
01:07:44.340 and compassionate
01:07:44.960 person knows
01:07:46.280 to be true
01:07:46.920 which is that
01:07:48.020 racism
01:07:48.840 and other forms
01:07:49.640 of bigotry
01:07:50.320 are odious
01:07:51.660 and that
01:07:52.840 Western culture
01:07:53.700 has been replete
01:07:55.360 with bigotry
01:07:55.900 of all types
01:07:56.560 as has every
01:07:57.740 culture
01:07:58.680 and there is
01:08:00.820 certainly some
01:08:01.460 residual racism
01:08:02.400 and bigotry
01:08:03.280 left to expunge
01:08:05.000 and there is
01:08:06.100 nothing that
01:08:06.920 you and I
01:08:07.540 will say
01:08:08.160 that should suggest
01:08:10.120 an unawareness
01:08:11.120 much less a denial
01:08:12.160 of these facts
01:08:13.640 so I don't know
01:08:15.980 if that's
01:08:16.420 I'm sure you can
01:08:17.420 more or less
01:08:17.980 sign on to that
01:08:18.580 I don't know
01:08:18.800 if you have anything
01:08:19.220 to add
01:08:19.740 No, I'd just add
01:08:21.240 that it seems
01:08:22.880 to me
01:08:23.500 that one of the
01:08:25.100 great disappointments
01:08:26.060 of what I describe
01:08:26.800 as the re-racialization
01:08:28.180 of the public square
01:08:30.060 one of the great
01:08:31.220 disappointments about it
01:08:32.180 is that you even
01:08:32.940 have to talk about
01:08:33.660 yourself in terms
01:08:34.520 of skin color
01:08:35.240 and that it seems
01:08:36.480 so obvious to me
01:08:37.480 and has done
01:08:38.060 for as long as I can
01:08:39.340 remember
01:08:39.860 that you would
01:08:40.740 in any way
01:08:41.780 identify yourself
01:08:42.820 because of it
01:08:44.200 it's like
01:08:45.360 somebody asked me
01:08:46.460 recently
01:08:46.980 what I was proud
01:08:48.380 about
01:08:48.720 about being a man
01:08:49.940 and I said to her
01:08:50.880 this is the second
01:08:52.260 stupidest question
01:08:53.340 I can imagine
01:08:54.000 after being asked
01:08:54.900 what I was proud
01:08:55.540 about about being
01:08:56.200 white
01:08:56.600 you know
01:08:56.980 I'm not proud
01:08:58.140 of things I haven't
01:08:59.160 done
01:08:59.640 I don't see why
01:09:01.160 you would be
01:09:01.940 it's like being
01:09:03.000 proud of being
01:09:03.820 5'10
01:09:04.620 like what the hell
01:09:06.500 is that
01:09:07.320 so yes
01:09:08.980 I think it's
01:09:10.120 it's just
01:09:11.040 already deplorable
01:09:11.820 at the start
01:09:12.300 that we are being
01:09:14.000 urged to think
01:09:15.240 of ourselves
01:09:15.740 in these terms
01:09:16.620 because they are
01:09:17.740 precisely the terms
01:09:18.680 that I had been
01:09:19.520 brought up to
01:09:20.380 regard as being
01:09:21.360 so unimportant
01:09:22.640 that we didn't
01:09:24.320 talk about that
01:09:25.060 well I think
01:09:26.480 you and I agree
01:09:27.140 that the
01:09:27.620 the appropriate
01:09:28.560 goal here
01:09:29.640 with respect to
01:09:31.140 political and
01:09:32.320 moral progress
01:09:32.980 is to arrive
01:09:34.580 in some
01:09:35.100 happy future
01:09:36.320 where race
01:09:38.100 simply does
01:09:39.480 not matter
01:09:40.040 it has
01:09:40.820 no moral
01:09:42.080 or political
01:09:42.840 valence
01:09:44.200 to it
01:09:44.780 no one cares
01:09:45.620 you said to me
01:09:46.020 once I think
01:09:47.660 maybe in a
01:09:48.200 previous discussion
01:09:48.920 that it would
01:09:49.740 end up having
01:09:50.200 as much importance
01:09:51.040 as your hair
01:09:52.260 color
01:09:52.600 yeah
01:09:52.920 and if you just
01:09:54.700 roll that back
01:09:55.680 in the other
01:09:55.960 direction
01:09:56.280 imagine
01:09:56.760 how insane
01:09:58.160 and counterproductive
01:09:59.880 it would seem
01:10:00.540 if we could
01:10:02.020 look ahead
01:10:02.780 and predict
01:10:04.120 that at some
01:10:05.240 point in the
01:10:05.600 future
01:10:05.920 people were
01:10:06.740 going to care
01:10:07.280 about hair
01:10:08.100 color
01:10:08.460 to the degree
01:10:09.380 that they
01:10:09.660 currently care
01:10:10.320 about race
01:10:11.460 we'll want to
01:10:12.520 know how many
01:10:12.900 blondes got into
01:10:13.700 Harvard this
01:10:14.360 year
01:10:14.640 and if it
01:10:15.220 doesn't exactly
01:10:15.880 match the
01:10:16.540 population level
01:10:17.560 we've got a
01:10:18.100 real problem
01:10:18.820 on our hands
01:10:19.280 yes I'd go
01:10:20.600 into a bookshop
01:10:21.180 and say I'd
01:10:21.500 like to see
01:10:22.180 the section
01:10:22.680 written by
01:10:23.100 ginger head
01:10:23.740 authors
01:10:24.180 yeah
01:10:25.940 so yeah
01:10:27.100 I mean just
01:10:27.600 as that would
01:10:27.980 be a problem
01:10:28.580 the the
01:10:29.840 ethical daylight
01:10:31.120 ahead of us
01:10:31.760 is in
01:10:32.420 arriving at
01:10:33.980 some
01:10:34.440 colorblind
01:10:35.480 future
01:10:36.100 and yet
01:10:37.520 not only
01:10:38.520 is and that
01:10:39.160 was the goal
01:10:39.980 of someone
01:10:40.440 like you
01:10:41.260 know the
01:10:41.660 leading lights
01:10:42.440 of the
01:10:42.740 civil rights
01:10:43.400 movement
01:10:43.760 people like
01:10:44.660 Martin Luther
01:10:45.040 King Jr
01:10:45.500 but it's
01:10:46.480 not only
01:10:46.900 not the
01:10:47.520 goal of
01:10:48.540 the current
01:10:49.280 religion of
01:10:50.400 anti-racism
01:10:51.160 it's explicitly
01:10:52.140 not the goal
01:10:53.080 that goal is
01:10:53.940 disavowed
01:10:54.700 by many people
01:10:56.100 yes
01:10:56.720 yeah
01:10:57.260 well that's
01:10:58.320 that's I mean
01:10:59.020 this is the
01:10:59.540 great moral error
01:11:00.260 that's going on
01:11:00.920 and I think
01:11:01.920 there's a little
01:11:03.780 bit of a link
01:11:04.240 to my previous
01:11:04.820 book
01:11:05.080 because in
01:11:05.380 matters
01:11:05.740 of crowds
01:11:06.320 you know
01:11:06.760 I described
01:11:07.840 what I said
01:11:08.340 and what I
01:11:08.980 described
01:11:09.240 as the
01:11:09.640 temptation
01:11:10.180 if you'd like
01:11:11.880 to continue
01:11:12.300 listening to
01:11:12.760 this conversation
01:11:13.480 you'll need
01:11:14.320 to subscribe
01:11:14.920 at
01:11:15.140 samharris.org
01:11:16.060 once you do
01:11:17.280 you'll get
01:11:17.580 access to
01:11:18.060 all full-length
01:11:18.740 episodes of
01:11:19.300 the making sense
01:11:19.800 podcast
01:11:20.360 along with
01:11:21.180 other subscriber
01:11:21.880 only content
01:11:22.600 including bonus
01:11:23.840 episodes
01:11:24.460 and AMAs
01:11:25.500 and the
01:11:26.060 conversations
01:11:26.540 I've been
01:11:26.880 having on
01:11:27.260 the waking
01:11:27.540 up app
01:11:28.020 the making
01:11:28.960 sense
01:11:29.200 podcast
01:11:29.580 is ad
01:11:30.160 free
01:11:30.480 and relies
01:11:31.360 entirely on
01:11:32.080 listener support
01:11:32.740 and you can
01:11:33.720 subscribe now
01:11:34.520 at
01:11:35.020 samharris.org
01:11:35.920 on
01:11:41.760 Youtube
01:11:42.200 and
01:11:44.420 on
01:11:45.260 daily
01:11:45.980 time
01:11:46.880 on
01:11:47.860 night
01:11:49.400 the
01:11:50.160 day
01:11:50.840 it
01:11:52.340 is
01:11:52.560 the
01:11:52.860 night
01:11:54.620 we
01:11:56.060 miss
01:11:56.800 the
01:11:58.060 witness
01:11:58.660 I
01:12:00.040 was
01:12:00.060 I
01:12:01.160 I
01:12:01.600 was
01:12:01.860 I
01:12:03.220 I
01:12:05.360 I
01:12:05.500 I
01:12:05.800 I