How Twitter Censored COVID Dissent - David Zweig
Episode Stats
Length
1 hour and 12 minutes
Words per minute
180.95639
Harmful content
Toxicity
41
sentences flagged
Hate speech
9
sentences flagged
Summary
Author and journalist David Zweig joins Francis and Constantine to talk about his new book, The Twitter Files, and how he got access to one of the most powerful people in the world, Elon Musk. He also talks about how he went from a rock star to a journalist, and why he thinks we should all be worried about what s happening on the internet.
Transcript
00:00:00.120
So it's this idea of like, is a journalist just like a megaphone?
00:00:04.960
You know, legacy media outlets by and large were all on board with a certain narrative.
00:00:11.140
People needed a place like Twitter where they could be exposed to different ideas.
00:00:18.160
Yes, but many of those ideas were absolutely legitimate.
00:00:22.300
This is a place for legitimate debate by people who understand the issues.
00:00:32.900
There was a Trump tweet about how he said something like, don't be afraid anymore of COVID.
00:00:42.680
Shouldn't we be like, you know, labeling this as misleading or take it down?
00:00:46.580
And some of the other executives had to write back and say, optimism is not against our policy.
00:00:53.060
My God, this like super high level person actually even entertained the idea for something like that to be suppressed.
00:01:03.920
It was just a bizarre experience that I will never get over.
00:01:11.000
Hello and welcome to Trigonometry on the Road from the USA.
00:01:29.020
And this is a show for you if you want honest conversations with fascinating people.
00:01:34.420
Our brilliant guest today is an author and journalist, David Zweig.
00:01:40.240
So before we get cracking, tell everybody who are you, how are you, where you are.
00:01:43.860
What has been the journey through life that leads you to be sitting here talking to us?
00:01:55.140
I write news articles, investigative pieces, commentary.
00:02:04.100
Before that, I was playing music for a number of years.
00:02:06.980
The joke I like to tell people is, well, it wasn't a joke.
00:02:11.280
That after I'd been trying to succeed as a musician for many years, I went home and I
00:02:19.400
You know, and this is very hard of them having a kid who's, you know, trying to do that.
00:02:22.280
And I said, look, I'm no longer trying to make it as a rock star.
00:02:27.580
I said, instead, I'm going to be writing a novel.
00:02:37.260
But finally, it seems to have been working out.
00:02:40.280
So I've kind of hit my stride, I think, in the last decade working in nonfiction and doing
00:02:46.680
And here I am, I guess, chatting with you guys.
00:02:49.000
And one of the things we really wanted to talk to you about is the Twitter files.
00:02:52.540
And to see, I still don't know what I think about it, because on the one hand, I was one
00:02:57.400
of the people who was like super excited about seeing, you know, lifting the curtain, so
00:03:03.340
But I also felt that when we did, we kind of only really saw what we already knew.
00:03:10.360
Now, I'm not saying that's accurate, but that was my impression.
00:03:13.840
So what is it about the Twitter files that people should actually think, oh, that was worthwhile,
00:03:19.140
you know, all this hype and all the rest of it?
00:03:23.800
I can only speak about my own reporting, or at least I can only speak-
00:03:30.300
What was it like going in there and, you know, how did it work?
00:03:33.900
So I was brought out there by Barry Weiss, who, as most people probably know, Elon Musk gave
00:03:46.100
And then some short period of time after that, he gave Barry access.
00:03:56.540
And I had sent an email to Barry and one or two of her editors who I knew, because I had
00:04:03.240
And I said, because I've been writing about COVID stuff for, since the beginning of the
00:04:07.740
pandemic, I didn't know what they were doing, if they were going back.
00:04:13.380
Look, if you guys are going to be sending someone else back there, or if you're going back
00:04:19.480
yourself, these are the things I might look for.
00:04:23.380
Like, I thought it was a little maybe presumptuous of me to, I'm like, who am I to tell them?
00:04:27.160
But I was just, these are a few things I would do, do, do, do, do, do.
00:04:30.160
And like an hour later, I get an email back, can you get on a plane to San Francisco?
00:04:35.940
So they, you know, they figured, you know, I guess based on that and everything that
00:04:40.260
they know about me, that I should be the one to go out and at least initially look into
00:04:46.160
Because I think up to that point, the reporting wasn't specific to the pandemic.
00:04:50.620
It was on a variety of other things, Hunter Biden, laptop, and other stuff.
00:04:54.680
So I went out there, really the sort of, I don't know if this is the right word, but the
00:04:59.640
I'm like, this is the one lane that I'm going to travel in, which is related to COVID.
00:05:03.820
And so I was there, it was only a few days while I was there.
00:05:08.120
Michael Schellenberger was there, Leighton Woodhouse, Lee Fong.
00:05:12.120
So it was the three of them and me, and each of us working on our own things.
00:05:17.160
But everyone was helping each other as well, just because it was so challenging to look
00:05:25.920
I basically, I mean, I could get into a little bit of details about what the process,
00:05:31.000
Um, yeah, because I think there, I think there's still a fair amount of misunderstanding or
00:05:43.040
I think what a lot of people don't know or realize is it's not like going into some,
00:05:48.440
I don't know, company or some law firm where they're like, open up the books and you can,
00:05:54.540
There were very specific ways that we were able to search.
00:05:57.440
There were no restrictions on what we could look up or, or publish.
00:06:01.500
Um, that's important to note, but the process of actually searching was really challenging.
00:06:07.360
This is not like what we're used to as regular citizens, like Googling something or whatever.
00:06:12.440
The, the, the systems in place to actually do the digging were, um, were not very user-friendly.
00:06:19.420
So I would break it down into, there's basically two separate paths that we could search down.
00:06:25.520
One was looking at what I would call like the log files of individual accounts.
00:06:30.420
Um, and there was an engineer in the room with us on a special laptop and we could give him
00:06:34.880
someone's name and say, I would like to look up this person's account.
00:06:38.420
We were not able to see any like private information on anyone's account, which is
00:06:42.320
important for me to note because there were tons of lawyers involved in this.
00:06:45.560
Um, but what we were able to see were on someone's account, their activity and what Twitter internally
00:06:53.060
had, if there were any flags or marks on certain things on the account as a whole or on certain
00:06:58.420
And then there were tons of kind of files within files within files.
00:07:04.300
And then the other one was looking up, um, internal communications or internal and external,
00:07:12.380
And so the way that those searches were not performed in the room, we had to basically
00:07:19.640
I don't know if they were like next door to us or what, but they were, it wasn't directly
00:07:23.140
And then sometime later, a person would come in with a different laptop and say, here's
00:07:30.160
So if you imagine we're looking at, and we had to look up specific employees, you couldn't
00:07:34.580
just say, I want to find anything on myocarditis.
00:07:37.300
No, you had to give, and I'm not sure the reasoning behind this, but you had to pick certain
00:07:42.320
employees and they could look up that employee's emails and we could say, I want to see all
00:07:46.760
the emails from this person from this date to this date.
00:07:50.180
Um, and then, so we, we could get 4,000 emails from someone if you're looking at, so, and I
00:07:56.000
have, you know, X number of hours and wanted to look at this stuff.
00:08:00.520
It's more, it's not just like, oh, let's just do a word search and we're done.
00:08:05.980
So you're trying to sift through all this stuff to find what you're looking for.
00:08:13.620
Most of the time I knew what I was looking for and it was just seeing if it was there
00:08:17.700
And other times you don't even know what you're looking for.
00:08:19.460
It's just like, I know this is an important person.
00:08:21.840
I know this particular topic is something that we might have reason to believe something
00:08:28.720
If something's interesting or not interesting or newsworthy or not newsworthy.
00:08:32.020
And that was so, it was very laborious, very tedious to do these types of searches.
00:08:37.620
The stuff on the log files, I mean, it could take a very long time to like dig through and
00:08:42.320
you find a particular tweet and then, and then through there, there's all sorts of code
00:08:47.880
And then the engineer was basically explaining to me, okay, here we go.
00:08:50.700
I can see this tweet, what's labeled on here, um, or not labeled.
00:08:58.680
And I was going to ask, so what role did Elon have in this?
00:09:06.860
Did you have to say, look, this is what we found?
00:09:09.660
How involved was he in this entire investigation?
00:09:19.300
Yeah, I mean, again, at least from my, my experience, um, I had no personal interaction
00:09:29.080
Um, my understanding is, and this was through talking with Barry, was that, um, he, we had,
00:09:38.300
And, um, the only, uh, sort of rule was whatever we found, we had to publish it on Twitter first.
00:09:46.220
And, you know, and it didn't need to be far in advance, but it could be five minutes, but
00:09:51.760
Other than that, it wasn't, if you find X, Y, or Z, you're not allowed to do this, et cetera.
00:09:56.080
And as far as I'm aware, unless there was some incredible conspiracy happening, um, there
00:10:01.880
was, similarly, there was no, um, constraints placed on us to get the information.
00:10:07.220
So, both in regards to what we were able to look for and then what we found, neither of
00:10:15.700
You know, this engineer is, I'm literally looking over the person's shoulder as they're
00:10:20.240
I cannot imagine how that could have been in some way compromised without us knowing.
00:10:25.080
It's, some other guy had interviewed me and he kept saying, well, maybe the email searches
00:10:32.080
And again, we're talking about thousands of emails.
00:10:35.020
I can't imagine how they could possibly have done this.
00:10:38.280
And it took us, a team of four people in real time looking, so I don't see, because they
00:10:42.600
didn't know what we were going to be searching for.
00:10:46.440
It was very kind of like by the seat of your pants, just trying to do it.
00:10:49.960
I saw Elon while we were there, um, but there was no, he did not interact with me.
00:10:54.820
Or have any involvement, um, as far as what I was looking for and what I was able to publish.
00:10:59.760
And that being the case, what did, with your findings, what was the thing that shocked
00:11:10.180
I would, I think looking at some of the prior reporting, um, I think maybe I had a little
00:11:17.460
bit more of a, uh, of an assumption about the staff, you know, this guy, Yoel Roth and
00:11:23.580
And I have to say, um, many, many instances, as I'm reading through these internal Slack
00:11:29.420
communications and in emails, in many instances, these people at Twitter really were trying
00:11:35.220
their best to moderate content in a way that they thought was reasonable.
00:11:41.340
Now, I think they were wrong in how they performed this, but I, but there's that, you know, there's,
00:11:49.220
It's, but, you know, what's the thing about don't ascribe to, to, you know, nefariousness,
0.99
00:11:53.560
what can be, you know, actually is just sort of stupidity.
0.99
00:11:56.680
I'm not saying that they were stupid, but in the sense that I don't think this was some
1.00
00:12:00.600
big plot that, that, that these people genuinely were trying to sift out things that they believed
0.98
00:12:07.800
I think they were completely out of their depth with what they were doing.
00:12:11.620
And I think structurally, this was totally inappropriate about how they handled a whole
00:12:18.180
If so versus like on an individual level, I saw a lot of people with a lot of conversations
00:12:24.200
pushing back against what they perceived the government wanted, pushing back against what
00:12:29.560
other coworkers wanted in too many instances, bad choices were made, but it is not like there
00:12:35.440
was this thing constantly all the time where they were just, you know, taking out the red
00:12:40.000
pen against all sorts of tweets that they didn't like.
00:12:42.980
There was an internal debate that happened oftentimes.
00:12:48.260
One of them that was actually kind of both encouraging in the end result, but also revealing in the same
00:12:57.120
Jim, I forget the guy's name, but one of the lead attorneys at Twitter, there was a Trump
00:13:03.180
tweet about how he said something like, don't be afraid anymore of COVID.
00:13:09.140
And he was emailing people internally saying, does this go against our guidelines?
00:13:13.960
Shouldn't we be like, you know, labeling this as misleading or take it down?
00:13:17.640
And some of the other executives had to write back and say, optimism is not against our policy.
00:13:25.140
So it was both like good that there was that stopgap, but also like, my God, this like super
00:13:30.440
high level person actually even entertained the idea for something like that to be suppressed.
00:13:39.140
And also, I mean, I think if you remember, and most people want to forget that period of time,
00:13:45.340
because, you know, it was a bit, not a bit, it was very unpleasant for a lot of people and
00:13:51.260
One of the things I think a lot of people, particularly those of us who felt that both
00:13:56.340
government and big tech became quite authoritarian during the pandemic was this feeling that we're
00:14:04.380
not actually able to have a proper sensible discussion about something that actually really matters.
00:14:08.900
To what extent do you feel, based on what you saw, our ability to genuinely discuss one of the
00:14:18.340
most important issues in our lifetimes, actually, for a lot of people, was curtailed by what they
00:14:24.560
ended up doing at Twitter, whether it was maliciously motivated or whether it was incompetence
0.99
00:14:28.740
or stupidity, or they just thought that's the right way to do it.
0.99
00:14:32.040
How much of what happened was wrong, I guess, is what I'm asking you.
0.99
00:14:35.680
It's impossible to quantify, because I didn't do some sort of systematic, you know, analysis,
00:14:42.460
but I would characterize it, to my mind, it was profoundly wrong about what happened,
00:14:50.100
at least in the things that I researched related to COVID.
00:14:53.560
Like, what are those things that were profoundly wrong?
00:14:56.680
Right. So, the thing about Twitter is, I think a lot of us had hoped or saw social media in general
00:15:03.680
as an alternative to sort of, I guess, what's called the mainstream media, legacy media.
00:15:08.460
This is another, this is a venue for people who are shut out of these other platforms that they can
00:15:15.540
use their voice. And when I say people, it doesn't just mean regular people. I'm talking about
00:15:20.340
highly credentialed experts who just had views that were, you know, oppositional to what the
00:15:25.800
establishment wanted. And so, what I observed throughout the pandemic, and then what I was
00:15:32.520
able to basically prove, was there was a systematic suppression of content that went against the
00:15:43.900
establishment narrative, at least in America, about what was considered appropriate, and what
00:15:49.860
was considered correct. And anything, oftentimes, that went against that was labeled as misleading,
00:15:57.020
tweets were taken down, and accounts were suspended on numerous occasions. So, what I don't know is
00:16:03.820
the extent of that, because again, that would require some sort of analysis that, you know,
00:16:08.900
an individual reporter is not capable of doing. But what we do know is, I observed it as a
00:16:13.880
person deep in this for years, and what I wanted to do was kind of deconstruct, well, how did this
00:16:23.040
happen? You know, because so many times, I would see a tweet by someone like Martin Kulldorff, who's
00:16:28.620
this prominent infectious disease epidemiologist who's at Harvard Medical School, and I would see
00:16:35.560
stuff by him being labeled as misleading. I'm like, that's troubling. What's going on? So, man, when I got to
00:16:42.280
go to Twitter, I'm like, I can try to find out, well, how did that happen? What led to that? So,
00:16:47.620
I sort of wanted to work backwards to find out, because I was baffled. I'm like, what the hell is
00:16:52.500
going on? So, there were- What did you find? So, what I found, I view it as sort of like three
00:16:57.720
different, like, buckets about how they did this. There were, there was a system of bots that they set
00:17:04.320
up, this sort of like AI system where, I guess they crawl through Twitter, and they're trained,
00:17:11.400
whether it's through certain keywords or other mechanisms, they're trained to look for certain
00:17:16.340
tweets that were problematic, that the system, the algorithm viewed as problematic. So, some of the
00:17:21.580
tweets that were perfectly legitimate, perfectly appropriate, were caught up in this bot system.
00:17:28.080
If you think of like a trawling net in the ocean, you're looking for certain type of fish, but
00:17:34.580
inevitably, you're going to catch some dolphins in there. And I think this bot system probably did
00:17:39.600
catch a lot of garbage, you know, some crazy, you know, conspiratorial QAnon, you know, microchip
00:17:45.500
type of stuff. But in that process, they caught a hell of a lot of dolphins in that net, too. So,
0.89
00:17:51.820
one thing were these bots. The other thing, you had these independent contractors in places like
00:17:55.980
the Philippines, who were basically deciding the content, whether something is misleading or not.
00:18:02.760
They were given these decision trees, basically, this kind of thing where you would have a drop-down
00:18:06.740
menu. Maybe you see a tweet that has myocarditis in it. Maybe a bot would first flag the tweet. Then
00:18:12.020
it gets sent to this independent contractor, and they're like, myocarditis, click. Then there's a
00:18:15.960
drop-down menu, good or bad, click. And then, you know, there would be this system set up. But the idea
00:18:22.440
that some random guy sitting in a cube farm in the Philippines is going to adjudicate something as
00:18:29.020
complex as whether or not a tweet about myocarditis is misleading or not is insane. Like, of course,
00:18:35.340
they're not going to be able to do that. So, you had that. And then the third mechanism were people
00:18:40.180
themselves at Twitter. And there are many instances where either a particular tweet or account would be
00:18:45.000
escalated where people were looking at it, or sometimes they were just intervening, like that
00:18:49.660
example about the Trump tweet. So, you had real people doing this in Twitter, you know, at high
00:18:55.020
levels. You had these independent contractors, and then you had this algorithm. So, between those
00:19:00.200
three systems, that's how things were flagged. And, you know, we had tweets by people who were
00:19:06.800
physicians, who were, you know, licensed medical doctors, who were tweeting things from studies that
00:19:12.960
were published in peer-reviewed journals. Now, does it mean the study is well done? I don't know, but it's in a
00:19:17.620
peer-reviewed journal. Certainly not for Twitter to decide that something that's published is not.
00:19:22.680
And nevertheless, they were flagged. And in every instance, these were things that only went in one
00:19:27.760
direction. There never seemed to be something that was too pro-lockdown, pro-vaccine. That would never
00:19:34.820
be flagged, as far as I saw. But anything that questioned that, that is met by, you know, pulled up in
00:19:41.380
the net. And so, there were many instances where things that were true or that were legitimate from
00:19:47.460
published sources that were tweeted by credentialed people that were nevertheless flagged. And that,
00:19:55.240
to me, is just, like, so profoundly disturbing when you think about the information environment and,
00:20:01.860
like, what we perceive to be real or not real, what information we're all getting. And for someone
00:20:07.400
like me, who I was knowledgeable about this stuff, it was frustrating and enraging. But what I'm more
00:20:13.600
worried about are, you know, regular people who have normal jobs, they're not following this stuff
00:20:17.480
closely, and they see misleading tweet, you know, like a label put on it. I mean, that's obviously going
00:20:22.520
to have an incredible amount of influence over the public conversation and the narrative.
00:20:28.200
AI and risk-related compliance isn't tomorrow's idea, but rather today's edge. Moody's combines
00:20:34.820
advanced AI with one of the most comprehensive data estates, and automated workflows provide
00:20:39.700
faster insight, reduce bottlenecks, and drive more strategic action. Moody's helps banks,
00:20:45.040
corporates, and financial institutions navigate today's challenges and seize tomorrow's opportunities.
00:20:50.420
Stay ahead in an era of exponential risk. Visit moody's.com slash kyc slash ai dash study
00:20:58.200
Isn't the real problem here, I mean, to me there's two real problems, and I'll come
00:21:05.340
to that, the second question in a moment. But to me, this is the problem when you politicize
00:21:10.980
issues which have no basis in politics whatsoever. This is a virus released from a Chinese laboratory.
1.00
00:21:28.420
But isn't the problem when we politicize issues like this, and this is what happens?
00:21:34.280
Well, yeah, it's politicized, and it's also, at least related, you know, to COVID specifically,
00:21:41.240
there was this default that whatever the CDC says, we're going to go with them. And on one level,
00:21:48.120
I appreciate that. I get it. It's like, well, am I going to believe some random, you know,
00:21:52.640
doctor or some random, not even doctor, some random person? Or am I going to trust the CDC?
00:21:57.280
So on one level, I get that. But that's not how free speech works. And by the way,
00:22:02.340
the experts are wrong all the fucking time. And the CDC, nor a government, remember the leader of
0.98
00:22:09.200
New Zealand? I think at one point she said, I am the truth. Maybe she didn't say I, but like,
00:22:17.620
Remember that unless you hear it from us, it is not the truth. Dismiss anything else.
00:22:24.160
That is insane. That's horrible for any democracy, for anyone. So I think, you know,
00:22:37.340
it's like, you know, these are cliches, but sunlight is the best disinfectant. Like,
00:22:41.120
let the best ideas win out. Sometimes there will be bad ideas. But what I think this showed was that
00:22:47.040
our government has no confidence in regular people being able to be informed. There is this like
00:22:56.500
incredible sort of like paternalistic attitude of like, the masses are stupid, and they cannot be
1.00
00:23:04.580
trusted. So what I think ultimately happened was this idea of the noble lie, where over and over,
1.00
00:23:11.360
there would be these infectious disease experts and others within the government, where they were
00:23:17.200
saying something, whether it was about vaccines, whether it's about masks, where they were kind of
00:23:21.960
often just gilding the lily. There may have been the underlying truth, but then they would push over
00:23:26.340
the edge because they felt like they had to encourage people to quote, do the right thing. And so and I
00:23:31.880
think Twitter, most of the people who work there, I don't know now after the exodus when Elon took over,
00:23:37.020
but they were almost all of them were left leaning. And they were, you know, aligning themselves with
00:23:43.680
the Biden administration, which was aligned with the CDC. So there was this structural, you know, bias
00:23:50.540
toward a certain truth. But again, I mean, there were plenty of differences of opinion from experts
00:23:58.000
around the world on a lot of these ideas, and where different countries were doing things differently.
00:24:02.700
And within America, the idea that you're not supposed to listen to the opinion of a doctor at
00:24:11.040
Harvard Medical School, or someone at Stanford, or anyone for that matter, you have, I don't think
00:24:17.180
for obviously anyone should be able to say whatever they want in a country that values free speech. But as
00:24:23.080
far as a platform like Twitter, I don't think there should be zero constraint, because I think it will
00:24:28.820
create an environment where there's just tons of like violence and pornography and stuff that most
00:24:34.280
people probably don't want to be bombarded with that constantly. So it's not that there's no line.
00:24:39.440
But I think the line should be far, far over toward leniency toward like a wide latitude. And they were
00:24:49.220
far, far too tight on how they were controlling and policing information. And it's profoundly wrong that
00:24:58.440
they did that. And as I said, you know, it just, it altered the information environment. And
00:25:04.380
because, you know, legacy media outlets, by and large, were all on board with a certain narrative,
00:25:13.060
it was people needed a place like Twitter, where they could be exposed to different ideas. Were some
00:25:19.220
of those ideas completely nuts? Yes. But many of those ideas were absolutely legitimate. This is a
00:25:25.700
place for legitimate debate by people who understand the issues. And nevertheless, those views were
00:25:32.200
suppressed in many instances. That is very troubling. David, let's start, Francis. Yeah, I was going to
00:25:38.800
say, I have actually a great deal of empathy for big tech. And I can already see the comments,
00:25:44.640
how can you say it? But it's true, because just imagine. You're a cuck, man. Yeah, exactly. Anyway,
00:25:50.380
in the words of my ex-girlfriend. Now, but just put yourself in their shoes. So you've got this
00:25:58.100
pandemic. No one knows what the mortality rate is. We're seeing these images come from places
00:26:04.360
like Italy, where they look absolutely horrific. And then you have this incredibly powerful tool at
00:26:13.080
your fingertips, which I think it's fair to say, no one understands how powerful this thing is,
00:26:20.000
all the long-term implications of this technology. And you have to figure it out as you go along
00:26:26.780
through the pandemic. Isn't it natural that you would err towards a side of caution when dealing
00:26:32.540
with this particular matter? Because I would certainly do that. Yeah, I mean, I think as I started
00:26:39.020
out saying, I felt a lot of sympathy toward what these people were doing. And I think they were
00:26:44.980
trying their best. I think, you know, one of the things about society and institutions is that
00:26:53.520
ideally structures are put in place to help protect us from ourselves. Your instinct sounds like a
00:27:00.440
reasonable and natural one. Let's try to be cautious with like, let's try to, okay, the CDC says X,
00:27:06.440
Y, or Z. Let's stick with that. If someone's saying ABC, different from X, then maybe we should,
00:27:12.120
you know, the instinct is there and it comes from a good place, I think some people, but there needs
00:27:18.760
to be a structure in place to go against that instinct because that instinct is wrong oftentimes.
00:27:23.500
And, you know, there's just so many instances in society for, you know, all of history where the
00:27:32.620
experts think one thing and they were wrong. Right. And if anything, something like a pandemic,
00:27:38.620
we went in the opposite direction of where we should have gone. We should have, the public,
00:27:43.440
I really, and I'm writing a book about the pandemic, specifically about American schools
00:27:49.200
and what happened with them during the pandemic. But because that's part of the larger thing,
00:27:53.560
I spend a lot of time looking at it in sort of like a more panoramic view. And to me,
00:28:00.060
everything comes back to the public health authorities and that they created an environment
00:28:04.840
where a place like Twitter felt like they had to go with them because they spoke with an unearned
00:28:13.300
certainty about so many things. And as we saw over and over again, those things got walked back from
00:28:19.600
the masking guidance flip-flop to vaccines. We were told they would do one thing and then they did
00:28:24.200
something else. And the excuse of, well, the science changed. Sometimes that was true and sometimes
00:28:29.480
it wasn't. But the point isn't that you can't change your advice or guidance. The point was that
00:28:34.600
when they initially gave the guidance, when they initially told people about these things,
00:28:39.840
this was said with a strong degree of certainty. And it was not only was it politicized, but it was
00:28:45.840
moralized that you were a piece of shit. If you did not do what they said, if you didn't wear a mask,
1.00
00:28:52.120
you were a right-wing asshole. And well, the evidence on that has been pretty iffy for a long
1.00
00:28:58.420
time about mask mandates. The evidence on a whole variety of things, on school closures and all
00:29:03.300
these other things are actually not clear in the slightest bit. I'm spending, you know, a thousand
00:29:08.300
page book showing that it wasn't the way they said. So that to me, it comes back to that.
00:29:12.900
Well, yes. But I guess the reason we're asking you this question, David, is I think we're all in a
00:29:17.900
position where we're trying to understand what the true impact of this new social media technology is
00:29:23.820
on the world. And one of the things that it has done, it has shattered our shared consensus of
00:29:29.900
reality completely. And the other thing is it puts into question, you know, the two of us, and I'm
00:29:35.580
sure you as well, really feel that free speech is a crucial part of what not only is, you know, you
00:29:40.480
could say the First Amendment in America, but it's not about that. It's about the fact that the entire
00:29:44.560
Western project is based on the principle that people are allowed to express themselves.
00:29:49.260
And at the same time, I have to recognize, you know, we run a YouTube channel which has, you know,
00:29:54.540
half a million subscribers on YouTube and probably the same on a podcast. If we thought that, you know,
00:30:00.780
the comments on our YouTube videos had the potential to kill 10,000 people, we would think about what we
00:30:07.540
do differently, right? Even if our impression of that was wrong. And I think we're all in the place now
00:30:14.420
where the pandemic was really the first major test case of what we now think we're supposed to do
00:30:21.800
about the fact that we have communication technology that allows one person to, some would argue not
00:30:28.400
an insight, a riot or start, you know, we had David Icke in the UK who was talking about how COVID is
00:30:34.820
caused by 5G. And then the next day people are burning down telephone masks, right? So that I think
00:30:42.420
is probably where the good intentions of the people that you're talking about were. But at the same
00:30:47.760
time, like once you start censoring Nobel Prize winning scientists who are talking about their
00:30:52.360
field of expertise, that's, you know, way too far, right? So my point is the line is somewhere in
00:31:00.680
between and nobody knows exactly where it should be, do they? That's my view on it. The only thing that
00:31:07.360
I do know for certain is the line was drawn far, far too close towards censorship. And I think
00:31:13.720
while it shouldn't be no line at all that we're agreeing on, and, you know, something is clearly
00:31:18.720
inciting violence of something, you know, in these types of circumstances, but by and large, we need
00:31:24.680
to let ideas be free. We need to let people communicate. And again, when you have credentialed
00:31:31.740
people, but we shouldn't only say credentialed experts. There's plenty of people who didn't have
00:31:35.460
credentials who are also correct on a lot of things. So that's, it becomes, this becomes more and more
00:31:41.240
complicated. So, you know, part of being able to live in a society where there is free speech is
00:31:49.020
it's the freedom to be wrong. And that's important too. So there, it is complicated. And I think some
00:31:56.700
people probably have a highly binary, simplistic view, let everything go through. But I think that's
00:32:02.500
fairly unrealistic and also unappealing to most people again, you know, so for me, I view my role
00:32:12.780
in this doing what I did, which was, I know a lot about this topic. I have access to internal files
00:32:19.480
at Twitter. Let me actually bring this to light for people so they can see what actually happened.
00:32:25.060
These different mechanisms I was explaining by how these things were done. There's a lot more granular
00:32:28.720
details, but that in a nutshell, and then that starts a conversation. That's the start. So I,
00:32:35.420
it would be, I feel like I would be an asshole to be, you know, pronounce, making pronouncements
0.98
00:32:40.360
about, well, here's how everything should run now. You know, so I don't even feel comfortable saying
0.98
00:32:44.340
that. You can tell, you know, we're all sort of saying, we know what happened was wrong. We know
00:32:48.920
the line should be somewhere else, or maybe, maybe that's not even the right metaphor. Maybe it's not a
00:32:53.840
line, but it's like a zigzag. Who knows what? It's a levy over the, we can come up with different
00:32:58.220
metaphors, but we don't know exactly. But for my role was, let me just report on this and let's start
00:33:04.460
a conversation. And the idea that, that this is quote, like a nothing burger, as all these people were
00:33:11.280
saying, is so idiotic to me and so like phony, like bogus, naive, like everything about that was
0.99
00:33:22.220
wrong. I mean, I, I know that the last time I looked like my thread that I did on Twitter for the,
0.99
00:33:27.840
for my Twitter files that I did. I mean, it had more than 60 million views. And as far as I'm aware,
00:33:36.320
no one covered it in any sort of like legacy media. That was shocking to me. And so my view on this is
00:33:44.840
even if, and they would be wrong to say this, but even if someone said, well, what was reported on
00:33:51.020
here isn't important. Even if they thought that the mere fact that something had 60 million views
00:33:57.960
on it, hundreds of thousands of likes or whatever the, you know, whatever the statistics were that in
00:34:03.300
and of itself makes it newsworthy. I mean, that is the definition of news. When something is of great
00:34:09.860
interest to an enormous number of people, particularly on a topic like this, that is news in
00:34:15.880
and of itself, even if they felt like the content wasn't important. So that is something that as a
00:34:22.120
society or that people in media, I think need to reckon with. It is just a remarkable comment on like
00:34:31.680
the environment we're in today, the culture, that something, a thread, you know, and this isn't
00:34:37.220
about me. I mean, Matt Taibbi is reporting that their stuff was far more in depth than mine,
00:34:42.100
hasn't been covered. How is it possible that this stuff was basically ignored? That is remarkable.
00:34:48.860
I'll give you an example. We had it in the UK and I mentioned it at the time. The BBC on the day
00:34:54.060
that the first several instances of the Twitter files were released. They had a story about Elon
00:35:00.420
Musk firing cleaners who happened to be two women of colour. And they had a story, which was, it wasn't
00:35:07.340
a story, it was like a piece, which is who is the billionaire who's running Twitter, Elon Musk.
00:35:12.460
So they obviously felt that Twitter and Elon Musk were important and significant, but they didn't cover
00:35:17.360
the Twitter files at all. And that, to me, that was very telling. But the other thing I wanted to ask
00:35:22.120
you about, and this is obviously something that a lot of people felt on along political lines prior
00:35:27.860
to the pandemic. And then during the pandemic, it, I think, strengthened how people felt in this way
00:35:34.480
about it, which was, you know, Twitter is a company headquartered in California. We all know that
00:35:41.140
California, the values of people in California are not necessarily representative of the entire country
00:35:46.420
or indeed the West more broadly, which is where this conversation was playing itself out.
00:35:50.560
Now, did you find, did, as you were reading this, did you feel that this was a group of people who
00:35:56.380
were almost captured by a particular worldview or just forget captured, that's a loaded term,
00:36:01.920
who held a particular worldview? And as a result of that, they were making decisions that were not
00:36:07.340
necessarily reflective of the worldview shared across the country.
00:36:10.520
A hundred percent. There were exceptions. And like I said, I saw a lot of robust exchanges within
00:36:19.800
Twitter amongst employees, but by and large, absolutely. I mean, again, the fact that they
00:36:25.300
went with, well, here's what the CDC says, that's the truth. I mean, that's crazy that you can't
00:36:32.540
function in a society where a government agency is the truth and everything else is considered
00:36:39.280
misleading or disinformation. And all these people, not all, most of them, and this is a known thing,
00:36:49.460
are left-leaning, for lack of a better term. And because the pandemic in America was so politicized,
00:36:59.480
you were on one team or another. I found myself, you know, sort of floating in space. I had no incentive
00:37:07.220
whatsoever to align myself with Republicans necessarily. For years, I had been a Democrat
00:37:13.520
loosely. I mean, I have a complex, as you guys probably do too, a complex range of opinions on
00:37:21.400
a complex range of topics. So I don't align myself- You think for yourself, shocking.
00:37:26.560
I try, you know. Which makes you far right. Yeah, far right.
00:37:29.460
Exactly. That means I am a right-wing. I mean, I did an interview with, I forget who it was,
00:37:37.140
but someone who's like center-right. And I sent it to a group, this was, I don't know,
00:37:41.560
a year ago. And I texted with, I'm like a group chat with like a bunch of old friends of mine.
00:37:46.420
We've known each other for 20 or more years. And one of them wrote back to me and said,
00:37:50.160
are you in the Proud Boys now? You know, that like white supremacist group. And I was like,
00:37:55.020
come on, man. You know? And he was like, half joking? Jew and the Proud Boys. But anyway,
0.94
00:38:01.360
it was, point being, it's like, if you questioned any of these things, you automatically were thrown
00:38:09.560
into that camp on the other team. And, you know, so for me, as, I think I'm considered one of the
00:38:18.520
only journalists or maybe the most prominent one who was writing for a lot of legacy media
00:38:25.780
publications that tend to be left-leaning. But I was writing from a very much sort of contrarian
00:38:33.760
viewpoint, not purposefully. I wasn't like, I'm going to be the contrarian. It was, that's where
00:38:37.760
the facts took me. And that's what launched me on this early on in the spring of 2020, when this all
00:38:43.860
began, that I saw everything that was happening. I, like most of my neighbors and everyone around me,
00:38:49.460
oh my God, there's this crazy virus. Let's, let's like, you know, lock the door, everyone.
00:38:54.640
But very quickly, because of my professional background as someone, I like reading scientific
00:39:00.860
journals. I read academic journals. I talk to scholars and research people all the time.
00:39:07.280
So I have that sort of professional background in doing that. And also for better and for worse,
00:39:11.920
my temperament, I'm always poking holes in things. I always see kind of, I'm always skeptical about
00:39:18.640
things. And I very quickly started, well, let me just look and see what the actual evidence is for
00:39:24.160
this. And very quickly, it took me in a direction that was different from the direction that I was
00:39:31.160
supposed to be in. And all of a sudden I found myself like, you know, cast out of the boat, cast out
00:39:38.720
of the, of the, of the thing with everyone else that I was quote, supposed to be with, even though
00:39:43.560
there was no zero political incentive for this. And so, but nevertheless, I was branded that way.
00:39:49.700
You know, I'm, I'm a right winger. What are you fucking talking about? You know, I was a Bernie
0.99
00:39:53.720
supporter years ago. So this, this incredibly infantile binary approach to life is just so
00:40:04.220
tiresome and problematic. But what I've come to believe is that I think that's just how,
00:40:11.860
and this is perhaps a banal observation, but that's just how society operates. Most people
00:40:16.820
want to be in a certain tribe and their default is like in group, out group. And so I spend a lot of
00:40:25.220
time writing about this and digging into this in my book that I'm working on about this. I just,
00:40:29.560
it's, it's, it's incredibly fascinating to me. I, you know, what I know about both of you guys,
00:40:34.260
I think we're in the same, we are in our own little, it's like the Venn diagram. We're in like
00:40:38.300
this tiny, I don't know how, how large is our group? I don't know. Not that we all share the same
00:40:43.180
opinion, but our group of people who are sort of like not, I think there's something, do you think
00:40:47.900
there's something in your personalities? Cause I think I wonder this about myself. Is there something
00:40:52.380
about us and people like us that we seem to have less allegiance to groups than it seems other
00:41:00.120
people do? Do you think there's something like, how did we end up where we ended up?
00:41:04.300
For me, it's quite easy because growing up in the Soviet union, I kind of had a very early
00:41:08.820
experience of like, just cause everyone says, everyone goes along with something doesn't mean
00:41:13.600
it's true. And then it was replicated more generally. I mean, if you, if you're an intelligent
00:41:21.160
kid, you look around at the adults and you realize these people don't have a fucking clue
0.99
00:41:24.660
what's going on. Right. So that from a young age, I realized I'm going to have to think about
0.99
00:41:31.100
things for myself, you know, but I actually, you know, I think it can be very dispiriting
00:41:36.740
because online communication makes it feel like we are one of a handful. And then there's
00:41:43.760
the crazy people on the right and the crazy people in the left. But I think what we forget
0.82
00:41:48.540
and look, the success of our show is kind of reinforcement of this is there's a hell of
00:41:53.840
a lot of people in the middle who are never going to leave a YouTube comment, who don't
00:41:57.920
have a Twitter account, who just want to hear something because they are in the middle.
00:42:03.100
They're trying to make up their mind. They're not hard line. And because of that, you don't
00:42:08.600
see them. They don't show up. Right. And there are crucial points in the sort of trajectory of
00:42:14.540
society when, yes, you have to go in the voting booth and you have to pick red or blue.
00:42:19.340
But most of the time, I think most people aren't really that welded to those political ideologies.
00:42:26.980
And you don't see them as much because they're not on Twitter screaming because they're trying
00:42:30.560
to get their kid to school and put food on the table and whatever. So at least our hope,
00:42:35.720
certainly mine anyway, is we've always tried to speak to those people and stay away from
00:42:42.220
the crazies on both sides who are, you're right, at this point, shaping the debate.
00:42:48.540
I think you put this far better than I did. I think you're right.
00:42:59.140
You're right because I observed so many times. But I guess what worries me about that a little
00:43:04.220
bit is, and again, maybe this is just human nature or society, most people are either
00:43:10.040
disinterested or disinclined to speak up. So I think you're both correct that there is a larger
00:43:17.260
number of people like us than there are. But because most people, for whatever combination of
00:43:23.500
reasons, don't say, and I've witnessed this myself in my small town that I live in with my kids when
00:43:30.040
the schools were closed, there was this kind of very loud group of people on the local
00:43:35.840
parents' Facebook group, you know, and I was called a murderer for saying that I thought schools
00:43:40.560
should open, even though kids were in school in Europe at this point. Why do you want to murder
00:43:45.680
children? I'll never forget that. Someone responded to me that way. And I'm like, you're aware there
0.97
00:43:50.880
are millions of children in school elsewhere and they're all, they're fine. I think there, are there
00:43:55.240
still kids in Europe? They're still there. Just click. So that's the thing that there's sort of
00:44:02.300
this dual dynamic. There's both a lot of us, perhaps, but for whatever reason, and maybe a lot
00:44:08.180
of those reasons are good. Sometimes it makes sense why someone can't speak out for any number,
00:44:13.240
maybe their job, they're afraid. It's regrettable. I wish more people were a little more vocal or
00:44:20.240
involved then if they didn't, because the forces particularly, and I guess I have, believe me,
00:44:27.620
I have no shortage of condemnation for people on the far right, but I'm far more sensitive about the
00:44:35.800
left for two reasons. One is that I think I used to associate myself with the left for most of my
00:44:41.380
adult life. So I have to have like a greater sense of betrayal, I feel, in the way I feel like some of
00:44:48.440
these people behave during the pandemic. So that's number one. Number two, by and large, the left
00:44:54.340
controls the major levers of society. Yes, the politicians in Washington or in different state
00:45:01.780
houses or whatever change, you know, at each election cycle, but technology, the movie industry,
00:45:09.920
media, publishing, fashion, so many major pillars of our society, the cultural sort of cornerstones
00:45:21.440
are far dominated by the left. So that's why I have like a particular sensitivity of when the left,
00:45:30.040
by and large, moves in a certain direction, why that really needs to be kept in check so much.
00:45:36.480
Let's dive into the media a second, because we're talking about journalism here. And if you think
00:45:44.060
about movies, there was always the archetype of the intrepid journalist going where others feared
00:45:50.000
to tread in order to get the story. For example, if you look at superheroes, I don't think it's an
00:45:54.640
accident that Peter Parker, Spider-Man, is a photojournalist and Superman is a journalist.
00:45:59.260
Right. You know, they were the ones challenging authority. When did that stop? When did we start
00:46:07.580
seeing a journalist as someone to basically uphold what the government's saying?
00:46:15.100
One of the things that like blew my mind in the beginning of the pandemic and as what went on was
00:46:20.920
that I kept thinking and I've written a number of pieces that I believe it's fair to say that I kind
00:46:29.700
of blew something open in the sort of legacy media that that wasn't being talked about, whether I was
00:46:35.600
wrote very early on the very beginning of May saying we should be opening schools. Here's why.
00:46:42.040
Here's a long compendium of data to show why this is a good idea.
00:46:47.500
There's this, I don't think this was common in Britain or elsewhere, but we had this hybrid
00:46:53.620
schooling model here where kids go to school for one or two days a week and they stayed home and they
00:46:57.860
had that. And I was the first person to write a big piece on this and I had a number of experts in
00:47:04.940
the article saying like, this is completely idiotic for multiple reasons. There's no evidence that this
0.99
00:47:11.160
is going to reduce transmission. So I wrote that. There was a piece that I wrote questioning the CDC had
00:47:16.500
guidance. I'm going to get to my point in one second. The CDC had guidance before the summer
00:47:23.100
camp guidance where they wanted children to wear masks outdoors. And when I saw that the guidance
00:47:28.080
came out, I'm like, this is nuts. So I immediately, I reached out to some really prominent people in
00:47:33.040
the field. Dimitri Christakis is the editor-in-chief of JAMA Pediatrics, you know, the foremost pediatric
00:47:39.140
medical journal, immunologist at Columbia University Hospital, these people. And I reached out to them,
00:47:44.600
and I said, am I crazy or is this like completely insane? And, and they were like, oh, this is,
00:47:50.140
this is horrible. And, and, but, and, and I had a handful of other pieces on myocarditis. I interviewed
00:47:55.780
the guy, the Israeli scientist who wrote, basically released the first major report. I don't even know
00:48:01.320
how I got ahold of him, but I'm, there I am texting with this guy on WhatsApp and we're, you know,
00:48:05.680
put, so there were a bunch of instances where I wrote about things that I hadn't seen exposed yet.
00:48:11.220
Also the hospitalizations, sorry, now I'm just, take your time. This is what we do. So, okay.
00:48:16.660
Don't rush, take your time. I mentioned this not as an aggrandizement, but to a larger point than
00:48:21.540
I'm about to make. David is brilliant. Yes. I mean, he's so fucking good.
00:48:24.800
So, but there, the hot people have been murmuring for quite a while that, you know, maybe this
0.96
00:48:31.100
hospitalization numbers aren't actually reflective of people who are in the hospital. It rather,
00:48:36.560
it's with COVID rather than from COVID. Now this also ties into our conversation about the suppression
00:48:43.440
and censorship on Twitter and elsewhere. For a long time, that was considered a conspiracy theory
00:48:47.880
and you're crazy. Well, it turns out that's true. And, and, and I reported on these two studies that
0.94
00:48:54.840
were done on pediatric admissions. And at the time, at least 40% of the pediatric COVID admissions
00:49:02.280
were unrelated to COVID. A kid, maybe they broke their foot or something and they had to test them
00:49:07.680
for COVID and they just incidentally had COVID, but it had nothing to do with why they were in the
00:49:12.300
hospital. But nevertheless, that was added to the tally. Just very quickly, I had, I had a joke that I
00:49:18.180
tweeted during the pandemic. I said, my, my grandfather died with his wife by his side, but it was the
00:49:23.700
middle of the pandemic. So they decided he died of her and put her in prison anyway. Exactly. It was absurd.
00:49:29.800
So, but if you, but people said this were conspiracy theorists, you're some asshole, but it was true.
0.99
00:49:35.900
So anyway, so the reason I mentioned all of these things, the school closures, the hybrid shit, the
1.00
00:49:39.920
myocarditis, and then a handful of other things over and over, I kept thinking, I need to like get
0.95
00:49:47.580
this thing out. I got part of being a journalist. I generally like a slow moving on less in the news
00:49:53.520
cycle, but because I found myself in this, you want to be first. I'm like, this is exciting. And each time
00:49:58.500
I'd be in a panic because I turn an article at first, I pitch an article to, to an editor. Then
00:50:04.280
if they accept it, then there's, you know, then I'm writing it, I'm researching it. The editors go
00:50:08.920
back. It could take a week or two weeks or longer before you finally put the thing out, depending how
00:50:13.260
in depth it is. And I, every, I just be nervous waiting that someone else is going to scoop me.
00:50:19.000
And no one did any of those times. And slowly I started seeing this pattern. And that's the thing.
00:50:25.440
This is a very long response to your, to your prompt about where's the intrepid reporter. And
00:50:30.740
I'm not saying I am the embodiment of that, but what I would say is there was an absence of that
00:50:36.940
elsewhere. When I wrote that thing about the summer camp guidance, I think a day or two before my piece
00:50:43.760
came out, there was a thing in the New York times on it. And it was like the nukes, you know, CDC guidance
00:50:48.820
for camps is out. Here's how to keep your kids safe. And it just unquestioningly, they had a one
00:50:54.780
or two, you know, doctors who talked about it. No one questioned the, the, the validity of any of
00:51:01.320
these recommendations, including making a kid wear a mask while they're going to be playing tennis with
00:51:05.360
someone, you know, 30 feet away on the other side of a net. No, it wasn't questioned. So it's this idea
00:51:11.740
of like, is a journalist just like a megaphone? You know, maybe in some instances, instances that's
00:51:18.540
appropriate, but by and large, that's what these people were doing. And I was kind of astonished
00:51:23.180
that over and over, it seemed like these large media operations with millions or hundreds of
00:51:31.540
millions of dollars behind them. These are huge institutions with large staffs of people and
00:51:37.200
editors and this whole operation. And they weren't questioning the guidance. They weren't questioning
00:51:44.120
what was said. They were simply reporting it. And they had the same crew of experts oftentimes. And
00:51:50.480
I use experts in quotes because oftentimes this is someone like there was a particular emergency room
00:51:54.880
physician who, who I think was at Brown at the time, who was quoted constantly in the New York
00:52:01.720
Times and CNN and these other places. This person had no expertise on infectious diseases,
00:52:06.600
had no expertise on mitigation measures and their effectiveness. Nevertheless, this person was
00:52:13.440
quoted. They're on like the speed dial at a place like the New York Times. So you had no questions
00:52:20.480
asked of Fauci or CDC. And then when there was sort of questions asked, it was almost always from the same
00:52:27.020
pool of, quote, experts who were basically just supporting what was being said anyway. So that was the
00:52:35.520
weird thing for me. I just was, now when I look back, it's funny, but you can imagine me, I was so
00:52:39.880
nervous I was going to get scooped on each of these things. Someone surely is going to write a piece in
00:52:45.900
some major place about like, why our schools aren't open here when they're open everywhere else. And I'm
00:52:50.720
like, no one wrote that. Someone surely is going to write about the hospitalizations. Someone sure, all these
00:52:56.160
things. And it wasn't happening. So it's just been, it was just a bizarre experience that I will never
00:53:04.280
get over. Like it's permanently altered me. Seeing this, you know, from my own experience as a writer
00:53:13.520
And it's also has repercussions right the way down the line, because we've all agreed that the media
00:53:20.160
weren't being honest about COVID. The mainstream media weren't being honest. I think that's fair
00:53:25.380
to say. And then they report on Ukraine and everybody goes, well, how can I trust you? And
00:53:31.000
then you get a large portion of people coming out and going, well, they weren't honest about COVID,
00:53:36.840
AKA, they're not going to be honest about this. I don't believe it. And then what you have is a
00:53:42.880
fundamental breakdown between media and the people. And that's a really dangerous place to be in.
00:53:49.480
I mean, I think there's a couple things there. Journalism, in my view, should be adversarial
00:53:56.080
to authority. And for whatever complex range of reasons that, you know, we could go into,
00:54:06.080
but it's again, writing a thousand pages on the why, but for a long range of reasons, that sort of
00:54:12.760
skepticism of institutions was absent. It vaporized.
00:54:21.340
Right. Exactly. I mean, these are the, these are the things that traditionally-
00:54:25.960
People on the left, they used to, the left traditionally was like-
00:54:30.780
Yeah, exactly. It was like highly skeptical of these places, like big pharma, the government,
00:54:35.960
you know, these, but because, I think because it's so politicized, they then just threw their hat in
00:54:40.800
with this group. That's one piece of it. I don't think it's the only piece that that,
00:54:46.020
and then the public ends up not trusting it because they, or at least a large segment of the public.
00:54:51.780
And I think one of the things that's funny, not to get too far afield here, but, and if I'm not
00:54:58.280
already canceled, maybe this will cancel me, but like-
00:55:01.060
But here we go, is that there's so much talk about diversity, which is important, you know,
00:55:06.780
racial diversity and other types, but we don't see, and again, this is an observation many people
00:55:13.300
think, there's not an ideological diversity in a lot of these institutions. And like most of the people
00:55:19.500
who work at these, you know, at the Times or, you know, these major TV networks, these are, I mean,
00:55:26.000
I know a lot of these people personally, and I know of, they all went to Yale and Brown and Columbia
00:55:31.240
and Harvard. They all went to the same places. They're all there, inside there. Imagine if the
00:55:36.320
Times, instead of their diversity lens, instead, if it was trained on, well, let's, let's hire some
00:55:41.960
people who just went to some community college. Let's go and, let's get someone who grew up in,
00:55:46.580
in West Virginia and maybe didn't go to college at all, but it's like real smart and has some sort
00:55:51.680
of grit. Imagine how different the coverage would be of all sorts of issues, whether it's related to
00:55:57.060
the pandemic or otherwise. So there's certain types of diversity, but you're missing out on,
00:56:03.260
on working class people. You know, journalism used to be more of a working class kind of profession
00:56:08.580
at one point in certain regards, you know, you know, a number of generations ago. Now it's,
00:56:13.020
it's far more, um, I think elitist. So these people, and, and I think that there's a wonderful
00:56:19.760
article, I think it's Thomas Frank from years ago in Harper's. Um, and he wrote about how, um,
00:56:26.100
Bernie Sanders and he was running basically was murdered by the, his, his campaign, um, by the
00:56:34.160
mainstream media. And, and he roughly says this, and this is some of my own sort of projection onto
00:56:39.980
his thesis. But in order to get into a place like Yale or Brown, how do you get in there? Well,
00:56:46.400
you're a certain type of person who follows rules. Maybe some people get in there who are
00:56:50.960
iconoclasts. They're just brilliant. But most of the people you're getting in there because you're
00:56:54.420
an apple polisher, you're the worker B, you know how to get straight A's, you know, 4.0 grade average,
00:57:00.560
you're a perfect student. And then you get in there and you're, you know how to work your way within
00:57:05.100
an institution and network and you get perfect grades there. And then maybe you go to Columbia
00:57:08.780
journalism school and you follow this path. So there, the way that these people became
00:57:14.500
successful was through a certain path. That's their life. That's their lived experience is,
00:57:20.880
is doing this. So it makes sense. So then they saw Bernie as this outsider. How could this guy win
00:57:27.740
or who dare he? You know, how should he do this? You know, he's not one of us. And I think,
00:57:33.720
and I think we also see that in coverage of a whole range of issues, including with the
00:57:38.660
pandemic, that if you have all these people who all know each other, they all have the same
00:57:44.480
kind of background, the same way that they became successful, that to veer from that,
00:57:51.680
I'm not even saying this is a conscious thing. I think it's just like, it makes sense. This is just
00:57:55.660
how they view things. So that, that, that's my, that's my thesis on, on that.
00:58:01.960
And we haven't even reached, as far as I'm concerned, the real punchline of the pandemic,
00:58:06.400
which is the effect on children's education and attainment. And also as well, what that is going
00:58:12.220
to be doing to the ever widening gap between attainment, between rich kids and poor kids.
00:58:18.020
Right. Well, you know, there's, um, my focus on reporting has not been on what people will call
00:58:28.540
sort of like victim porn on the horrible things that happen to children. Um, because to me,
00:58:34.340
it's just like, so self-evident that this was absurd. Um, but there's so much just obvious
00:58:41.840
evidence from this. This was obvious before it even happened. And now we have also from the academic
00:58:47.420
harms, um, you know, from kids. And we know that children who are in less privileged backgrounds
00:58:54.000
were going to do, we knew this was going to happen. I wrote a piece for the New York times,
00:58:58.400
actually early on in the pandemic about what they call these pod schools. They probably didn't have
00:59:02.480
this in the UK where, um, you would pay what wealthy parents would pay generally wealthy.
00:59:07.960
You know, it could be $20,000 where a group of parents got together and they hired a teacher
00:59:12.120
and they had a pod of like, it could be five kids, 10 kids where they, because the schools were closed.
00:59:17.420
Um, so those kids who had the money, they had the pod or they had tutors. They have, you know,
00:59:23.820
they can get all this extra help, but some kid who's living in, you know, a tiny apartment in
00:59:28.980
the Bronx, um, with, with eight other people or whatever the circumstance may be. And there are
00:59:33.920
those kids, um, cause I've spoken to them and I've spoken to their parents and their teachers.
0.99
00:59:39.060
They, they were never going to learn jack shit. Um, they didn't have an internet connection.
0.98
00:59:44.600
There were kids who are sitting in a parking lot of the Taco Bell so they could have a wifi
0.94
00:59:49.700
connection to try to get on to some bullshit remote learning program. That is nuts. So for any,
0.99
00:59:55.980
so the idea, because I was branded a racist, as were many other people who advocated for schools
01:00:01.020
to open, they was considered, they said it's racist to want schools to open. It's white supremacist.
01:00:06.280
But the, the, the horrible irony of course, is that the greatest harms were, we knew this from,
01:00:13.460
this could be seen very easily from the beginning and it was obvious while it was happening. The
01:00:17.520
greatest harms were born by the kids who were Brian, uh, disproportionately black or brown, um,
0.99
01:00:25.320
and certainly moving away from race, just who had less resources, regardless of what your race was.
01:00:30.140
We knew the people who have the money, they're always going to be okay. Most of them. I mean,
01:00:34.540
those kids suffer too. Make no mistake. A lot of them, everyone's different. And it's not
01:00:38.500
automatically you're going to do horribly if you don't have money, but that's the broad trend.
01:00:42.560
And this was of course obvious. So school is more than just a place for learning math or something,
01:00:49.900
and particularly for younger kids. In America, school has a role of, of a societal function where
01:00:57.920
there are people, by the way, were dismissed and maligned for saying, Oh, you, you just want a
01:01:04.040
babysitter. It's a bunch of parents just want babysitters for their kids. Well, that's, that
01:01:09.500
should not be put down. Like care for children is really important. And that's a wonderful function
01:01:16.120
of school. That's not something that should have been like dismissed as like, you're an asshole for
1.00
01:01:21.040
wanting your, you know, six year old to go to a place while you're going to work. Like, but somehow these
0.99
01:01:27.140
people were branded as selfish. It's the sort of thing you say when you've got your own babysitter
01:01:31.380
that you can pay for. That's when you can say it. And that's where it comes from. But I know
01:01:36.960
you've got a book about this coming out. So tell everybody when it's coming out and what it's going
01:01:41.020
to be called. Well, it's called an abundance of caution. At least in the States, I feel like they
01:01:48.160
did in the UK too. This phrase was constantly used out of an abundance of caution, dot, dot, dot. And I
01:01:54.860
remember when they closed my kids' schools that the email was sent out and there it was
01:02:00.000
out of an abundance of caution, we need to close schools for a deep cleaning. And they
01:02:04.480
said, you know, meant nothing. But, um, so that phrase always stuck in my head because
01:02:09.680
the question is, well, caution in what direction? Um, and what, and, and, and what we, what I
01:02:19.700
spend a lot of time writing about, and God knows when the book's going to come out because
01:02:23.000
I'm still working on it now. It's probably not for a year or so. I saw that when I asked
01:02:26.240
David, there was that little pang. I know that I've written the book. It's like.
01:02:29.820
The pain. Every night, like my son, he's like, dad, are you still working on that book? I'm
01:02:34.040
like, yes. Stop asking me. I'd like, my like temple is pulsing. Yeah. But, um, but you think
01:02:41.340
about a lot of these interventions, including school closures. I think there's good evidence
01:02:47.120
that the most important word that is not talked about enough about relation to the pandemic and
01:02:53.380
these interventions is time. And a lot of things can be effective over a very short period of time.
01:02:59.420
If the physician puts on a very tightly fitted N95 respirator and sees a patient for a 10 minute
01:03:06.200
checkup in a room, that mask may very well be protective to some extent. Um, but that's different
01:03:12.880
from wearing some bullshit cloth mask in a classroom with 20 other kids for eight hours.
0.96
01:03:18.340
And similarly, you know, pulling the master switch and keeping everyone home for a week or two weeks
0.89
01:03:23.640
or not everyone, but the people who can stay home, that can definitely have an effect. I think there's
01:03:28.320
good evidence. The problem is a pandemic's a marathon. It's not going to end and, you know,
01:03:34.400
quickly. And over time, we know that people's ability to comply with things wanes. You just
01:03:41.940
wearing a mask is fucking annoying for most people. Um, it's not natural. And whether you
0.99
01:03:47.360
purposely don't want to do it, or you're just like, even subconsciously, you're tugging at it.
01:03:51.220
People are rubbing their eyes. They've done studies on this. You can see people constantly touching
01:03:54.420
their face or whether it's something like school closures, people are going to spread their ability.
01:03:59.560
And there's Google mobility data for this too. You could see people just started moving around
01:04:04.820
more and more as time went on. And you saw this happen even in places where the restrictions
01:04:09.800
were still in place, that that didn't stop it, that you can't like hammer people with
01:04:15.520
this stuff. Human beings, most of them are not comfortable being completely isolated from
01:04:21.100
each other. They can't do it. And so what happens is this, this confusion about, well,
01:04:27.940
this works or that doesn't work. Of course, if people don't go to school, there'll be less
01:04:31.600
transmission. Well, yeah, maybe for a week or two, but over a month or two months or three
01:04:36.700
months. And that's the main misconception I think that a lot of people have about these
01:04:41.120
interventions. That yes, I'm not someone who says masks don't work or this quote doesn't
01:04:47.220
work. That's too simplistic of a way of framing these things. It's some things may work to some
01:04:53.840
extent in this particular circumstance for a certain period of time. It's not a sexy answer.
01:04:59.460
It's not as easy to just tweet that, but that's the truth. You need all these little qualifiers.
01:05:04.740
So we knew that over time, this stuff was simply not going to be effective. And that's what I
01:05:10.040
believe the evidence shows. Well, I was going to wrap up there, but just maybe one more thing on
01:05:15.040
that. I think one of the things that you touched on there, but I think we could expand on a little
01:05:20.500
bit is that too many people forget that safety isn't the only thing that we are ever pursuing.
01:05:27.020
I mean, freedom for this country would not exist if people didn't get together who valued something
01:05:32.580
other than safety above safety, right? The United States was created by a bunch of people who put
01:05:38.540
themselves in harm's way. That is the very definition of being the opposite of safety for
01:05:43.400
something else. That something else is called freedom in their case, right? So one of the reasons
01:05:49.060
I think we're stuck in the place that we're stuck in is that people have forgotten that safety is not
01:05:54.820
the only value. It's not the only thing that we care about and we cannot care about. Otherwise,
01:06:00.380
we'd never leave the house. And I would say two things. One, agree with you a hundred percent on
01:06:05.240
that, that, you know, there's a reason why we, right. People go swimming, even though a certain
01:06:09.060
number of people drown every year swimming, we get in a car, we do everything. If you, the second you
01:06:14.020
step out of your house, there's some degree of danger, depending on what you're going to do,
01:06:16.960
but that's part of life, including we also endanger other people because the argument against what
01:06:21.860
you're saying, people say, well, that's different. That's you just endangering yourself. But with the
01:06:26.360
virus, you're endangering other people. Well, guess what? Every time you get in a car, you're
01:06:30.160
endangering other people too, but we all choose to partake in that. And we don't force everyone to
01:06:34.780
drive 10 miles an hour. We laugh. So there's a whole range of things we do where we are endangering
01:06:40.120
each other. And we accept that as a society because we value these other things for human flourishing
01:06:45.420
beyond just putting ourselves in bubble wrap. But the other thing is, staying home is not safe
01:06:51.500
either over time. And that's the other flip side, again, about this idea of caution that it's like,
01:06:56.180
well, there are people, and I wrote this piece about this church that got in a huge lawsuit with
01:07:01.360
a county in California where they were barred, all churches there were barred from having anyone
01:07:08.280
gather indoors for many, many, many months. And so this church ignored the God.
01:07:15.420
People got together anyway, and they accrued an extraordinary number of fines. It's an amazing
01:07:21.960
story. I mean, they were spied on. They were special health officers peering at them through
01:07:27.520
a chain-link fence. They were going into the church and monitoring private activities. It's an insane
01:07:33.540
story. But the message from it, though, to your point, is a lot of the people I interviewed,
01:07:41.100
these were people from the church, church members. These were people, and by the way,
01:07:45.120
I don't attend church. I had no skin in the game about support. But these are people suffering from
01:07:49.620
addiction, elderly people who were profoundly lonely. There was a guy who, I think there were more than
01:07:56.160
one, I spoke to one myself, who wanted to commit suicide because he was so profoundly lonely. He had
01:08:02.440
gone through a breakup. And for these people, church was their support system. Is it mine? No.
01:08:08.260
But I have to understand and respect for them. They needed this. So the idea that they're safe
01:08:14.840
because they're kept home was bullshit. These people, some of them were literally at risk of
0.99
01:08:20.380
dying themselves. And we know that there is epigenetics with how your environment affects
01:08:26.380
your immune system. Telling a bunch of people you need to be isolated, particularly elderly people,
01:08:31.720
oh, you're in an old age home, and your family can't visit you. Like, of course, that has a profound
01:08:37.600
effect on people's underlying health. So we're seeing this, and it's very complicated, but with
01:08:42.460
excess death rates and all sorts of other measures about what makes for a good life. So, you know, to
01:08:48.560
your point, this sort of like myopic focus on a virus and not looking at all these other aspects of what
01:08:57.520
makes for a flourishing life, again, maybe that made sense for a couple weeks in the beginning.
01:09:03.240
But very quickly, that, you know, the ratio changes about what is reasonable, both from a public
01:09:09.340
health standpoint and also a mental health standpoint. So that's how I view it.
01:09:15.400
David, it's been a fantastic interview. Thank you so much for coming on the show.
01:09:19.540
So the final question we always ask all our guests is, are you vaccinated? No, I'm joking.
01:09:28.400
Because if you're not, you're killing people. No, you're a murderer.
0.99
01:09:30.900
You're a murderer and murdering kids. Anyway, it's what's the one thing we're not talking about
1.00
01:09:39.740
I would say, just moving away from this topic, one thing that has long interested me
01:09:48.540
is everyone talks about climate change as this macro concern. And I'm not even going to touch on
01:09:56.860
the validity of this. But one thing that I think is really important for people who care about the
01:10:02.160
environment, and I'm someone who's long cared about the environment very much, is climate change is a
01:10:08.860
very abstract idea. But there are very concrete things that are happening to our planet and to
01:10:14.900
animals that we actually can stop. Because it seems like we failed at climate change. Every year,
01:10:20.080
they have the meetings. They go to Davos or Kyoto or wherever else, and they talk about it. But
01:10:30.960
Right, exactly. But the thing we can do is, there are an enormous number of sharks that are killed
01:10:38.920
every year so people can have soup with their fin. Are you fucking kidding me? And they slice the fin
0.98
01:10:45.460
off this majestic animal, and they just toss them back in the ocean. And they're left there to just
01:10:51.600
slowly bleed out and die. And they don't even use the rest of the animal. So shark fin soup, where you
01:10:58.560
think about, there are still multiple countries that do commercial whaling. I think, I might be wrong,
01:11:04.360
I think it's like Iceland, Norway, Japan. That is insane, to my view. And palm oil. I'll stop here.
01:11:14.240
You know, most of the garbage that we eat, you know, all these processed foods, they have palm oil in
01:11:19.340
them. That is a massive source of deforestation. They burn or chop down incredibly important and rich
01:11:27.380
rainforests or other really old-growth forests and put up a bunch of, you know, palm trees for making palm
01:11:33.900
oil. That is horrible for the species that used to live there, who are now dead, because they don't
01:11:38.200
have a home. Horrible for greenhouse. So those are the things that I think get lost. Once it's, you know,
01:11:43.420
once again, it's easy to just be like, we need to help stop climate change. Well, guess what?
01:11:47.540
That messaging failed. It doesn't seem to work, because we still keep producing all this stuff.
01:11:51.960
So instead, I would love for people to focus more on how do we save specific animals? How do we stop
01:11:58.920
specific practices that, to my mind, are completely unacceptable? And in 50 years from now, we will
01:12:05.460
be embarrassed that these things were happening, from factory farms to the sharp fins, etc. That's
01:12:11.060
what people should talk about. Absolutely, David. Thank you so much for coming on the show. Really
01:12:14.820
looking forward to reading your upcoming book. Of course, you have got this one as well, which we
01:12:18.880
didn't get to talk about. Buy this book, everybody. Buy this book, everybody. And we're going to head over
01:12:23.500
to Locals for your bonus questions. So join us there. All the links for David's Worker in the
01:12:28.600
description. Take care. We'll see you on Locals very shortly.
01:12:33.200
Does David think the reason Twitter agreed to censor information was due to a willing alliance
01:12:37.580
with the government, out of fear of the government if it didn't comply, or out of ideological conviction