The Megyn Kelly Show - December 14, 2020


Sam Harris on Political Tribalism, Cultural Divisions and Finding Inner Peace | Ep. 37


Episode Stats


Length

1 hour and 38 minutes

Words per minute

170.76505

Word count

16,905

Sentence count

879

Harmful content

Misogyny

14

sentences flagged

Hate speech

23

sentences flagged


Summary

Summaries generated with gmurro/bart-large-finetuned-filtered-spotify-podcast-summ .

Sam Harris is a philosopher, an author, and host of the Making Sense podcast. He's a liberal, but he's not woke. And he's one of those guys who just makes you feel like, Why couldn't I put it that way? Super smart, very big, big-brained guy. But he's also pensive. I think you'll find him illuminating on how you could quiet your own mind and enrich it after you quiet it.

Transcript

Transcript generated with Whisper (turbo).
Misogyny classifications generated with MilaNLProc/bert-base-uncased-ear-misogyny .
Hate speech classifications generated with facebook/roberta-hate-speech-dynabench-r4-target .
00:00:00.520 Welcome to The Megyn Kelly Show, your home for open, honest, and provocative conversations.
00:00:11.980 Hey, everyone, I'm Megyn Kelly. Welcome to The Megyn Kelly Show. Today on the program,
00:00:16.000 we've got Sam Harris. This guy is a philosopher. He's an author. He is host of the Making Sense
00:00:23.580 podcast. And I would say more than anybody else, he is the name that has been suggested to me
00:00:29.140 as a potential guest on the program. And I've listened to Sam in the past, but I was fascinated
00:00:34.140 to see what is it about him that has made him so very popular with the people who listen to this
00:00:39.900 podcast. And now I get it. Now I get it. He's a liberal, but he's not woke. And the way he talks
00:00:47.260 about the woke is incredibly eloquent and thoughtful and smart. And he's one of those guys who just
00:00:52.180 makes you feel like, oh, why couldn't I put it that way? Super smart, very big, big, big-brained guy.
00:00:58.140 But he's very thoughtful. He's pensive. I think you'll find him illuminating on how,
00:01:04.980 A, you could quiet your own mind and, B, enrich it after you quiet it. It may require you being
00:01:11.800 quiet for two months in a row, but we'll talk about that later. But I think we had a good spirited
00:01:16.660 exchange on things like Trump and his fans and sort of the media and their coverage of
00:01:24.340 politicians and their, quote, lies. Anyway, we'll get to Sam in one second. But first,
00:01:30.260 I want to talk to you about Jan Marini's skincare. You guys, this is a beautiful line of products.
00:01:37.560 I myself tried them. And number one, they don't smell, which I like. I don't like a lot of odors
00:01:43.900 in my products. Number two, they idiot-proof it for you. They tell you in the regimen,
00:01:48.560 this one goes on in the morning, this one goes on in the evening, this one you can put on both.
00:01:52.080 And so, you know how, like, you're a busy person, you don't have time, so, like,
00:01:55.020 they walk you through it. So Jan Marini Skin Research is a recognized leader and an innovator
00:01:59.700 in skincare. The easy-to-use products keep your skin feeling nice and refreshed, nice and hydrated.
00:02:05.260 You don't get, like, shiny. And the Jan Marini buzz is that it's one of the fastest-growing
00:02:10.120 professional skincare brands in recent years. That's what everybody's saying about it. And it's true,
00:02:14.060 because it's got a bunch of awards. It's used by multiple movie and TV production sets. It's
00:02:20.180 being used on the set of Spider-Man, I guess. Yay. And it's a five-step daily care system.
00:02:25.040 You cleanse, you rejuvenate, you resurface, you hydrate, and you protect. Their skincare
00:02:29.060 management system has been awarded 10 consecutive years by New Beauty Magazine as Best Skin Care
00:02:33.880 System for Aging Skin. Jan Marini Skin Research has earned more beauty awards from New Beauty than
00:02:38.840 any other skincare company. The products are hydrating, they're calming, they have numerous
00:02:44.240 clinical studies conducted by leading dermatologists, and you can get them anywhere, really.
00:02:47.700 They're at med spas, aesthetic offices, spas throughout the country. So go to janmarini.com
00:02:52.340 to find locations near you or purchase directly from the website. Plus, they've got some great
00:02:56.940 holiday offerings available and always with just two-day free shipping. Transform your skin
00:03:02.580 with Jan Marini. And now, Sam Harris.
00:03:07.500 Really excited to talk to you and also a little nervous because I've been watching a lot of your
00:03:13.180 interviews and listening to your podcasts and you seem like an intellectual giant and I'm
00:03:20.700 feeling ill-equipped to talk to you.
00:03:24.820 Well, listen, I want to put you at your ease because, first of all, I think you and I agree
00:03:31.020 about many things and I'm a fan of yours. I'm happy to talk to you about anything and where
00:03:37.160 we disagree. I think it'll be fun. So let's just go.
00:03:41.440 Good. Okay. I feel a little better. I like to start where I can with news of the day.
00:03:45.040 And you've been so smart and easy to listen to on the topic of wokeness and the religion of
00:03:51.520 wokeness. And just today, we saw a school in Virginia, an elementary school, announcing that
00:03:57.600 it is dropping part of its name, Thomas Jefferson, quote, due to the pain his legacy can cause,
00:04:04.500 not even actually is causing, but can cause black students, despite an overwhelming majority 0.98
00:04:10.360 of the parents saying, we don't want this. We are not in favor of this at all, but it's
00:04:16.040 happening. And then here in New York, it was announced not long ago that at one of these
00:04:21.960 sort of expensive private schools, it's actually happening at more than one, parents must now
00:04:27.320 outline their commitment to anti-racism when applying. And they have to attend anti-racist
00:04:34.100 training before they can even get into the school. The anti-racist label is a rhetorical
00:04:39.840 trick. So does all of this concern you? I know you've got daughters and you're concerned about
00:04:45.780 wokeness in general, but I think in the schools, it's especially pernicious. 0.80
00:04:49.340 Yeah, well, it does. At first, I think we should just bracket all the heresy we're about to download
00:04:56.360 with an acknowledgement that racism is real and it's a problem and it's been an excruciating
00:05:04.180 problem throughout our history, right? So it's not mysterious, this kind of moral panic we're seeing
00:05:11.580 around the issue of racism now. I mean, we know what the past was like. The problem is that people seem
00:05:20.040 to, one, not want to acknowledge the progress we've made, right? So there's something deranging
00:05:26.280 about acting like this is 1964, right? Given all that's happened in the last 50 years.
00:05:35.580 You know, we had a two-term black president and that counts for absolutely nothing. We have a generation
00:05:41.520 that's acting like, you know, being on the three-yard line with respect to racism is, you know, a moment
00:05:51.300 of moral emergency, right? So I'm not saying that racism is gone. I'm not saying that there are no
00:05:56.340 ways in which there may be policies that disadvantage people in various groups and we want to untangle all
00:06:07.360 that and respond to all that. But we have made enormous progress and there are just not that many racists
00:06:17.340 out there, right, who want racist outcomes, you know, starkly unequal and unfair outcomes for people.
00:06:24.880 And so to not acknowledge that is deranging. And so now we have a kind of an activist cohort in our
00:06:34.480 society. It's, it's not clear just what percentage of, of American society it is, but it doesn't have to be
00:06:40.720 all that big to completely derail our conversation about these issues. And yeah, so now, you know,
00:06:47.800 defenestrating Thomas Jefferson, um, uh, or anyone else from our, from our history who, who's, you know,
00:06:59.060 obviously, uh, record on race was, was imperfect, uh, to say the least, but still you can't, you can't
00:07:07.820 discount the fact that this is one of the most important people in American history and, uh, still
00:07:14.740 the person who made, you know, an outsized contribution to creating our country. Um, so it's, uh,
00:07:24.060 yeah, this, this particular activism makes no sense. And yeah, it's, it's deeply concerning that
00:07:32.520 we, you know, these are educators. These are people who will be teaching children, uh, almost without
00:07:41.020 fail at this point to, to view the world through the lens of race in a way that, that, you know,
00:07:48.640 arguably was appropriate 50 years ago, but really isn't appropriate now. It's in fact,
00:07:54.840 totally dysfunctional now. I mean, some of the critics have said, this is, this is David Duke's
00:08:02.360 dream realized where everything really is about skin color. It's immutable. It's not overcomable.
00:08:09.980 It makes all the difference in character and how well you can do in society. You know, these,
00:08:16.400 the differences are innate and, um, should dominate anyone's perception of another or just upon
00:08:23.140 first meeting them. And that's where we're going. And I know you've said, I thought you made an
00:08:28.320 interesting statement because you're not big on false claims of victimhood. And you were saying
00:08:33.720 they, they can diminish the social stature of any group, including one that really has been victimized.
00:08:41.280 And so constantly, as you say, you know, we're on, we're on the three yard line with race, 0.82
00:08:47.560 constantly pretending that we're not, that we're already, we're still all the way down the playing
00:08:51.480 field can actually set a group back. I mean, it can actually, it's setting back race relations and
00:08:59.520 black people as a group. Yeah. Because it's, it's dishonest, right? It's just, it's not,
00:09:05.340 um, and it also just, it just violates the basic principle that will get us into the end zone.
00:09:12.300 I mean, to, to, to, let's just acknowledge what the goal is here. The goal is to wake up in a world
00:09:19.420 where these superficial differences between people, the skin color, uh, has absolutely no moral or
00:09:27.940 political significance, right? That's, I mean, that, that was certainly MLK's dream and it should
00:09:34.280 be ours, right? And, and activists on this issue and a host of other issues as well are acting like
00:09:42.160 these differences between people are indelible, morally important, uh, politically essential to
00:09:50.580 recognize at every turn and that, that any disparity we see in our society, you know, if we, if we inventory
00:09:58.480 our fortune 500 companies and, and our various professions, we find that there's not a perfect,
00:10:04.320 uh, representation of the general population in all of those places, right? If, if exactly 50%
00:10:12.660 of cardiologists are not women, if exactly 13 to 14% of, of, uh, uh, people in the C-suite at Apple
00:10:22.580 aren't black, right? The only explanation for this is bigotry, right? Now that is just,
00:10:29.740 there are many things wrong with that, but first is that it is almost certainly untrue, right? I mean,
00:10:35.300 at minimum you would need real evidence to make that allegation. And there's so many other
00:10:41.000 explanations that, that, uh, promote themselves here. And it's so the, the dishonesty of it is toxic,
00:10:49.660 uh, not to mention the fact that it's seeing yourself as a victim perpetually and, and locating
00:10:55.100 your, your social power, your, your status in victimhood, which is really the, the algorithm
00:11:01.560 that is running here. Uh, it's just, it's, uh, it's intrinsically divisive, right? I mean,
00:11:07.460 if you're, if you're, if you're politics is based on the politics of identity rather than,
00:11:11.720 than, uh, looking for solutions that benefit everybody, looking for systems that are intrinsically
00:11:18.580 fair, right? You're, it's just, you can't, you can't possibly converge with other people because
00:11:26.440 all you're doing is ramifying your differences. Uh, so you're, you're, you're viewing everything
00:11:31.100 as zero sum in principle. Uh, and it's just, it's, um, yeah, it's, it's regressive. And again,
00:11:38.420 it, it, it doesn't acknowledge any of the progress, progress we've made or the, the principles by which
00:11:44.900 we've made that progress. One of the problems we're having is the complete stifling of conversation
00:11:52.320 on this and the messaging to white people and often white men that you just can't speak.
00:12:01.100 If you're white, you shouldn't speak on the issue of race. If you're a man, you shouldn't speak on 0.72
00:12:04.860 the issue of sexism. If you're cis, meaning you identify with the, your biological sex that you
00:12:10.820 were born with, um, you shouldn't speak on the issue of transgender and so on. And it's a really
00:12:16.140 effective way of silencing people in those groups because you're told if you even try to speak out
00:12:22.700 about it, nevermind, speak out critically. If you dare say anything critical of it, um, you've,
00:12:28.900 you've offended just by opening your mouth, you've offended. And I know this summer when we were in
00:12:35.100 the midst of the BLM protests and defund the police cries and all that, you, you were very outspoken in
00:12:42.320 a very, very powerful podcast that I I've listened to a few times. And you made a point up front about
00:12:48.700 saying you, it was a conscious choice not to put on a black scholar, uh, an intellectual to talk to
00:12:57.040 you and to make these points for you with you across from you, because you want to disabuse
00:13:04.420 people of this perception that the color of one's skin matters for a discussion on race or that,
00:13:11.920 you know, your gender has to determine whether you can speak up on trans or et cetera. I thought 0.96
00:13:19.240 that was really brave. And I think more of us need to say that. Yeah. I mean, you know, my,
00:13:23.280 obviously my bravery is, is to some degree founded on the fact that I've taken prudent steps to not be
00:13:29.560 cancelable. Right. So in the end, it's, it's not all that much bravery to, to be honest, uh, because I
00:13:35.420 just have, uh, I have, I knew what I wanted to be able to do. And I've, I've taken steps to ensure
00:13:43.600 that I, I, uh, run a fairly low risk of, of, uh, suffering some, you know, fatal error, you know,
00:13:51.180 bit, you know, career wise, uh, for doing it. So, and I mean, as you know, you've had adventures in,
00:13:56.360 in cancellation and it's, you know, I mean, you're an example I've actually spoken about. I mean,
00:14:00.720 someone at the absolute top of media, uh, saying one thing and being hurled from the ramparts for it.
00:14:09.780 Um, and, you know, and I would, I would say, you know, it was a, a malicious framing of,
00:14:15.560 of what you said. Um, and I mean, perhaps we can, we can, I don't know if you want to talk about that,
00:14:21.940 but I think there's a, the underlying problem here is that people want to hold people to the least
00:14:31.060 charitable interpretation of what they've said or done at every turn. And they're really not concerned
00:14:38.320 to know what was actually intended, what was actually going on in their minds or what, or what,
00:14:42.880 what their, their real aims are. Right. So if you can discover that someone has said something
00:14:49.420 that can possibly be construed as racist or sexist or transphobic or whatever it is, um, well then
00:14:57.140 that possible construal is the thing that you, you, you will amplify as an, as an activist or even as
00:15:04.580 a journalist now. And the goal is just complete obliteration of a person's reputation and, and,
00:15:12.720 you know, obvious aim to make them, them unhireable. Uh, and so, so yeah, I mean, I, I've, I saw the
00:15:22.740 writing on the wall there and I have, um, you know, kind of created a platform for myself where I can say
00:15:28.520 more or less anything I want, but yeah, the reality is that if you are a white guy who is talking about,
00:15:34.740 uh, in this case, police violence, uh, against, uh, you know, everyone, but in particular young black
00:15:42.960 men, uh, you are, you're on, you, you, you know, you have to be on the back foot, uh, to, to even think
00:15:51.420 you should be saying something, especially at a moment as fraught as immediate aftermath of the,
00:15:56.900 the George Floyd killing. Um, but the truth is the color of a person's skin has absolutely no
00:16:04.160 relevance to a conversation about the, the, the actual statistics of police violence and, and
00:16:11.240 crime and violence in our society. Uh, and it's, it's, it's bizarre and dysfunctional to think it,
00:16:18.600 it does have relevance there. And, um, you know, we, we have, uh, on that particular point,
00:16:25.340 we're, we are suffering a kind of public hysteria around this topic. I'm not saying that there isn't
00:16:32.000 too much police violence. I'm not saying that many of our cops are poorly trained and we should be
00:16:37.140 hiring, you know, better recruits and training them better. And, and, you know, far from defunding
00:16:41.640 the police, we should, we should give them more funds for, you know, to, to recruit better people
00:16:46.840 and to, and to train them better. Uh, but the flip side of that is the cops have the hardest job in
00:16:52.680 the world practically, right? I mean, they're thrust into situations that are, that are, um,
00:16:57.960 that the public, you know, don't understand how to interpret, right? I mean, we're in a country
00:17:02.620 where there are 400 million guns on the streets. This is not like policing in Japan where you can
00:17:07.580 assume when, when the guy turns around and races to the front seat of his car, he's going to, he's,
00:17:13.840 he's not going to pull out a handgun. In fact, you, you have to assume that he will in this case,
00:17:18.560 right? So cops are in all these situations where they have to make split second decisions about
00:17:24.220 what someone's doing with their hands. And it's, um, you know, yes, they kill a thousand people every
00:17:30.120 year, but they don't kill people, uh, in a way that suggests that there's an epidemic of racist
00:17:37.760 violence perpetrated by racist cops in our society. They're just, that's not what the data show.
00:17:43.740 And yet to say that in the aftermath of George Floyd, um, yeah, that's, you know, I think it
00:17:50.340 was rightly perceived as risky and, you know, many people would have been fired for recording a
00:17:55.160 podcast like that, but, you know, happily, I, I have taken pains to be unfireable.
00:18:00.060 I mean, the, the whole stifling of conversation around COVID first and the, the riots over the
00:18:08.000 summer, black lives matter, the gender stuff. This is really what made me want to get back out there, 0.99
00:18:12.780 just get in front of a microphone again, because the more you tell me, I can't say something,
00:18:17.260 the more I want to say it. And I've always been that way. I don't, I, you know, my executive
00:18:20.820 producer, Tom Lowell used to say to me at, at, at Fox, MK, you like to go to the place that hurts.
00:18:27.100 And it was true. And it wasn't, you know, gratuitous. It was because when no one wants to
00:18:33.280 talk about a thing, to me, it becomes ever more important to talk about the thing. There's nothing
00:18:37.740 wrong with talking about the thing. And we've gone completely crazy on this where just talking
00:18:45.780 is a fireable fence, just a fireable offense, just talking about something can get you fired.
00:18:52.720 It's, it isn't right. And it, it scares me that so many people have, forgive the term,
00:18:59.580 bent the knee on that. They, they don't talk anymore. They're afraid.
00:19:04.500 Yeah. Yeah. Well, this is a, um, I mean, it's a, uh, a great inoculation against this
00:19:10.540 comes in, in certain areas of academia. Although I got to imagine those are closing down a little
00:19:17.400 bit now, but, you know, having a background in philosophy, you know, in a philosophy seminar,
00:19:21.360 you know, just as thought experiments, you can talk about anything, right? Cause you're,
00:19:27.180 you're looking, you're probing for the foundations of ethics, right? You're trying to figure out what
00:19:31.300 makes something wrong. You know, if in fact things can truly be wrong in this world, you know, as,
00:19:37.740 as I think they can, uh, you know, in a, in a philosophy seminar, you might say, well, you know,
00:19:42.160 why can't, what is the difference between abortion and infanticide? And, you know, why can't people kill 0.74
00:19:48.780 their children if they don't want them anymore? Right. You know, if they do it, you know, within 15 minutes
00:19:53.760 of birth, right. If abortion is, is legal. I mean, so these are the kinds of things, sentences you
00:19:59.660 would speak in a room filled with, with people searching to understand ethics. And yet now one
00:20:09.100 would fear that somebody would just take that quote out of context and say, well, look at this maniac.
00:20:16.360 He doesn't understand why you can't kill babies. Right. Um, and it's, we need to be able to speak
00:20:23.060 without being paranoid. Uh, and we need our, we need our actual intentions to matter. I mean, it's,
00:20:31.480 it's, it's not that hard to figure out if somebody is actually racist and racist in a way that that
00:20:38.960 should matter. Right. Or actually sexist, you know, sexist in a way that should matter. And people don't
00:20:45.940 tend to conceal this stuff. Right. And, and it's, and people should be held accountable for what
00:20:51.600 they really are trying to do to the world. Right. So it's the fact that a, a bad joke or,
00:20:59.460 or something that can be misunderstood or a, or an honest question of confusion, um, get, can get
00:21:06.440 spun into, uh, a, a career ending reputation canceling offense. That's, uh, I mean, maybe that
00:21:16.780 was always possible in some way, but it does seem genuinely new in a way that it achieved scale based
00:21:23.200 on, you know, our, our, our new, uh, technology. I mean, just social media has, has leveraged this
00:21:30.320 into something that's, that's deeply, uh, unhealthy for us as a society. And, and, uh, we have to find
00:21:37.300 some way to pull back. Well, I know you've said all we have between us and the total breakdown of
00:21:43.820 civilization is successful conversations. That's, and we're on the brink, but it made me think because,
00:21:51.200 you know, having been in media for a long time, I do wonder whether these conversations are working.
00:21:58.140 The ones that we were having before the complete shutdown of all of them and even the ones that
00:22:04.280 we're starting to have now, you know, it's, there've been studies that say people don't want to
00:22:09.060 hear opposing ideas that they really, they try to avoid listening to them. That's why we have Fox
00:22:14.300 news and CNN and MSNBC on the other side. People go there for confirmation bias and just to feel good
00:22:20.480 about themselves. It's like hearing the sweet nothings about how right you are. And I really wonder
00:22:26.440 whether we've leaned so far into that, that conversation's over, it's canceled conversations
00:22:33.380 canceled. Yeah. Well, so I do think conversation is the only tool we have to ensure that 8 billion
00:22:42.500 of us can collaborate in an open-ended way. I mean, is, is the human project truly open-ended? You know,
00:22:50.620 is it possible that we're going to get through this century and into the next million years of
00:22:57.720 conscious life, you know, that, that is directly descended from who we are now, you know, whether
00:23:06.760 we will be recognizably human at that point is certainly an open question, but you know, is,
00:23:13.360 is, is this project doomed and is it doomed in the near term or are we going to, to, to thrive in
00:23:21.860 some truly open-ended way? I think conversation is really the, the whole story that will decide
00:23:30.840 whether, you know, which future we, we land in here, because it is the only thing that allows us to
00:23:38.220 modify the behavior of perfect strangers without violence, right? We have a choice between conversation
00:23:47.400 and violence and in all its forms coercion, right? So, you know, laws are also a form of violence in
00:23:53.780 the, in the end, you know, if we pass a law against you doing or saying certain things, well, what
00:24:00.260 happens when you break that law? I mean, people show up to your house with guns and you know, if you do
00:24:05.860 the wrong thing with your hands, you might get shot, right? So it's, it's, um, we have, we have our
00:24:12.660 ability to persuade one another based on common principles of reasoning and, and appeal to facts
00:24:20.720 that can be, you know, mutually appreciated, uh, or we have force and, uh, we should be very slow to
00:24:29.600 make an appeal to force for obvious reasons. Uh, so we have to get better at,
00:24:35.860 speaking to one another about difficult issues and we appear to be getting worse at it. And that's,
00:24:42.020 and we, and we appear to have crafted a, an information ecosystem in the media and social
00:24:48.040 media now that is, uh, has all of the signs of being a, a psychological experiment to which no one
00:24:55.780 consented, uh, whose purpose is to see how crazy it can make us, right? I mean, we, we, we, we're now in,
00:25:02.560 you know, mutually canceling and irreconcilable, uh, echo chambers. And there are many of them.
00:25:11.020 And, you know, I mean, just to look at, you know, what's happening now currently in our politics
00:25:14.620 around the election and, you know, what has been happening four years, uh, under Trump,
00:25:19.900 we're seeing a, a fragmentation of media and social media. Uh, and, you know, you, you know,
00:25:28.640 in the case of social media, we've built a built platforms that, that, you know, have been maliciously
00:25:33.720 gamed, you know, where the, where the echo chamber effect can be accentuated by the very business
00:25:39.060 model. Um, so it really is, uh, something that we have to get straight. And I think the prospect of
00:25:47.280 us just maintaining this particular course where we're this fragmented, where it's, where it's this
00:25:53.520 difficult to talk about the most important things that face us, you know, you know, a pandemic,
00:25:59.540 right? We can't even figure out how to talk to one another about what, what we should do in the face
00:26:05.540 of a pandemic, right? And that's, um, and, and trust in our institutions has eroded, right? And for,
00:26:13.160 for obvious reasons, but for reasons that we have to figure out how to nullify, right? So you just want
00:26:18.580 to take something that will be, uh, I would imagine dear to the hearts of many of your listeners,
00:26:24.320 I mean, the hypocrisy of public health officials, right? Where they castigate people protesting lockdown
00:26:32.500 as being, you know, murderously irresponsible for having gone out in public without masks, uh, or,
00:26:39.800 or gone out in large groups at all, uh, to express their, their political opinions. You know,
00:26:46.540 the way that flipped when the protests were for black lives matter, and you have the same public
00:26:52.420 health officials in many cases by the, I think by the thousand signing open letters in support of,
00:26:59.340 uh, these protests, which, you know, from an epidemiological perspective were just as crazy.
00:27:05.420 In fact, probably far crazier than any of the, the protests that were happening, uh, in protest over the,
00:27:11.700 over lockdown. Um, yeah, it harms the, the, the stature of our most important, uh, organs of
00:27:23.620 information in a pandemic, right? If science is going to be that politicized, uh, it's easy to see
00:27:31.820 how trust breaks down. More with Sam in just one second, but first let's talk bloomsy box. Are you
00:27:37.600 still struggling with what to buy your mom or your family member, your long lost friend for the
00:27:42.560 holidays? Here is a great idea. You can send flowers from bloomsy box. Your loved one is going to light
00:27:50.600 up when they get your flowers and they're going to show them to all their friends because they're
00:27:54.200 spectacular and they're special. These are not your average flowers. They are better blooms. You're not
00:27:59.780 going to believe the look on somebody's face when your bloomsy box flowers arrive. It's kind of
00:28:03.400 magical. And the reason is they're sustainably grown on family farms around the world. So you
00:28:09.720 place your order, your flowers are handpicked. They're arranged at the farm just for you. It's
00:28:15.400 like sending a personal one of a kind flower gift and they are delivered farm fresh straight to your
00:28:20.660 loved one's door. I'm going to do this for my mom. So she's going to get them and they're not going to
00:28:24.240 die in two days. Like you get in New York city. They're, they're going to last because they came from 1.00
00:28:28.020 the local farm, get an incredible price, a huge selection of artisan designed arrangements. There
00:28:33.720 are no hidden fees, no endless upsells and free shipping with your subscription. So check it out.
00:28:39.160 Whether you're going to send a single holiday arrangement or a subscription, uh, for someone
00:28:43.380 special to receive flowers every month, you got to go with bloomsy box. And I got you a special
00:28:47.940 discount. Go to bloomsy box.com and enter MK to get 15% off and free shipping. That's promo code MK
00:28:55.180 for 15% off at bloomsy box.com. There is a difference between the way the New York times
00:29:08.420 tries to get its facts straight, even when it fails, then something like Breitbart or Fox news or,
00:29:15.900 or, um, you know, any, any organ of, of, uh, uh, you know, I would say pseudo news on the right.
00:29:24.320 And that, that asymmetry is something that, you know, we have to figure out how to correct,
00:29:29.640 uh, because it's, it's, um, you know, it's, it's shattering our society.
00:29:35.460 Would you, I'm just curious, would you, do you think CNN and MSNBC are in the New York times camp?
00:29:42.900 No, no, I think they're worse. I think they're obviously worse. Most of the time. MSNBC is,
00:29:48.560 is obviously worse. CNN is, is often worse. Um, and it's, yeah, I mean, I find them both more or
00:29:57.700 less unwatchable now. I mean, I never watched MSNBC, frankly, but, um, but I think we have to be
00:30:06.680 honest about why this is because I, so I'm sure, you know, my opinion of Trump. Um, but I think,
00:30:15.880 I think Trump really is as, um, I mean, I think it is true to say that he is the most dishonest
00:30:25.340 and corrupt person who has, has, who has appeared in, in public life, in our country,
00:30:32.720 in our lifetime. I think that is a fact about him. This is not, I don't think that's just my
00:30:37.920 opinion. I don't think it's, I don't think this is really debatable. I mean, I think it's like saying
00:30:42.960 he's got a, you know, a vaguely orange hue to his skin or, you know, fairly colorful hair. Uh, I mean,
00:30:49.820 these are, these are, these are, these are facts about him that you can observe based on being able
00:30:54.920 to observe him for now decades, right? I'm not talking just about the last four years. I think
00:31:00.720 he is, he is a, um, he's not a normal politician. He's not a normal person in this specific regard and
00:31:09.380 his, his level of dishonesty and his level of, of selfishness, I mean, his, the malignancy of his
00:31:17.020 self concern, right? And everything gets sucked into that, that kind of moral black hole around
00:31:27.200 him, right? And he, and he attracts people into his orbit who will put up with that, right? You know,
00:31:33.500 people who are for their own reasons are willing to, who don't view that as the, the, the moral and
00:31:43.040 political abomination that, that it, that it actually is. And the, the media has had to figure
00:31:48.420 out how to respond to this. And, uh, you know, I would be the first to say that they've done a
00:31:54.540 terrible job. Um, and it has been deranging for them, right? And he said, because when they see that
00:32:00.980 he's, you know, he, he functions by a very different physics, you know, your reputation, like he's managed
00:32:06.200 to dissect out a, a kind of, you know, a personality cult within our society.
00:32:13.940 Now I got to stop you. I got a couple of thoughts. Um, I would suggest to you that one of the reasons
00:32:18.340 you don't see Hillary Clinton as in the running for most dishonest, corrupt politician ever is 1.00
00:32:23.640 because in part, the way the media didn't cover that they run cover for their favorite Democrats.
00:32:30.220 Like we saw with the Hunter Biden story, which has now come out as confirmed. The FBI has been
00:32:35.120 looking into him reportedly looking into foreign dealings, possible money laundering, a story that
00:32:40.660 Twitter and Facebook and the other media outlets wouldn't even touch. We were told it was untrue.
00:32:47.200 We were told it was a smear. It was an October surprise, not based in fact. Well, guess what?
00:32:51.840 That's not true. And Hillary Clinton at the time, she wasn't the most beloved democratic candidate,
00:32:56.640 but once it was her versus Trump, they were all in on protecting her stories about the Clinton
00:33:01.700 foundation. She lied every two minutes, but we didn't have a lie ticker going on her because the 1.00
00:33:09.060 media wasn't interested in doing that. Cheryl Atkinson was just on talking about that, about 1.00
00:33:12.840 this reporter who was telling her, we calculated out how many times Trump lies. And this is the number
00:33:17.460 of lies per minute. And she said, how many for Hillary? This is when they were running against each
00:33:21.260 other. And they said, Oh, we didn't have the time or the staff to do that. Well, it might be
00:33:25.820 interesting. She's, she's not, she doesn't have an adult relationship with the truth either. I
00:33:29.740 don't, I don't argue that, that Trump is truthful, not for one second. And I understand he may be in
00:33:36.460 a special category, but he may not be as far ahead of people like Hillary as you think.
00:33:42.780 No, I mean, I would definitely dispute that. Well, first I will say you're not going to find
00:33:46.380 much of a defense of Hillary Clinton or the Clintons as a couple in me. I mean, I just, I, you know,
00:33:53.800 I completely get why people were allergic to her and her candidacy. Um, and it was a lesser of two 1.00
00:34:02.700 evils, you know, choice from my point of view, but it was, it was much lesser for, for a host of
00:34:09.160 reasons, not just to this, this difference in, in dishonesty and, and personal corruption. Uh, but it's
00:34:17.880 just, I mean, the, the, the, or the ordinary range of lying and self-dealing and, and, um, you know,
00:34:29.500 perverse incentives and all the stuff we, we recognize in, in many, if not most politicians,
00:34:36.520 we have a sense of the general shape of that. Right. And I, you know, I, I, I get that if you
00:34:43.980 scratched the surface on, on most of the people in power and then certainly most of the people who
00:34:48.280 seek the office, right. I mean, it is kind of, it's a self-selecting group and there, there are
00:34:54.020 people, you know, who are in this game for the wrong reasons. And there, and there's just, you know,
00:34:59.820 many of the, the incentives are perverse and yet it's not glamorous, right. It's, it's, it's hard to
00:35:05.140 be idealistic about many of these people. And I would say the Clintons were, um, especially cynical
00:35:13.360 and, and yes, there's a lot of dishonesty that you can find in their, in their backstories. Um,
00:35:21.020 but it's just, it's, you know, Trump is orders of magnitude worse. I mean, he, he's done literally
00:35:30.180 hundreds of things, any one of which would have destroyed the political prospects of any other
00:35:37.840 normal politician. Well, that's true. But what I'm saying to you is that when you say we have a
00:35:43.240 sense of what these politicians were and how far, you know, down the dishonesty lane they,
00:35:48.620 they were, I'm, I'm just positing to you that you should put an asterisk there because I think the
00:35:55.320 media has exposed itself during the Trump era. And I think they have gone nuts. They've, they've taken
00:36:00.600 their dishonesty to a new level because they are deranged by him, but their bias has always been
00:36:05.800 there. And I do think it's worthwhile for people to stop and ask themselves how they've been
00:36:12.700 manipulated, how their perception of a politician like Hillary or Barack Obama was manipulated by a,
00:36:19.180 by the glowing coverage they received in general from outlets like the New York times. And to your
00:36:25.260 second point that you were making earlier about when I asked you about CNN and MSNBC,
00:36:28.720 they don't get a pass for, you know, well, they, they're nuts, but he drove them to it. Well,
00:36:35.560 too bad. That doesn't, you know, some of us had a very difficult time with Trump and some of us were
00:36:41.920 not huge fans of his for personal reasons. Right. But you can get past that if you're a strong person
00:36:48.600 and you're an ethical person, you understand what the job of a journalist is, which is not to make it
00:36:54.040 about yourself. And what we saw here with, with CNN and MSNBC was not a couple of journalists fell,
00:37:01.000 you know, like Jim Acosta, who used to be a straight news guy. He fell. He, he just went
00:37:05.040 totally partisan, got the derangement syndrome, couldn't be relied upon for truth anymore.
00:37:10.000 They all went down. All of them. I would like, it was shocking to me to see somebody like Anderson
00:37:16.780 Cooper, um, sacrifice his credibility because of what appeared to be his inability to see Trump in
00:37:23.660 any way that approached fair. Um, so I don't, you know, you, I, I, I felt like you may, may have been
00:37:30.060 sort of trying to explain their surrender to their nonstop attacks on him by, well, he drove them to
00:37:36.660 it. They drove themselves to it. I would just say that the, the asymmetry here really is difficult to
00:37:43.260 navigate. So, I mean, for instance, just to take the Hillary Clinton coverage, I think the media,
00:37:51.060 uh, is rightly, uh, concerned that they got Trump elected, right? I mean, they, first of all,
00:38:00.160 they gave Trump something, some insane ratio of coverage. And there's something like 20 times
00:38:07.340 the amount of coverage that they gave. I can't remember if that was, that was for Clinton or for
00:38:12.080 Bernie Sanders, but there's some, there's some comparison between the amount of airtime Trump got,
00:38:15.860 you know, in 2016. It was the other GOP candidates in the primary. Okay. So, so it's just,
00:38:21.740 they did that, right. And they, they have good reason to worry whether that was counterproductive
00:38:28.860 given what they wanted to happen, uh, in the election. Uh, and also it's, I think it's fairly
00:38:35.960 well established at this point that the, the Comey, uh, you know, reactivation of, of the email,
00:38:43.420 uh, investigation, whatever it was 11 days before the election in 2016, the coverage of that and
00:38:51.520 the sort of the, the, the, the scrupulosity of that moment, both from the FBI side and from the
00:38:57.320 press's side, like, Oh yeah, we really got to talk about this now from, you know, you know,
00:39:01.620 yet another 24 hour news cycle when there's only 10 days left left. Let's talk about Hillary Clinton's 1.00
00:39:07.460 emails. They rightly think that that probably cost her the election right now. Obviously there are
00:39:13.980 many other things that, you know, cost her the election. She was a terrible candidate and
00:39:18.100 she didn't go to Wisconsin. Yeah. You, I mean, you might, yeah, you could say that her
00:39:22.160 election was, was overdetermined, but when the people who were tracking the polls in those last
00:39:27.980 days of the campaign just saw the, the, the direct effect, or at least now claim to have seen the
00:39:33.560 direct effect of, of that coverage. So, so it's given that history. Yeah. What do you do with a
00:39:41.080 Hunter Biden story? Right. It's a hard problem. You know, I, and because it's, it's, it's not clear.
00:39:47.840 It isn't hard. You're so much more forgiving of these people than I am. First of all, it was very
00:39:52.640 obvious that they were giving Trump too much free airtime. Even to me, I was a person responsible
00:39:58.160 for putting points on the board when it came to ratings during the Trump rise, the GOP primary and
00:40:02.560 thereafter. And this is in my book, but we, we had many, a meeting, my executive producer, my team
00:40:08.060 and I did about the need to stop, to pull ourselves away from the crack cocaine. You know, like Trump
00:40:15.680 was amazing television. Every time you put him on television, the ratings soared, but it wasn't fair
00:40:20.800 because Scott Walker was terrible television and incredibly boring to watch. I like him. I like him
00:40:27.040 just for the record, but he is not a dynamic television personality and it was totally unfair.
00:40:32.240 What we were doing. So even though we wanted the crack cocaine, we didn't snort it, right? We,
00:40:37.260 we made a conscious decision to try to get ratings the old fashioned way, which was interesting debate
00:40:41.380 and reporting the news in an entertaining way, but not, not putting our thumb on the scale.
00:40:45.640 They did it intentionally. They knew what they were doing. Jeff Zucker knew what he was doing.
00:40:49.980 Noah Oppenheim knew what he was doing on the today show, allowing Trump to phone in,
00:40:53.460 to do phoners as a presidential candidate when nobody else would have been afforded that. Right.
00:40:57.180 Um, they did it for ratings. They sold their souls out for numbers, Joe and Mika too. They kissed
00:41:03.420 Trump's ass when he was running. Why? Because their numbers would soar on a show that was not doing
00:41:08.000 well. So I, that I don't, I don't forgive them one bit of that. In other words, I don't allow them
00:41:13.060 to use that to justify their overcompensation and trying to ruin him every moment of his presidency.
00:41:18.620 They, they, they made, made their bed and they needed to lie in it. And the Hunter Biden story
00:41:23.140 was very clearly legitimate. He, he had never denied it. There was an FBI subpoena that was
00:41:31.360 verifiable. There were third party witnesses who had come forward to say, I participated in this.
00:41:36.980 And let me tell you about it. Whose credibility was not assailable. They just didn't want to do it.
00:41:43.160 They didn't want to do anything that could hurt Biden.
00:41:46.360 Yeah. Yeah. Well, well, I understand it. I mean, you know, it's, you know, I've always been,
00:41:51.980 uh, I've always made an effort to be fair in my criticism of Trump because I do, I do think it's,
00:42:00.940 I mean, I think intellectual honesty is the, the master value here. And I think, I think you,
00:42:06.000 there's, if you think he's racist, well then, you know, argue on, based on real evidence that he's
00:42:11.640 racist and not pseudo evidence. Right. And I, and I think it's, you know, and this is true for 0.67
00:42:17.040 anything else you might want to allege about him. Um, but I, I think that, you know, the worst
00:42:23.640 problem with him beyond anything else that has, that he's done or tried to do is, uh, the way in
00:42:32.820 which our, our trust in our institutions and, and a whole style of, of speaking and thinking about
00:42:40.620 institutions, uh, has, has just fully eroded. I mean, we're just, you know, it seems that it
00:42:48.220 seems that half the country thinks we don't need institutions anymore. Right. Now that we can just,
00:42:52.820 let me ask you about that, but let me ask you about this. So I, I've been following this too,
00:42:56.140 and I've been experiencing it myself as a citizen, to some extent, not, not fully, but to some extent,
00:43:02.440 right. Because take, take Comey and the FBI. I defended Jim Comey. I thought he was a man of honor.
00:43:08.620 Um, the whole story of Robert Mueller and Comey and, you know, having each other's backs and
00:43:12.780 protecting each other and whatever the ethical, uh, quandaries they faced and rose above.
00:43:18.220 I defended him. And now I see him as a partisan hack. I really do. I've completely changed on him.
00:43:23.780 I defended the FBI. I know a lot of these FBI agents. I, I tend to defer to law enforcement.
00:43:28.840 I have law enforcement in my family. And I think a lot of women are deferential to law enforcement for 1.00
00:43:33.340 reasons having to do with nonstop crimes being played on the evening news as we were growing up,
00:43:37.820 but that's just my armchair theory. Anyway. Um, then I saw the, the Peter Strzok emails and the
00:43:45.020 other emails coming out from the FBI and how partisan they were and how determined they were to
00:43:48.280 bring down Trump. And I thought, maybe I have been too deferential. Maybe Trump kicking these tires
00:43:56.600 is not such a bad thing. Maybe we have been too trusting in these institutions. And of course,
00:44:02.600 we've discussed media and how people were trusting media in a way they shouldn't have been. Um,
00:44:08.340 I don't, I don't, I agree. It's gone too far. I don't think you throw out the baby with the
00:44:13.360 bathwater, right? I think we're sort of now at the point where we trust no one and we've become
00:44:17.980 conspiratorial and things are getting weird. There's becoming like a cult, like love for Trump
00:44:26.080 that we're in, which he looks deified, that's concerning. And therefore his offhanded need to 0.98
00:44:33.280 attack every institution that says something bad about him or does something he doesn't like
00:44:37.060 or person, you know, governor Kemp of Georgia, who's his fan is, it isn't healthy. I see all of
00:44:43.900 that, but how, how do you ascribe, you know, how do you deal with those revelations of dishonesty,
00:44:51.900 um, versus maintaining a country in which we do have to trust institutions and move forward?
00:44:58.800 Well, I just think you have to put them in their proper perspective, right? So again, what you,
00:45:04.620 you drill down on anything and you start reading people's private emails or emails they thought
00:45:09.300 were going to be private, you start reading their texts. Yes, there's no question you're going to
00:45:14.280 find things that, that, you know, are embarrassing and, and are all too human. Uh, and that's totally
00:45:24.800 understandable. I mean, these, these things weren't intended for public consumption, right? So, you know,
00:45:29.680 a hack of the DNC or, or, uh, you know, subpoenaing people's text messages when you, when you're
00:45:36.580 investigating, you know, anyone that, you know, attorneys or members of the FBI, yet you'll, you'll
00:45:41.960 find evidence of bias and all of that. Right. So, but that, no one's surprised by that. What people
00:45:46.760 should be surprised by is, uh, you know, to take one institution institution in particular,
00:45:52.700 that you are under an election, um, and all the systems that support it. And we, we should be
00:45:58.200 surprised that we have a sitting president who in the run-up to the election would not commit to a
00:46:06.180 peaceful transfer of power in the event that he lost. And then on election night with millions of
00:46:11.980 votes still coming in, claimed to have won, right. And demanded that the voting be stopped,
00:46:17.620 right. Whether the ballots, you know, the ballot counting stop, um, that just, just that, if you
00:46:24.040 just, just make a little documentary about that moment and what it says about where we've come and
00:46:32.300 the, the divergent reactions to that in our society, right. The fact that we had something like half of
00:46:40.300 the society that simply didn't care this was happening, or they had some construal of it,
00:46:45.920 where this is all, this is, this is not only benign, this is good, right. He's, he's disrupting
00:46:52.680 everything now. Now he's disrupting our expectations about the peaceful transfer of power and, and, uh,
00:46:59.300 the integrity of our elections, right. He's calling a fraud on an election that, that, that is in
00:47:04.580 process, you know, where we haven't seen any evidence of fraud yet. Right. He, he called the
00:47:09.820 election he won in 2016 fraudulent. Um, and that's, you know, and this should be the lens
00:47:16.380 through which we look at everything that has happened subsequently in the last month and a
00:47:20.840 half or so, right. It's, it's, it's not that if you don't go looking for fraud, you'll never find
00:47:27.340 it. Of course, that there's, there's going to be some ambient level of election fraud that there
00:47:32.120 will be, and there'll be this, there'll be crazy behavior. Um, and if you put yourself in a
00:47:37.600 position to just look for crazy behavior, you will find it. But the question is, can you find it at a
00:47:43.320 scale, uh, that reveals this election to be, uh, completely fraudulent? Uh, you know, there's no
00:47:51.360 evidence for that at the moment. No, I get it. I get it. But I have to say, I think that I look at
00:47:55.060 the media who's that's, that's how we're learning about the substance of these election fraud claims,
00:47:59.580 right? We trust, we need to trust them to tell us what has he filed? What is the court saying?
00:48:04.780 Does he have the proof? It's, it's the same as those doctors who justified the BLM protests
00:48:11.960 coming out now telling us to stay inside and not go shoulder to shoulder. We're looking at them
00:48:17.720 saying, I don't believe you anymore. You already sacrificed your credibility. I don't know who to
00:48:23.060 believe, but when, when confused with who to believe people do revert to tribalism. They go to
00:48:29.940 the person who's wearing their team Jersey. And I think that's why a lot of these Republicans
00:48:33.980 who have lost all trust in media are believing Trump, even though I've said on this show many
00:48:40.080 times, I don't think any of his claims are robust. I've, there's been a couple his, the last one he
00:48:44.560 filed, um, in the state of Georgia that challenged like the failure to match up the signatures and how
00:48:50.080 many votes were thrown out versus in prior elections. And it went down this year, even though the vote was
00:48:53.920 five times higher by mail, blah, blah, blah. You could, you could make an argument on, on a couple
00:48:58.520 of them, right? So we're trying to give him his due without losing touch with reality. But I
00:49:03.500 completely understand why people that it's gone. Their, their trust in the, in the information
00:49:09.920 deliverers is gone. And just for the record, the FBI thing was not just the Peter struck emails,
00:49:14.240 but as you know, an FBI agent pleaded guilty for doctoring a subpoena. They got subpoenas in the,
00:49:19.900 in the case against Trump from the FISA court by using the, the, um, the dossier, which had been
00:49:26.300 discredited and they knew it wasn't true. Like there were a lot of things that were exposed.
00:49:30.860 Why? Because Trump fought back and he made outlandish claims at the time that weren't true
00:49:36.100 in defending himself, but he wasn't guilty. That was the truth at the end of the day. And,
00:49:41.160 and same thing with impeachment. There's just been so many overreaches in order to destroy him
00:49:45.760 that in essence, they've endowed him with the credibility to come out and challenge anything.
00:49:52.300 I'm not, I don't excuse Trump for throwing wild claims around. I'm just in the way you're trying
00:49:57.060 to explain the media's distorted minefield when it comes to Trump. I'm trying to explain why he now
00:50:02.340 has this ultimate credibility with all these folks, because the other side is just, it's collapsed.
00:50:07.860 The, the information deliverers have collapsed and it matters.
00:50:11.720 Yeah. Well, it does matter. It hasn't completely collapsed. And I think people need to be sensitive
00:50:16.560 to, to, uh, the difference between, you know, plausible interpretations of events and, you know,
00:50:26.580 completely unprincipled conspiratorial, you know, tinfoil hat, crazy interpretation of events. And,
00:50:32.900 and there, there, there are reasons why we have this, this category of conspiracy theories that,
00:50:38.620 that doesn't subsume all of our thinking about everything all the time. Right. And there,
00:50:44.040 there's a reason why there's a stigma associated with conspiracy thinking because it, it, it reliably
00:50:51.680 manufactures errors, right. It rests on, on not acknowledging, uh, the power of incentives,
00:50:59.860 right. And so to take the, you know, the case of the, the election fraud, uh, conspiracy right now,
00:51:05.220 again, I'm not saying there isn't some level of fraud, but there are many reasons to think that
00:51:08.900 whatever level of fraud there is, it almost certainly happens on both sides in the election,
00:51:14.640 right. And it's, there's not a lot of incentive for individuals to commit fraud and, and, and doesn't,
00:51:21.080 there doesn't seem to be the apparatus to allow us to really commit it at scale, uh, across multiple
00:51:27.160 states. Um, and you know, that's a good thing. I mean, obviously we want to, we want an election
00:51:32.640 system that we can be confident in and that, and that is, uh, designed in a way to truly minimize
00:51:37.960 fraud and error. And we have to, this is a project we, we need to, to engage, uh, if for no other
00:51:44.920 reason than to, to restore confidence in, in our election system. But the, the reality is, is that
00:51:51.000 all of this is happening in a context where many of the people in power, right. The governors,
00:51:56.820 the secretaries of state, the legislators, the election officials, the judges who have to hear
00:52:02.880 these cases, many of them are lifelong Republicans, right? Many of them, many of them surely voted for
00:52:08.960 Trump in this election. And so you have to, to, to, you're arguing to be really conspiratorial about
00:52:15.900 this. You're arguing that these people are somehow incentivized to risk going to prison in order to
00:52:23.360 help Joe Biden. Back to Sam in one second. But first, have you ever Googled yourself,
00:52:28.180 your neighbors? The majority of Americans admit to keeping an eye on their online reputation and
00:52:32.960 why shouldn't you? But Google and Facebook are just the tip of the iceberg when it comes to finding
00:52:37.160 public records. There is an innovative new website called truth finder, and it is now revealing the
00:52:43.160 full scoop on millions of Americans. Truth finder can search through hundreds of millions of public
00:52:48.720 records in a matter of minutes. Truth finder members can begin searching in seconds for
00:52:53.780 sensitive data like criminal traffic and arrest records. Before you bring someone new into your
00:52:59.620 life and around the people you care so deeply for, consider trying truth finder. What you may find
00:53:05.220 may astound you. This might be the most important web search that you ever do. So do it. Go to
00:53:10.620 truthfinder.com slash Kelly right away to start searching and be prepared for what you find.
00:53:16.700 Again, that's truthfinder.com slash Kelly. Okay, we're going to get back to Sam in a second. But
00:53:22.480 first, we want to bring you this feature we call Sound Up, where we talk about some of the sound
00:53:26.500 bites making the news that we think are interesting and want to share with you. And today, we are going
00:53:31.720 to talk about Hunter Biden. Remember old Hunter Biden? If not, you could be excused since it was a
00:53:38.460 story that was totally buried by the mainstream media. CNN, Politico, others reported last week
00:53:44.700 that Joe Biden's son is being investigated by the FBI. His taxes, his dealing with China,
00:53:51.720 money laundering, and more. We don't know how deep this goes, but it doesn't sound good.
00:53:57.080 The story has a lot of the same hallmarks as the New York Post's reporting on Hunter Biden
00:54:02.140 in October. Remember that? It was a big scoop. It was about what was on the Hunter Biden laptop.
00:54:08.200 That story, however, which was published before the election, was suppressed, totally censored by
00:54:15.540 Twitter, Facebook, and other social outlets. Basically, no one wanted to go with this. It
00:54:19.260 wasn't just social. It was print magazines and print newspapers and television. Nobody wanted to
00:54:24.120 touch this thing. They decided it was not to be discussed. And they said it was because it was
00:54:29.200 unreliable. But the truth is more likely that it was bad for Joe Biden. Anything bad for Joe Biden has
00:54:34.520 to be suppressed, according to the media. So how did the media treat the story at that time prior to
00:54:40.200 the election? Listen to some MSNBC highlights. Watch for President Trump to go after former Vice
00:54:46.120 President Joe Biden's son, Hunter Biden, and unverified emails about his business dealings,
00:54:51.300 a story that many intelligence experts say has all the hallmarks of a foreign interference campaign.
00:54:56.760 When there is a New York Post article that is false,
00:54:59.760 it's much better for Twitter to let people read the New York Post article and sit there and laugh
00:55:07.840 at the hokey story of a computer repairman looking at a computer going, this sure does look suspicious
00:55:14.600 to me. I'm going to call Rudy Giuliani. Let that out. OK, because people read this story and then
00:55:23.380 they'll go, this is really one of the stupidest October surprises I've ever seen before. What did
00:55:29.740 he have? X-ray vision? Oh, my God. Why? Why when they're trying to portray a dumb person, do they
00:55:35.700 always put on a Southern accent? It's so irritating, honestly. And then they're like, me, an elite? Oh,
00:55:42.240 wait, I love being an elite. Yes, I am. I am. Screw everybody else. Anyway, so that was Morning Joe.
00:55:48.260 Here's Christiane Amanpour at CNN, who seems to forget the role of a journalist. 0.91
00:55:52.860 As you know perfectly well, I'm a journalist and a reporter and I follow the facts and there has
00:55:58.180 never been any issues in terms of corruption. Now, let me ask you this. Yesterday, the FBI-
00:56:04.100 Wait, wait, wait, wait. The FBI- How do you know that?
00:56:07.140 I'm talking about reporting and any evidence. I'm talking to you now-
00:56:12.380 OK, I would love if you guys would start doing that digging and start doing that verification.
00:56:16.420 No, we're not going to do your work for you. I want to ask you a question.
00:56:20.640 There's no reports.
00:56:22.860 So we can't report on it. Why were there no reports again? Oh, because it was buried by
00:56:27.140 people like you who just decided without investigating it that it was untrue. Well,
00:56:31.120 guess what? The FBI has been investigating it. And frankly, it looks like they've been
00:56:35.480 investigating it for a long, long time because in order to get the subpoena to get that Hunter
00:56:40.020 Biden laptop from the legally blind guy who was repairing it. Remember, Hunter left his laptop there
00:56:46.080 and he never went back. And finally, the FBI came in and got it. That was back in December of 19,
00:56:52.920 December of 19. So for the FBI to have gotten that, it suggested that they had an open investigation on
00:56:57.540 Hunter Biden. That's how the FBI can't just like randomly throw out subpoenas back in December of
00:57:02.580 19, which they would have known had they bothered to look at it at all, at all. But no, it was the New
00:57:07.940 York Post. That's a Rupert Murdoch publication. So this has got to be a lie. And, you know, Jack at
00:57:14.260 Twitter, who's got some far left guy deciding what gets censored, spoke out and said, no, we're not
00:57:19.020 going to do it. And so the media lemmings followed. And now it turns out it's a very big story. And if
00:57:24.740 it were Republican, they'd be treating it like it were white hot. So now fear not. Now they're on it.
00:57:31.380 They got Politico. They got CNN. They are on it now that Joe Biden appears to be headed to the White
00:57:36.880 House. Not surprising. Back to Sam.
00:57:43.280 I've been thinking about it when it comes to race and we've been talking about it on the show when it
00:57:46.480 comes to gender. And the reason the gender thing is important is because the scientific world has
00:57:53.220 collapsed, too. Right. Like science is done and you're not allowed to say that there are only two
00:57:59.420 genders and there's only two biological sexes and you're not allowed to question whether it might
00:58:04.680 not be healthy to put a girl who suddenly at age 14 for the first time says she might actually be a
00:58:11.260 boy on puberty blockers or cross sex hormones or allow her to get, quote, top surgery before she's 1.00
00:58:18.000 even reached her 18th birthday because that's not allowed. So you go to a scientist, you go to a
00:58:22.620 psychiatrist to check out whether that's true for your daughter. And the standard of care is for him to
00:58:27.860 just affirm, yes, you're trans, you're trans, you're trans. These are the standards of care.
00:58:32.200 It drives me insane. And I know you're a neuroscientist. We had Deborah Sell on the
00:58:36.100 program talking about how she left the field. She's now a journalist covering that field because 0.77
00:58:40.340 she didn't think she could say it was scientifically true. I feel like people feel they need to fight
00:58:45.820 back. You know, if Biden gets in there, we're going to see the return and the emboldened
00:58:49.320 nature of all of this. No one's going to shut them down. You've got people like you, Sam,
00:58:54.400 who speak honestly about the dicey issues. You're obviously a, you know, a Democrat, a liberal.
00:59:01.240 But you speak honestly about those issues. But so few people do that people are getting hurt now
00:59:06.540 in the scientific field. They're getting hurt. What do you make of that?
00:59:09.840 Yeah, well, a few things here. One is scientists are just people, right? And not everything they do
00:59:15.440 is science. So, you know, you have scientists who express their opinions on many topics,
00:59:21.140 you know, as I have here on many topics. And it's not, you know, there's no guarantee that
00:59:28.280 what they're saying is, you know, scientifically defensible or convergent with what they would say
00:59:35.240 when they have to put their scientist hat on. And yes, I would agree with you that it is
00:59:41.920 very costly to the reputation of science and any specific institution, you know, scientific journals
00:59:50.860 when they express opinions that are at minimum, you know, highly debatable in terms of their,
01:00:02.060 you know, ethical integrity and connection to science. And it's wrapped up in the mantle of,
01:00:10.080 you know, this is now a scientific opinion, right? So that's, that's a problem. And there,
01:00:15.920 you know, there are many issues, you know, some of which you just raised where
01:00:20.120 they're just, they're, they're fraught issues for a reason. It's hard to know what to do about
01:00:26.420 certain things. I mean, you take the transgender issue, right? I have no doubt that transgenderism
01:00:32.540 is a real phenomenon, right? It's not just made up. It's not just a product of culture.
01:00:39.500 It's not just a, a social contagion, but is there a degree of social contagion riding on top
01:00:47.440 of a real phenomenon that we have to worry about? I mean, specifically with the issue you,
01:00:52.740 you mentioned, I'm sure, I think you're probably referencing Abigail Schreier's book
01:00:57.540 there. And just, you know, just among girls transitioning to, to being boys or wanting to,
01:01:04.940 um, yeah, that's, that has to be discussable, right? And, you know, as her efforts to get, 0.97
01:01:13.020 you know, her side of this discussion out have shown, it's, it's very hard to discuss,
01:01:17.960 right? There are people who want to, to cancel her over this. And, you know, somebody like JK Rowling,
01:01:23.880 you know, gets, uh, you know, uh, at least attempted, there's an attempted cancellation of
01:01:29.880 her, um, based on something absolutely benign. She said about, you know, the trade-offs between
01:01:36.680 women's rights and trans rights. And there were, there are trade-offs there. And it's just, there
01:01:40.980 are, there are moments that are hard to navigate based on, uh, appeals to, you know, the primacy of
01:01:48.560 identity, uh, around those issues. And, um, you know, when she was objecting to the corruption of
01:01:54.040 language where we can't talk about women anymore, we have to talk about, about people who men's
01:01:58.780 straight, um, you know, it, it is in fact true to say that if she were not this, you know, billion 0.76
01:02:05.940 dollar colossus of a writer, she probably would have had her career ruined over the, the absolutely 1.00
01:02:12.020 anodyne thing she said about trans issues there. I mean, this is, there's, there's no evidence at
01:02:18.820 all that she's remotely bigoted against trans people. Um, so yeah, it, it's a, um, the, the stuff
01:02:28.340 is hard to talk about. And the only thing we can appeal to really, if this is going to work is a good
01:02:35.320 faith, uh, engagement with facts and arguments and let the, let the best arguments and the,
01:02:43.300 and the most searching, honest engagement with facts win. I heard John McWhorter on your show
01:02:49.160 saying, uh, he believes these sort of woke, wokesters are in good faith, that they think
01:02:55.300 they're doing good. They're going to help you understand, you know, how racist you are, how
01:03:00.000 transphobic you are, but he also said you, they, they're not persuadable. The only answer is to 0.96
01:03:06.720 fight them. Do you agree with that? Yeah. Well, I think, and fight by, by which he meant in most
01:03:13.700 cases, ignore them, you know, go around them, no longer give them any power. Um, but you know, I,
01:03:20.820 I, I have to think that more people are persuadable in the fullness of time. We're not dealing with
01:03:27.400 a different species of person, you know, in all of these camps, we have recognizably human people
01:03:34.780 who have been persuaded of certain dogmas or certain bad ideas, and they're not disposed to,
01:03:44.340 to run the, the, the, the newest operating system, uh, that would debug these ideas. Right. And so
01:03:53.460 it's, it's, it's, it's, we, we have to just keep advertising the importance of, of getting a
01:04:00.520 firmware upgrade here. And it's, it's true. It's true on both sides. And it's, it's true on the far
01:04:05.760 left. It's true on the, on the right or whatever you want to call Trumpistan. It's not clear how,
01:04:11.400 how that relates to conservatism now, but, um, I mean, to come back to a point you made earlier,
01:04:16.620 I hold out hope for it being easier to deal with the wokeness and the hysteria on the left
01:04:25.720 under Biden than, than under Trump, because Trump was such a, a super stimulus, you know,
01:04:33.020 he was such a, such a confirmation of the worst fears or an apparent confirmation of the worst fears
01:04:39.560 of the left. I mean, cause again, much of what is, has been alleged against him on those particular
01:04:45.880 points, you know, with respect to, to race in particular, I think is, is not true. Right.
01:04:51.780 I mean, I, I happen to think, you know, I happen to believe Trump is a racist, but I don't think
01:04:56.040 he's a white supremacist. And I don't think he has, you know, I think he has been unfairly tarred with,
01:05:00.660 you know, the, the good people on both sides or the fine people on both sides, hoax, you know,
01:05:05.700 that, that is, you know, you go back to that press conference and yes, he did condemn white
01:05:09.600 supremacists and neo-Nazis. Clearly my hope certainly, and my, if I had to bet money on it,
01:05:16.500 I would, I would say it's going to get easier under Biden to recognize how dysfunctional wokeness is
01:05:26.360 politically and ethically, uh, because we, they won't have Trump to point to, or, you know,
01:05:31.340 they, they probably won't have Trump to point to if he doesn't emerge in some other.
01:05:35.700 So far it doesn't look like it's going that way. I mean, so far it does not look like it's going
01:05:39.480 that way. You've got, um, critical race theories coming back thanks to what Biden says will be his
01:05:45.160 first executive order. Uh, the mandated sessions at the federal government amongst, uh, its workers
01:05:50.140 and its contractors, they're already saying that they're going to try to undo the restoration of
01:05:54.560 due process rights for men who get accused on college campuses of sexual assault. Um, those are
01:06:00.720 not good signs, not at all. No, no. And yeah, I mean, it, again, I just, I just know what it's
01:06:08.200 like. I know what it'll be like for someone like me to criticize all that without having to bracket
01:06:15.720 everything I say with an acknowledgement of how crazy things are on the right. You know, it's just,
01:06:21.360 it's, it's not, um, it'll, it'll just be very easy to do now. Uh, you know, maybe this wave is just
01:06:30.920 now cresting, you know, maybe there, there is, there is a bizarre effect here where when the problems
01:06:37.780 get smaller and smaller, the people who are, who are most focused on these problems get more and more
01:06:45.960 agitated and more and act more and more like things have never been worse. Right. It's just,
01:06:51.700 it's on some level, this is the, the narcissism of small differences. Right. I mean, just, if you're
01:06:56.960 not singing from precisely their hymn book, you know, on, on any of these woke issues, well, then
01:07:04.780 you're a Nazi. Right. And, and we're talking about things like, you know, whether you can compliment 0.91
01:07:09.560 someone on their hair, right. You know, is, is that a racist microaggression, right? We're not
01:07:14.420 talking about people getting lynched. We're not talking about, you know, people, um, having to
01:07:21.320 function under the rule of, of, of, you know, race-based laws, right. We're talking about, um,
01:07:27.420 you know, off color jokes, you know, someone gets into a, into a elevator at an academic conference.
01:07:33.920 And when they ask me, you know, what floor he says, women's lingerie, right? Like that is a,
01:07:38.120 that is a life deranging cancelable offense because he did a, he, he, you know, offered up a,
01:07:44.420 a dad joke from the 1950s. Uh, that's, that's where we are. And it's, it's eminently
01:07:51.920 criticizable. Right. And there's so many people, I mean, you know, to, to mention some of the black 1.00
01:07:56.820 intellectuals who I, I didn't invite on that podcast. There's so many great people like John
01:08:02.200 McWhorter and Glenn Lowry and Thomas Chatterton Williams and Coleman Hughes. And, you know, many of
01:08:07.360 you, I know you, uh, many of whom I know you've spoken with, um, and Camille Foster and, and, uh,
01:08:15.240 Chloe Valdary. And, and I mean, there's just, there, there are people who, you know, it shouldn't
01:08:19.740 matter that the, the, the people I just mentioned are black, uh, but it does matter. Right. And they,
01:08:26.820 they really have a, an enormous responsibility and, and they are shouldering it, uh, to perform an exit,
01:08:34.920 perform an exorcism here. Right. I mean, this is just, because the, the people, uh, on the far left
01:08:42.900 simply cannot hear it from people like us, right. I mean, if you, if you're white and, and, you know,
01:08:52.180 obviously privileged, you know, you have all the privilege marks, uh, that could be, you know,
01:08:57.460 ascribed against anyone in, in our society at this point. Um, you by definition don't get it
01:09:06.040 and can't talk about it. Uh, but there are many people who, um, really can talk about it and,
01:09:12.440 and for whom they, I mean, they really are a kind of kryptonite and it's not an accident that
01:09:17.280 no one really wants to debate them. I mean, the people are not lining up to debate Glenn Lowry and
01:09:22.380 John McWhorter about race issues or Shelby. Or Coleman Hughes, Coleman Hughes, who's been out
01:09:27.040 there saying, Hey, Ibram X. Kendi, just let's talk about your, your book. I don't believe in it.
01:09:32.160 I think you've made mistakes. You've been sloppy. Let's talk about it. And he won't debate Coleman
01:09:36.400 Hughes, a 24 year old guy who's done his homework because he's afraid. But, but to your point earlier,
01:09:43.640 they, right. They can't assail Glenn Lowry and, and Coleman and certainly Thomas Sowell and the way
01:09:49.680 that they could come after you or me, but they won't hear from them. Number one, they get called
01:09:54.480 uncle Tom's number two, they do not get invitations to appear on shows. No one's interested in putting
01:10:01.740 them out there to say how they feel. And I'll just give you one other small example. Um, it's a stupid
01:10:09.020 story, but my, I am in the Bethlehem central hall of fame. Perhaps you didn't know that Sam, but I
01:10:16.040 no, I did not know that. Now I'm, now I'm intimidated. I was inducted years ago. It's
01:10:20.600 not Stanford, but it happened. And, um, they, there's a push by some kids there to get me booted
01:10:27.020 out. Why did I say, cause I said something controversial about race. No, uh, because I
01:10:32.740 retweeted two black men, two prominent black men who criticized the constant focus on race in this
01:10:42.120 country. One was Jason Whitlock of outkick, right? Formerly ESPN, a journalist who's super smart.
01:10:48.600 He's coming on the show. He, I love Jason Whitlock. He's been brilliant on these issues and really
01:10:52.540 brave. And he's been called an uncle Tom by everybody. And one, um, Leonidas Johnson, who's
01:10:57.240 got his own podcast. So now it's to the point where even retweeting these, you know, black men with 0.55
01:11:05.140 heterodox views of this race dogma is problematic, right? That's potentially cancelable, whatever
01:11:12.660 they'll, they'll do what they're going to do. But I think it's, no one wants to hear from them.
01:11:17.720 Why isn't Coleman Hughes a household name? If he were saying the stuff the left wants to hear,
01:11:24.420 he would be. Yeah. Uh, well, it's a problem. Uh, you know, I'll, I'll grant you that it's been,
01:11:31.020 it's been a problem for a long time on some adjacent issues. I mean, this is what happened
01:11:36.680 to, to Ayaan Hirsi Ali, who I, I think you probably know. Um, she's my friend and she was on the show
01:11:42.440 recently. Yeah. Who, I mean, she's a dear friend of mine and, you know, so, you know, when she was,
01:11:49.220 you know, speaking critically about the treatment of women under Islam and you'd think she would have 1.00
01:11:55.820 standing to do this, having, you know, come out of Somalia and, and suffered, you know, all of the,
01:12:01.740 the, um, uh, uh, the collateral damage of that experience that you might expect. And then,
01:12:09.520 uh, literally recapitulating the entire enlightenment project in her own life and becoming a, a, uh,
01:12:19.020 member of parliament in Holland. And then, you know, being, uh, you know, hunted by jihadists and
01:12:24.360 theocrats and, and, you know, essentially becoming the next Salman Rushdie, uh, the, uh, the left
01:12:33.280 didn't want to hear from her, right? I mean, she's, she was much, a much better candidate for, for
01:12:39.360 taking a position in a, a left wing think tank than a right wing one, but only the AEI, the American
01:12:46.740 Enterprise Institute would give her a perch, uh, when she really needed one. Right. And I, you know,
01:12:53.120 while I don't agree with, uh, everything that comes out of that organization, I, you know,
01:12:57.860 never, uh, cease to be grateful to, to them for doing that. Um, and, you know, so she had this
01:13:05.960 experience that many people have had where, when you begin making sense on one of these issues,
01:13:12.820 you know, in her case, the, the, the problem of, of Islamist theocracy, um, and, uh, it, it becomes,
01:13:22.540 um, radioactive enough on the left that you, you have a, it's a, it's a very, it's a disorienting
01:13:31.500 social experience. You, you, you, what you, what shows up in your inbox is, uh, you know, utterly
01:13:39.320 disparaging and crazy and bad faith attacks from the left. And on the right, you kind of get love
01:13:49.020 bombed by a cult, right? I mean, you just, you meet, you know, really friendly people who, who
01:13:54.320 don't agree with much or, uh, of the rest of what you may believe. Right. And so, you know, people on
01:14:02.180 the right, when they, when they hear me, uh, criticize Islam and his connection to jihadism
01:14:08.720 and terrorism, you know, very following very much the line that, that, that Ayan would, would 0.95
01:14:15.800 follow here. Um, people on the right are, are so happy to have, to have someone, you know, left of
01:14:24.360 center making sense on this issue that it's, you know, it really is just a completely congenial
01:14:31.200 meeting of the minds, despite all of the other things I believe and will argue for that they
01:14:37.120 find just odious. Right. And then you're like, surprise, surprise. Yeah. Well, yeah. So there's
01:14:42.900 that, I mean, that, I don't know if you ever saw my initial, my, my interviews with Bill O'Reilly,
01:14:48.160 but they, they all went this way where I would say something about Islam and it was a, you know, 0.96
01:14:52.180 a perfect meeting of the minds. And then I would switch to Christianity and, and, you know, 0.96
01:14:56.280 it was basically at the end of the segment. Um, so that's all the time we have. Yeah, exactly.
01:15:02.960 But, but the, but the truth is the, the, the worst, um, the most dishonest, the most hostile,
01:15:12.280 the most gaslighting, there's the, the most insufferable attacks tend to come from the left.
01:15:19.780 Right. And it's now there's an asymmetry here or kind of an optical illusion perhaps, because I'm
01:15:26.880 not, you know, I'm not dealing with the far right. I'm not talking to neo-Nazis and antisemites. Right.
01:15:32.640 So obviously what, what they would have to say to me would be, you know, just as despicable and,
01:15:38.160 and, and dishonest, I'm sure in the end, but, um, but the truth is I'm not even sure given how crazy
01:15:45.160 the far left has been. And, you know, I on experienced this and, and so many people have
01:15:50.960 experienced this and what many people have, have suffered, I would say, uh, you know, I've, I've
01:15:58.060 certainly resisted this, I think successfully, but not everyone has, there's kind of a tractor beam
01:16:03.640 effect where when you're getting nothing but disingenuous, uh, you know, nausea inducing craziness on
01:16:12.460 the left and the right is, is showing itself willing to just bury the hatchet again and again
01:16:19.060 and again, and we can agree to disagree about all these other things, but you know, we're nice guys
01:16:23.640 over here. Um, you see people get pulled to, into the getting, getting captured by a new audience.
01:16:31.620 And, you know, I'm not going to name names here, but there, you know, we have mutual friends who,
01:16:37.060 who I think just now can't really make a lot of sense when talking about Trump and, and the election
01:16:44.800 say, because they have been captured by a right-wing audience that really treated them well,
01:16:51.680 you know, when the left treated them just despicable. Um, and, and it's always, it's a kind
01:16:57.540 of social and psychological experiment. And it's, um, you know, I, you know, I think it's something to be
01:17:03.060 on guard for, um, not for political reasons, but just for reasons of, of, um, you know, making sense
01:17:10.480 on these issues. We, I know, I like what you're saying. I feel like this is illuminating. This is
01:17:15.380 illuminating because I, I know on the subject of religion, you're, you're not a subscriber. I know
01:17:21.340 you don't, you don't like the term atheist because you don't need to be called that just because you
01:17:25.460 don't happen to subscribe to religiosity, um, in any form, but it's really a rejection of dogma.
01:17:32.360 That's what you're, you're, you're a rejecter of dogma. And I, I actually do like that. I feel
01:17:37.100 like I too am a rejecter of dogma though. I am. I don't think I can say a practicing Catholic,
01:17:43.700 but I'm Catholic. I do believe in God. I don't believe in every story in the Bible. I'm trying
01:17:49.200 to raise my kids Catholic, but I don't, I don't really subscribe to the dogmatic religious thinking.
01:17:54.320 I kind of take what I want from it and use it to reinforce moral and ethical principle
01:18:00.180 principles. I believe in, I use God to threaten them, which works brilliantly. And, um, that's
01:18:07.300 sort of where I am, but I politically am very reticent to sign on for anybody's dogmatic thinking.
01:18:13.960 And I think that's an advantage to me as a journalist, but I, that's why I'm a registered
01:18:19.100 independent. I, I, why, why should I sign on to some party and their platform that I'm invariably
01:18:26.620 going to have many disagreements with nobody out there has got exactly my ideological outlook.
01:18:31.600 And why, why would I just put on their team Jersey? I'm talking about as a citizen. Now I just say
01:18:36.500 like, I I'm going to support you. I think it's, I'm always surprised when someone says they agree
01:18:41.220 with everything on the, on the Republican platform or everything on the democratic platform. It's like
01:18:45.860 all of it. Did you look at it? Did you think about it for yourself? How did you get to that place?
01:18:50.800 And so like, I will confess that I, when I see, I mean, I can't even deal with the far left. I'm so
01:18:56.980 over them. I don't want, I really don't want anything to do with them that I have shut down
01:19:00.860 my willingness to converse with them. I don't think they're honest brokers. I don't think they're,
01:19:05.780 I don't think they're coming at it in good faith. That's what, that's what I want to say about it.
01:19:09.100 The left, I feel differently about liberals. I feel differently about, and I think Republicans,
01:19:13.760 you know, my experiences with them have been largely positive. And so I understand what you're
01:19:18.660 saying, the temptation to put on the Jersey, but I haven't, I, and I, I won't, it's just not the way
01:19:24.800 I'm built. And I'm, I'm more skeptical of these groups and parties, uh, than I am loyal to them.
01:19:31.680 And it's one of the reasons why I'm like a little concerned about, as we were, as I was saying before,
01:19:36.340 the deification of Trump, like I understand defending him and giving him a fair shake. And,
01:19:41.340 and I understand just thinking he's awesome, right? Like I get those people,
01:19:45.000 but what I see happening right now at some of these rallies where people are like, I will do
01:19:49.760 whatever my president tells me to do. I will do what Donald Trump, he's, he sacrificed everything
01:19:54.720 for me. I would die for him. I would die for him. People are saying that. And I don't totally
01:20:01.040 understand how they got there or where that means we're going. Yeah. Well, this is where it crosses over
01:20:09.080 into something like a political religion or a, a kind of pseudo spiritual awakening, right? It's,
01:20:17.540 it's just kind of a mass movement and it's happened on the left. And I would argue that what's happened
01:20:22.680 around BLM has that character as well, right? Like it's, it's just, it's not even trying to get in
01:20:30.340 touch with facts, right? It's, it just, it feels too good to be right about this particular thing
01:20:34.840 that you just don't want to, you know, you've achieved escape velocity somehow from, from the
01:20:41.380 normal constraints of, of public discourse. And you're just soaring above the earth. And that,
01:20:48.660 and that's happened in Trumpistan. Um, it's, uh, it's very strange. I mean, it, you know, it's worth 0.65
01:20:57.880 looking at the literature on cults to, to understand it. I mean, it's functioning by the same dynamic
01:21:04.760 the difference between, um, a cult and a, a religion is, is really just in numbers of subscribers
01:21:13.840 from, from my view. I mean, once you get a billion subscribers, well then it's simply indecent
01:21:19.040 to call it a cult. I mean, that's a pejorative term here, here you're talking about most of the
01:21:24.580 people in, in any given society. Um, but if there's, there's only 15 people in a house with,
01:21:30.340 you know, a lot of, you know, burning candles and they've got pictures on the mantle that no one
01:21:34.640 can recognize, um, well then that's a cult and what the hell are you people up to? Uh,
01:21:40.120 and what are you teaching your kids and, and all the rest. And so that, but if you really
01:21:44.960 want to have an honest conversation about the way the world is and how we should all live
01:21:51.520 together within it so as to stand a chance of maximizing human wellbeing or escaping the,
01:21:58.600 the worst possible outcomes that are, that are in fact possible, well then you, you need to appeal
01:22:04.280 to something deeper and something universalizable, right? Something that isn't born of the mere
01:22:09.340 accidents of, of birth or geography or, you know, who you're, you know, what religion your parents
01:22:15.000 happened to, to have, or, you know, what, what politics they happened to have. And we know many
01:22:20.080 people inherit their politics very much like a religion, right? You tend to just be following the line
01:22:25.620 of your parents. Um, and it is, yeah, it's weird. I mean, to come back to the point you made about
01:22:32.860 platforms, it's weird that if you know someone's position on gun control, you know, you stand a good
01:22:40.420 chance of knowing their position on climate change, right? Or on, on a dozen other things that should be
01:22:47.280 unrelated to, and so it's, um, it's a sign that people aren't thinking these, these problems through
01:22:56.120 based on first principles. They're, they're joining a team. They're, they're joining a religion. They're,
01:23:01.280 they've, they're, they're part of a social experiment on some level. It feels good to, to be part of a team.
01:23:08.140 Yeah. It's tribalism. I mean, we, we have, we're deeply tribal and we, um, you know, we're, we're
01:23:16.480 apes in that regard. And we're trying to, we're trying to leave the monkey behind here. And again,
01:23:23.660 all we have is conversation by which to do it. Now, I want to pick up on something you said about
01:23:29.800 going forward in life and being focused on, you know, what matters and who we are as, as human
01:23:34.520 beings. You are, you're deep into meditation. You've studied it for years. You've practiced it
01:23:41.880 for years. You've read all the books you've spoken with all the gurus. And, but the thing that stood
01:23:47.380 out to me and just reading up about this piece of your life was, um, you, I read, this is a quote
01:23:53.500 from you. I've gone into silence for a week and meditated 18 hours a day just to see what can be
01:24:00.200 revealed through disciplined use of attention through introspection and to see how it can inform
01:24:07.240 the study of the mind. And then you didn't say anything more after that. And as somebody who
01:24:15.520 doesn't meditate, I was wondering, could you like short form it for those of us who didn't do that?
01:24:21.800 That seems like an important thing to know. Um, well, yeah, so I, I've spoken a lot about
01:24:26.440 meditation in, um, I wrote a book, uh, it appears in several of my books, but I wrote a book on the
01:24:33.540 topic called waking up. And I have a, an app by that title where I, where I, I and other people
01:24:39.080 teach various techniques of meditation and talk about its, its connection to understanding the mind
01:24:45.260 scientifically and, and, and just living an examined life altogether. I mean, just, just kind of
01:24:50.760 rebooting the, the ancient philosophical project of, of developing a philosophy of life that actually
01:24:57.980 matters, right. That actually changes one's moment to moment engagement with the world and aligning
01:25:03.900 one's ethics and one's emotional life and, and really trying to, to live a life that you don't
01:25:12.080 regret in the end. I mean, you don't regret at the end of any given day or a given hour, but you don't
01:25:17.700 regret at the end of your life. And what would it mean to, to, to do that? And, and how can we do
01:25:22.760 that? Um, so that's, that's really the kind of the center of gravity of, of my, my interest at this
01:25:28.400 point. But yeah, with respect to sitting silent retreats, yeah, I did a lot of that, you know,
01:25:36.500 mostly in my twenties, I spent about two years on silent meditation retreats and, um, the longest
01:25:42.840 one was three months. Um, and I did a couple of those and then, you know, several, you know,
01:25:48.800 one month and two month retreats. And, and then yes, you know, you can't talk at all.
01:25:54.080 No, it was on some retreats. You, you have an interview with a teacher every other day for
01:26:00.320 about 10 minutes. So you, you, you, they just need to check in on you and guide your practice and
01:26:05.920 make sure you're not losing your mind. Um, as some small percentage of people do under those
01:26:11.120 conditions as you, as you might imagine. And, and, uh, but basically it's, it is silence
01:26:16.320 and it's, you're not even making eye contact with people. I mean, you're really, you really
01:26:20.220 are kind of locked down. If you have no perspective on the nature of, of mind, uh, prior to concepts,
01:26:31.160 prior to your, your thinking incessantly about everything, uh, prior to the conversation you're
01:26:37.280 having with yourself, uh, you are a mere hostage of that conversation. And it's an, it's an amazingly
01:26:45.640 distorted conversation. I mean, you'll, you'll tell yourself the same thing 10 times in a row
01:26:51.180 and never get bored, right? If someone walked into the room and spoke the same sentence to you over
01:26:56.540 and over again, you know, you, one, you'd think they were crazy and two, you would get out of the
01:27:02.340 room, right? You'd say this, this is, this is not worth my time. But when you look at the kinds of
01:27:08.000 things you will tell yourself and, you know, you know, every hour on the hour, every minute on the
01:27:14.880 minute, right? When you're perseverating on something, when you're trying, when you're, when
01:27:18.060 you're really caught by something, um, it is, it is a psychotic dream really. I mean, the difference
01:27:26.320 between you and a psychotic in that case is, is, is that you have the good sense to keep your mouth
01:27:30.600 shut? And the psychotic is verbalizing everything, you know, out on the sidewalk, but that's basically
01:27:36.340 the difference, right? You're, you know, we, if you could just imagine your thoughts broadcast on a
01:27:40.840 loudspeaker every moment of the day for all to hear, you know, we'd, we'd all sound crazy under
01:27:46.020 those conditions. And meditation is a technique for recognizing the mechanics of all that and
01:27:53.180 relinquishing it, you know, if, if only for moments at a time. And as you get better at it,
01:27:58.200 you can, you can, you know, get off the train for longer. And what you discover when you do that is
01:28:04.120 that the mind is a, is, is the basis for, for all the wellbeing you have ever experienced in your
01:28:11.260 life. You know, it really is, there's an intrinsic quality to consciousness, you know, before anything
01:28:19.060 changes, you know, in the very midst of any ordinary experience, you know, before you, you know,
01:28:24.520 before, before the pain in your knee goes away, I mean, even in the midst of an unpleasant experience,
01:28:30.380 there is, there is a real freedom, right? A real, a real sense of, of, you know, compassion for
01:28:39.720 yourself and for others and, and just a radical openness, right? And, and I mean, this is something
01:28:47.400 that many people experience first, you know, taking, you know, one or another psychedelic, right? And this
01:28:53.460 is what happened to me, you know, when I was 18, I took MDMA. Um, and I had to look that up. That's
01:29:00.040 ecstasy. Yeah. Yeah. And, you know, and ecstasy, you know, since it became very popularized as a club
01:29:08.040 drug and people took it in, in comparatively, you know, frivolous ways. And it's not to say they
01:29:13.660 didn't have a lot of fun in the process, but, you know, originally it was, it was, you know, designed
01:29:19.820 and, and taken by people, you know, very much with the intention of discovering something about the
01:29:25.460 nature of their minds. And, and that was really the, the, the framing I had when I, when I took it. And
01:29:31.080 yeah, I discovered that it was possible to be much less of an asshole than I was tending to be,
01:29:38.460 right. I mean, it was possible to be deeply at peace with myself and the world and, uh, and to be
01:29:48.200 happy, right. Like really, really happy down to my toes. And then, and then I think the H word
01:29:53.920 and, and, and, and then, and then you, you know, then you lose that and then you become interested in
01:29:59.700 why, right. Like what is, what is it about how I'm using my attention that reliably produces
01:30:07.700 something less than the, the deepest peace and satisfaction and love and connection that I've
01:30:14.960 ever experienced, right. How am I, how am I failing to actualize that day after day? And that's when
01:30:21.680 a practice like meditation becomes relevant to people. You, can you get yourself there? Can you get
01:30:27.980 yourself to the ecstasy version of yourself through meditation? Yeah, yeah, definitely. Although
01:30:35.820 that isn't actually the, the primary goal is the primary goal of, of certain types of meditation.
01:30:43.320 And so in a Buddhist context, there's a meditation called Metta, which is the Pali word for loving
01:30:49.800 kindness. And, and in that practice, you, you are trying to create a specific state of mind,
01:30:55.600 very much like that, you know, what many people have experienced on, on ecstasy, which is, you know,
01:31:02.640 unconditional love for lack of a better word. But I mean, it really is that it's, it's, you recognize
01:31:08.800 that love is a, a state of being that you can fall into more and more deeply. And it really just,
01:31:18.960 it's not, it's not transactional. It's not like you, you love someone because, you know, you,
01:31:24.060 because of your history with them, because of all the good things they did for you because of,
01:31:28.540 you know, because of how much fun you have in their presence. No, it's, you can actually recognize
01:31:33.640 that you really want other people to be happy. You, you really want them to be free of suffering
01:31:40.840 and the depth of that wanting, the depth of that commitment to the wellbeing of other people,
01:31:47.720 even people you've never met, right? Even people who are your enemies, who, who are working hard to
01:31:53.160 make themselves your enemies, right? I mean, you, you, you can stand back from your kind of the, 1.00
01:31:59.160 you, your, your reactivity and your kind of the, the, the personal aspect of those collisions
01:32:06.700 and recognize that on some basic level, everyone is suffering. Everyone is going to lose everything
01:32:14.800 they love in this world. You know, everyone is, is we're all in this astonishing circumstance
01:32:21.520 together. And what you want, even for the bad people is an end to suffering. I mean, you want
01:32:31.540 people to be happy. And, and that is the, that wanting is a state of mind that you can focus on.
01:32:40.860 And it's, and so metta practice is the, is the, is the practice of, of, of amplifying that intention
01:32:46.720 and emotion to the point where, yeah, it, it just obliterates everything else in your mind for the
01:32:51.960 time that you're doing that practice. So you just, you just feel a depth of love for everyone,
01:32:58.560 you know, for, for no reason, other than the fact that that's, that's what you feel for them,
01:33:04.300 right? You actually just, because no matter how bad they are, and it's, I mean, it's, this may
01:33:09.040 sound, you know, bizarre or, or, or its own, in its own way, pathological to people, but I mean,
01:33:15.940 just take, you know, take one of the worst people who's ever lived. Right. And so there are many
01:33:21.480 people on this list. Um, you know, I wouldn't put Trump on this list has, it might surprise some
01:33:26.640 people, but, um, you take, I was going to ask you whether you've meditated on him. Yeah. I mean,
01:33:31.420 so he's, he's, he's harder than most people to feel compassion for, for reasons that are
01:33:36.300 interesting, but, um, let me take someone like Hitler. Yeah. Yeah. No, I, my, my favorite for 0.96
01:33:42.180 this, this use case is, um, Uday Hussein. Right. So one, one of Saddam Hussein's sons, you know,
01:33:48.080 his worst son. Right. So this, I mean, he is just the prototypical evil person, right? He's no
01:33:54.840 question. He was a psychopath. And this was a guy who, when he was driving through Baghdad with his
01:34:00.320 bodyguards and he would see a wedding in progress, they would descend on the wedding and he would rape
01:34:05.900 the bride. And in certain cases he killed the bride. I mean, he did this more than one case.
01:34:11.080 Right. And it's just, just the most despicable human being you can imagine. Right. So how could
01:34:16.740 you feel love or compassion for Uday Hussein? Uh, well, just look at his, his lifeline as a whole,
01:34:26.940 right. Just to just roll back the clock on him and think of him as the four-year-old Uday Hussein.
01:34:32.840 Right. So how do you feel about the four-year-old? Right. Well, he's, he, I mean, he may have been a
01:34:38.360 psychotic kid too. I don't know. He may have been a scary little boy. I wouldn't, I wouldn't doubt it,
01:34:43.580 but he was above all a really unlucky one. Right. He, first of all, he had Saddam Hussein as a father, 0.89
01:34:51.380 right. I mean, how unlucky can you get? Um, and he was the four-year-old boy who was on track to become
01:34:59.520 the very scary man who, you know, we wound up killing, you know, happily. That's, you know,
01:35:06.400 that's exactly what we should have done with him, uh, given who he was and given the fact that we
01:35:11.400 couldn't capture him. But, um, it is appropriate to feel compassion for the four-year-old boy who
01:35:20.540 became Uday Hussein. I mean, that is an unlucky life for, through no fault of his own. He didn't pick 0.94
01:35:26.040 his genes. He didn't pick his parents. He didn't pick the, he didn't, he didn't decide to be born
01:35:32.220 into a war-torn honor-based society that would, would amplify all of his flaws. Right. Uh, and so
01:35:39.140 you can feel compassion for that boy. And then the question is, at what point is it illegitimate to
01:35:46.000 feel compassion for him? Right. When he's five, when he's six, when he's seven, when he's eight,
01:35:50.020 like, when does he cross over into no longer being an appropriate target for your, your, your
01:35:58.800 well-intentioned wish that he'd just be happy, that he just overcome suffering. And you can get
01:36:05.720 there. I mean, you can really get there with the worst people. Um, but that actually is not the,
01:36:11.000 the center of the bullseye as far as I'm concerned for meditation in general. I mean,
01:36:16.040 in general meditation is not about producing specific states of mind, like loving kindness is
01:36:23.740 about recognizing that, that ordinary consciousness. I'm just, just the, the consciousness that is,
01:36:29.960 that is hearing my words right now. I'm just the consciousness that's allowing the two of us to
01:36:34.420 have a conversation, um, is already free of, of self. Sam Harris, living and
01:36:46.020 examined life. I love that. That's, that's inspirational to me. It does. And, and see where
01:36:51.820 that goes and see how it makes you feel and get really honest about the answers to both of those
01:36:58.940 things that I think I can do. Today's episode was brought to you in part by Jan Marini skin research,
01:37:05.640 dramatic results. Dermatologist recommended. Get your award-winning skincare system now
01:37:10.880 at Jan Marini.com. Want to tell everybody that, uh, next show, which is on Wednesday is going to be
01:37:18.100 with two people who are spectacular. Andy McCarthy of national review, who is the one lawyer who's been
01:37:24.940 super smart on all the Trump legal challenges, really fair to the president too. Unlike most
01:37:29.780 and Selena Zito, who I've corresponded with a bunch online, but I've never actually met her. And, uh,
01:37:35.120 she's somebody who I love her voice and I love the angles she pursues on stories. She's one of those
01:37:39.960 folks who gets fly over country and isn't disdainful of it. So I think those are two great
01:37:44.240 people to talk to. Well, it'll be, you know, our first show after the electoral college meets and
01:37:48.980 votes today and we'll get their take on where we are and what's going to happen, you know, between now
01:37:54.060 and January 20th. Sound counsel and thoughts from two smart, likable people. That's the kind of show I
01:38:01.120 love. If you don't want to miss it, go over there and subscribe. Make sure you're a subscriber. That helps
01:38:05.900 that, that lets me come right into your inbox in the morning, the top of your phone saying, Hey,
01:38:10.640 don't forget me today. Come listen to me. Talk to Selena. Um, and then of course you've got to
01:38:15.720 download and you've got a rate five stars and more than anything, send me a review. Will you?
01:38:19.900 It's been fun to read them. People make me laugh. Some people swear. Some people ramble on
01:38:24.820 some people write weird sexual things. Don't do that. But I do like hearing from you and, uh,
01:38:33.420 any guest ideas are good too. A lot of new names that I hadn't even heard of. And then I send my
01:38:38.140 team to go Google and find and, and call in some circumstances. So anyway, it's been a pleasure
01:38:43.660 and to be continued after we have results from the electoral college. Talk to you Wednesday.
01:38:49.640 Thanks for listening to the Megan Kelly show. No BS, no agenda, and no fear.
01:38:54.820 The Megan Kelly show is a devil may care media production in collaboration with Red Seat Ventures. 0.56