Real Coffee with Scott Adams - May 01, 2022


Episode 1730 Scott Adams: The Golden Age Is Upon Us, Trump Was Right Per Chomsky. Wow. What A Show


Episode Stats

Length

1 hour and 12 minutes

Words per Minute

147.68286

Word Count

10,768

Sentence Count

751

Misogynist Sentences

12

Hate Speech Sentences

32


Summary


Transcript

00:00:00.000 And that concludes the ASMR portion of our program.
00:00:07.280 The rest of it will be such high excitement that if you tried to sleep to this,
00:00:12.120 my God, I pity you.
00:00:14.000 The excitement that you are about to experience,
00:00:16.080 the full body pleasure, the dopamine rush,
00:00:20.300 unbelievable.
00:00:22.380 Usually the show is great, I have to say,
00:00:25.440 but today, it's other-level greatness.
00:00:30.420 The greatest greatness there's ever been.
00:00:31.920 There will be new concepts, new ideas, shocking twists and turns.
00:00:36.720 The sort of thing no one's ever seen before.
00:00:40.220 But can you enjoy this without being properly primed?
00:00:44.700 Of course you can.
00:00:45.720 But you can enjoy it more,
00:00:47.360 and all you need is a copper mug or a glass,
00:00:49.400 a tank or a gel, a canteen jug or a flask,
00:00:51.240 a vessel of any kind,
00:00:53.980 filling with your favorite liquid.
00:00:57.680 Gosh darn it, I like coffee.
00:01:00.480 And join me now for the unparalleled pleasure.
00:01:04.480 It's the dopamine hit of the day.
00:01:08.180 It's the thing that makes everything better,
00:01:11.000 and it'll fix everything in your day.
00:01:14.240 Go.
00:01:14.460 Go.
00:01:18.380 Ah.
00:01:21.360 Did you see that in the news,
00:01:23.980 was it a jeet pie?
00:01:29.880 Was getting some grief for having a gigantic coffee mug?
00:01:35.920 To which I say,
00:01:37.940 the more gigantic, the better.
00:01:40.400 Well, let's talk about all the news,
00:01:42.140 but before we do that,
00:01:43.740 special hello to Susan.
00:01:47.080 Susan at the Moss Beach Distillery.
00:01:49.100 And you, too, someday could get a special hello.
00:01:55.060 All you have to do is ask for it.
00:02:00.640 Turns out it isn't that hard after all.
00:02:03.220 But hello to Susan.
00:02:04.280 Everybody say hi to Susan.
00:02:06.580 And the golden age is upon us.
00:02:10.900 I don't know if you've noticed.
00:02:11.840 Let me start out with a bang.
00:02:15.980 Have you noticed
00:02:17.060 that
00:02:18.880 all of the biggest problems in the world
00:02:22.820 are heading toward a solution at the same time?
00:02:27.060 It's one of those things you don't notice,
00:02:28.800 because we always talk about the trouble.
00:02:30.980 So you don't notice anything positive.
00:02:33.620 What the hell?
00:02:34.320 Well, apparently something has exploded on me.
00:02:40.240 Okay?
00:02:40.880 I don't know what that is.
00:02:44.020 So here's some examples.
00:02:45.960 So we've got the pandemic is winding down.
00:02:49.440 That's good.
00:02:50.360 And we probably learned a whole bunch
00:02:52.120 about avoiding future pandemics
00:02:53.760 or minimizing them
00:02:54.780 or what to do and what not to do.
00:02:57.180 Wouldn't you say?
00:02:58.800 I would say so.
00:03:00.000 I'd say the whole pandemic thing,
00:03:01.920 as horrible as it has been,
00:03:03.160 and, you know,
00:03:04.000 there'll be lasting negative consequences.
00:03:07.020 But I think as a civilization,
00:03:09.140 we're stronger than we've ever been.
00:03:12.600 Because we know how to take care of the next one
00:03:14.760 way better
00:03:15.560 than we were prepared for this one.
00:03:18.760 How about inflation?
00:03:20.360 Inflation's out of control.
00:03:21.820 Gas prices are up, right?
00:03:23.020 So that's the way the news handles it.
00:03:24.680 It's all true.
00:03:26.280 But there are smart economists saying today,
00:03:30.060 according to the news,
00:03:30.940 that they believe we may have peaked.
00:03:34.180 That we may be at the top of the inflation.
00:03:37.360 And that you would see the,
00:03:38.540 what they call the flexible part
00:03:40.280 of the inflation coming down rapidly
00:03:42.300 in the summer.
00:03:44.100 There's still a sticky part
00:03:45.440 that'll take a while to work out.
00:03:47.420 But it's not the big part.
00:03:49.680 So we could be at the peak inflation
00:03:51.640 right now.
00:03:53.080 So the next month or so,
00:03:55.200 you might see it stabilize
00:03:56.160 or start to go down.
00:03:58.420 How about climate change
00:04:00.600 and, you know,
00:04:02.200 the catastrophe ahead?
00:04:04.220 Well, nuclear power
00:04:06.600 and nuclear energy
00:04:08.360 is going to be the solution to that.
00:04:10.220 And now it's a bipartisan agreement.
00:04:13.980 Democrats and Republicans
00:04:15.220 both say by majority,
00:04:17.640 yes, do nuclear, we need it.
00:04:19.260 We're starting to keep older plants operational.
00:04:24.040 There's a budget for that.
00:04:25.040 Biden's doing that.
00:04:26.620 And there's a fusion
00:04:27.740 that's starting to become practical.
00:04:30.100 Maybe in the next 10 years,
00:04:31.280 we'll have a fusion plant.
00:04:34.380 We've got these small Rolls-Royce type,
00:04:36.480 actually Rolls-Royce is a company
00:04:37.980 making actual nuclear reactors.
00:04:40.560 And they're making small factory-made modular units.
00:04:44.020 So they can be sort of approved once
00:04:46.420 and then you can implement them many times.
00:04:48.380 Cost-effectively.
00:04:50.560 Smaller size, already approved,
00:04:53.080 no new engineering,
00:04:54.400 just slapped together a bunch of them.
00:04:56.220 So basically the economics
00:04:57.540 and also the risk of handling the waste,
00:05:00.940 these problems are all solved.
00:05:03.220 Or they're right on the border of solved.
00:05:05.060 You know, you've got to work out a few kinks.
00:05:07.520 But they're all so solvable
00:05:09.080 with what we know how to do already,
00:05:12.640 you're really sort of at the worst of it
00:05:14.560 and it looks like it's going to turn
00:05:15.840 in the other direction very quickly.
00:05:17.200 What about free speech
00:05:19.420 and all the fake news?
00:05:21.620 That's the biggest problem in the world,
00:05:24.380 in my opinion.
00:05:25.780 Or maybe the highest priority to fix.
00:05:28.500 Because if you don't get the free speech fixed,
00:05:31.160 then everything else breaks.
00:05:33.480 But enter Elon Musk buying Twitter.
00:05:37.720 Now I'm not going to tell you that I know
00:05:39.360 that having one billionaire in charge
00:05:42.620 of such an important lever on free speech,
00:05:46.700 I'm not going to say that necessarily
00:05:48.660 that's going to be solved.
00:05:51.240 But it sure looks like it.
00:05:52.960 If I had to bet,
00:05:54.480 I would put a very large bet,
00:05:56.560 a very large bet.
00:05:58.520 I would easily bet a million dollars
00:06:00.800 that free speech will look better
00:06:02.980 in a year than it looks now.
00:06:04.280 specifically in the Twitter lever situation.
00:06:09.200 I would bet a million dollars that's true.
00:06:11.660 Would you take the other side of that bet?
00:06:14.100 I mean, anything's possible, right?
00:06:16.380 You know, Elon could turn out to be
00:06:17.940 an evil person who's been hiding it
00:06:20.080 for years or something.
00:06:21.000 But I'd bet a million dollars,
00:06:25.160 and I'd feel comfortable with that bet,
00:06:26.800 that if there were any way to measure it,
00:06:29.220 that we'd be way ahead in free speech.
00:06:31.920 And that, because it's Twitter,
00:06:33.900 and that influences the whole chain of news
00:06:37.740 and everything else,
00:06:39.600 that's heading in exactly the direction
00:06:42.160 I'd want it to head.
00:06:44.060 How about the experts being defanged?
00:06:47.580 Is that good or bad?
00:06:48.520 That our understanding of what experts
00:06:51.000 can and cannot do for us
00:06:52.420 is completely altered.
00:06:54.560 Is that good or bad?
00:06:56.920 It's good.
00:06:58.340 Because we're less susceptible to bullshit.
00:07:02.860 And then you add on top of that
00:07:04.320 that free speech may be coming back,
00:07:06.740 and that's starting to look good.
00:07:08.880 So our ability to communicate instantly
00:07:11.340 across the globe is in pretty good shape.
00:07:14.200 You know, the Internet is just amazing.
00:07:16.920 Everybody can talk to everybody.
00:07:18.200 That was a big thing for the pandemic, especially.
00:07:22.040 But if we're trying to figure out
00:07:24.420 what to do about the next emergency,
00:07:26.940 we really have to understand
00:07:28.520 the limits of our experts
00:07:29.880 and understand what they can and cannot do
00:07:32.420 in the fog of war.
00:07:34.720 And we should also be more forgiving
00:07:36.220 when they get it wrong.
00:07:39.200 So we should be improving in both directions,
00:07:41.780 and I think we will.
00:07:43.120 I think we should be more forgiving
00:07:44.560 in the fog of war,
00:07:45.760 and we should be even, let's say,
00:07:49.360 less forgiving
00:07:50.340 and more skeptical
00:07:52.140 when things are settled down, right?
00:07:55.020 It would be good to go hard
00:07:56.540 in both directions at the same time.
00:07:59.120 More skepticism,
00:08:00.700 but also more forgiveness.
00:08:02.040 I think that's happening.
00:08:06.160 What about, you know,
00:08:07.320 Ukraine and Russia
00:08:08.540 could turn into a nuclear confrontation,
00:08:11.640 but I don't think people expect it.
00:08:14.380 I don't expect it.
00:08:15.880 So I don't think that's going to happen.
00:08:17.620 You have to worry about it.
00:08:18.640 But I think the most likely outcome
00:08:22.260 would be a permanent understanding
00:08:25.640 that you can't attack your neighbor anymore
00:08:27.580 if you're a certain type of country.
00:08:31.540 You know, maybe far less industrialized countries
00:08:35.040 that nobody's paying attention to
00:08:37.080 can do some bad stuff to each other.
00:08:39.780 That'll last for a long time.
00:08:41.140 But in terms of a tank war,
00:08:44.520 I think this is the last tank war.
00:08:48.980 Does anybody disagree?
00:08:51.800 I think this is the last tank war
00:08:54.480 where you invade your neighbors
00:08:56.580 with heavy equipment.
00:08:58.200 I just don't see it happening again after this.
00:09:01.100 It's just so obvious
00:09:02.360 that the defensive weapons
00:09:03.680 are better than the offensive weapons
00:09:05.200 in this situation.
00:09:07.160 Somebody says no.
00:09:08.320 Maybe they'll be better tanks.
00:09:09.440 Who knows?
00:09:10.580 But I think that war,
00:09:13.140 war in terms of the World War II style
00:09:15.880 that we're seeing, unfortunately, again,
00:09:18.220 I think this is the example
00:09:21.160 that will just seal it for everybody.
00:09:22.860 It's like, okay, these tank wars don't work.
00:09:25.760 Because even if Putin gets what he wants,
00:09:28.040 it's not going to look like it was a good idea
00:09:29.900 to everyone else, right?
00:09:32.520 Even if Putin convinces Russia
00:09:34.280 that it was a good idea
00:09:35.240 because he controls their information,
00:09:37.420 nobody else is going to think
00:09:39.340 it was a good idea,
00:09:40.320 even if he gets control of Ukraine.
00:09:42.420 It's just not going to look like a good idea.
00:09:44.540 So that's good.
00:09:46.980 I think the supply chain issues,
00:09:48.780 which are bad
00:09:50.000 and might worsen a little bit
00:09:51.220 in the next month or so,
00:09:52.620 these are the kinds of things
00:09:54.580 that humans fix really well.
00:09:57.200 If there's one thing
00:09:58.200 that you could depend on humans
00:09:59.860 to fix,
00:10:01.340 it's the supply chain.
00:10:02.560 Because there's so many people
00:10:04.460 who have vital interests
00:10:05.820 and it's the most important thing
00:10:07.520 in the world right now, actually.
00:10:09.460 And there's so many ways
00:10:10.620 you can communicate.
00:10:11.740 There's so many ways
00:10:12.440 you can alter transportation,
00:10:14.940 emergency resources, etc.
00:10:17.920 It just seems to me
00:10:19.300 that this is exactly the kind of thing
00:10:21.780 humans are good at.
00:10:22.860 I'm not too confident
00:10:27.980 that the poorest countries
00:10:29.540 are going to come through this okay.
00:10:31.760 Maybe that's going to get pretty dire.
00:10:34.620 But I think that
00:10:35.500 we're going to go through
00:10:37.340 an uncomfortable bump
00:10:38.960 no matter who you are
00:10:40.620 and then it will be better.
00:10:43.180 And it will probably be better forever
00:10:44.500 because we'll figure out
00:10:46.280 where all the weaknesses are
00:10:47.460 and then we'll have a workaround
00:10:49.600 for every weakness
00:10:50.440 in the future.
00:10:51.440 These are really, really big things
00:10:54.820 that look like
00:10:55.640 minor efficiency improvements.
00:10:58.180 But these are the things
00:10:59.000 that make civilization survive.
00:11:01.960 And these go right to
00:11:03.080 survival of humanity.
00:11:06.100 So I think the food issues
00:11:07.360 will be fixed
00:11:07.980 and we'll probably rethink
00:11:09.060 our entire
00:11:09.620 how do you create food,
00:11:11.220 how do you grow it,
00:11:11.900 how do you fertilize it.
00:11:13.220 The fertilizer thing
00:11:14.080 is a big problem, by the way.
00:11:15.840 That was a big, big problem.
00:11:18.040 But
00:11:18.340 I think we'll have time
00:11:20.820 to put enough human ingenuity
00:11:22.460 into it
00:11:22.980 to fix it.
00:11:24.400 Because remember,
00:11:25.040 there was a time
00:11:25.540 when we thought
00:11:26.060 we would run out of food
00:11:27.140 entirely
00:11:27.720 and then somebody
00:11:29.220 invented fertilizer.
00:11:31.960 That was,
00:11:32.620 that plus
00:11:33.380 I assume
00:11:35.760 irrigation methods
00:11:37.000 would be the two things
00:11:38.940 and then pesticides,
00:11:40.000 I guess the third thing.
00:11:41.100 So those are three things
00:11:42.280 that I don't think
00:11:43.060 anybody foresaw
00:11:44.540 the inventions of.
00:11:47.840 That was a sentence
00:11:48.820 I wish I'd bailed out of earlier.
00:11:51.820 Right?
00:11:52.820 So don't you think
00:11:53.580 that there will be
00:11:54.260 new ways to make food,
00:11:56.360 new technologies,
00:11:57.800 you know,
00:11:58.000 just as fertilizer,
00:12:00.440 pesticides,
00:12:00.820 and irrigation techniques
00:12:02.540 could not have been foreseen
00:12:04.200 at one point,
00:12:05.400 there must be things
00:12:06.380 we don't see coming.
00:12:07.220 And I'll bet
00:12:10.040 we'll be surprised
00:12:10.700 on the upside.
00:12:11.580 The only thing
00:12:12.140 that I think
00:12:12.640 is broken
00:12:13.120 with no plan
00:12:14.260 in place to fix it
00:12:15.380 is election credibility.
00:12:19.380 Wouldn't you say?
00:12:20.560 If you're going to look
00:12:21.260 at problems,
00:12:22.520 it's not,
00:12:23.600 you could argue
00:12:24.240 it's not the biggest problem.
00:12:26.440 You could make that argument.
00:12:28.780 But
00:12:29.180 it's the one
00:12:31.680 that doesn't seem
00:12:32.300 to have anybody
00:12:32.880 working on it.
00:12:35.960 Which makes it unique.
00:12:37.220 right?
00:12:38.140 Because the other ones,
00:12:39.020 everybody says,
00:12:39.620 okay,
00:12:39.880 we're trying to fix
00:12:40.580 climate change
00:12:41.860 or we're working hard
00:12:43.380 on the pandemic
00:12:44.020 and everything.
00:12:44.880 Everybody recognizes
00:12:45.960 those as problems.
00:12:47.700 But even though
00:12:48.520 the public
00:12:49.040 has great skepticism
00:12:50.760 about the credibility
00:12:51.640 of the election system,
00:12:53.300 you know,
00:12:53.460 just the ability
00:12:54.120 to give us
00:12:54.620 an accurate count,
00:12:56.420 I don't really see
00:12:57.760 anything happening
00:12:58.420 there, do you?
00:13:00.020 It's the biggest problem
00:13:01.860 that doesn't have
00:13:02.520 anything that looks
00:13:03.400 like a solution
00:13:04.100 percolating in any way.
00:13:07.220 But I think
00:13:08.740 we'll get there too.
00:13:10.300 So I think
00:13:10.820 you could add
00:13:11.240 to this list,
00:13:12.180 we're in this weird,
00:13:13.080 weird place
00:13:13.660 where we definitely
00:13:14.880 had,
00:13:17.160 I mean,
00:13:17.640 civilization
00:13:18.320 just had the shit
00:13:20.060 slapped out of it
00:13:20.940 between the pandemic
00:13:22.740 and now the war
00:13:23.920 and the supply chain
00:13:24.900 and every other
00:13:26.060 damn thing,
00:13:26.640 inflation.
00:13:27.820 And I think
00:13:28.980 we're going to
00:13:29.380 actually handle it.
00:13:30.180 And if we do,
00:13:31.960 we're in really
00:13:32.800 strong shape.
00:13:33.980 So I think
00:13:34.300 that's where
00:13:34.620 we're heading.
00:13:35.360 So that's my
00:13:35.820 positive for the day.
00:13:38.040 The White House
00:13:38.520 correspondence dinner
00:13:40.160 happened.
00:13:41.380 I can just tell you
00:13:42.480 some of the better jokes.
00:13:44.240 The only thing
00:13:44.940 that's worthwhile
00:13:45.640 from it.
00:13:47.200 You know,
00:13:47.980 when you see
00:13:48.840 the White House
00:13:49.520 correspondence dinner,
00:13:52.060 to me that feels
00:13:53.160 like they're pulling
00:13:53.840 the curtain back
00:13:55.960 and they're admitting
00:13:57.680 to the world,
00:13:58.980 you know this is
00:13:59.600 all theater, right?
00:14:01.360 Like all the things
00:14:02.300 we say to each other,
00:14:03.620 you know we're all
00:14:04.260 lying and acting
00:14:05.220 and then when we
00:14:06.880 do this thing,
00:14:07.460 we'll mock the fact
00:14:08.400 that any of it
00:14:09.660 is serious.
00:14:12.140 So in a way,
00:14:13.520 I would say that
00:14:14.620 the White House
00:14:15.100 correspondence dinner
00:14:16.200 greatly decreases
00:14:18.380 the credibility
00:14:19.200 of our system.
00:14:21.260 Does anybody
00:14:21.900 feel that?
00:14:23.640 I would be
00:14:24.880 way more comfortable
00:14:25.680 if the correspondence
00:14:27.400 and the government
00:14:28.580 did not get together
00:14:29.860 and make jokes
00:14:30.560 and pretend
00:14:31.900 it was all
00:14:32.520 just an act
00:14:33.540 because that's
00:14:36.080 what we're afraid of.
00:14:38.820 Isn't it?
00:14:40.360 Isn't the thing
00:14:41.240 you're most afraid of
00:14:42.220 that the government
00:14:42.900 and the journalists
00:14:44.480 are colluding
00:14:45.420 is literally
00:14:47.180 one of our
00:14:47.800 biggest fears.
00:14:49.440 And so these
00:14:50.820 fucking idiots
00:14:51.560 came up with
00:14:52.340 a festival
00:14:52.980 to celebrate
00:14:54.180 our greatest fear
00:14:55.300 and to put it
00:14:56.260 in our faces
00:14:56.800 with lots
00:14:57.360 of visuals
00:14:57.860 so you can't
00:14:59.020 miss it.
00:15:00.060 Hey, here's
00:15:00.460 your greatest fear
00:15:01.300 that the journalists
00:15:02.720 are all in on it
00:15:03.700 and they're basically
00:15:04.780 just good friends
00:15:05.960 with the elites
00:15:06.640 and when they
00:15:07.720 get together
00:15:08.620 they all laugh
00:15:09.300 about all this
00:15:09.940 bullshit that you
00:15:10.720 think is real.
00:15:13.340 I'm looking at
00:15:14.080 the comments
00:15:14.580 and some of you
00:15:15.340 feel the same way,
00:15:16.140 right?
00:15:16.680 To me,
00:15:17.260 this is a massive
00:15:18.180 mistake
00:15:18.900 and it's one
00:15:20.960 that Trump
00:15:21.980 got right,
00:15:23.700 didn't he?
00:15:25.320 Trump got this
00:15:26.320 completely right
00:15:27.380 by not going.
00:15:30.760 What do I keep
00:15:31.660 telling you?
00:15:32.220 That Trump
00:15:32.740 would look better
00:15:33.640 the longer he's
00:15:34.860 out of office,
00:15:35.920 the better he
00:15:36.760 would look.
00:15:37.940 You just see
00:15:38.500 an example
00:15:39.040 almost every day.
00:15:41.060 You know,
00:15:41.420 here's another one.
00:15:43.200 There's no doubt
00:15:44.120 that as a citizen
00:15:46.360 when I'm watching
00:15:47.380 this,
00:15:48.000 I feel creepy
00:15:49.080 about it.
00:15:49.880 Like,
00:15:50.260 oh,
00:15:50.420 this is creepy.
00:15:51.920 They're supposed
00:15:52.500 to be on different
00:15:53.240 sides.
00:15:54.280 Our whole system
00:15:55.280 depends on them
00:15:56.120 having a proper
00:15:58.200 relationship.
00:16:01.880 And I'm not sure
00:16:02.560 that joking about
00:16:03.500 how serious
00:16:04.200 any of this is
00:16:05.000 is a proper
00:16:05.540 relationship.
00:16:07.100 Is it?
00:16:09.560 Well,
00:16:10.320 yeah,
00:16:10.620 we'll get to
00:16:11.040 Chomsky.
00:16:12.220 so here's
00:16:14.100 some of the
00:16:14.360 good jokes.
00:16:16.340 And these are
00:16:17.100 Biden's jokes.
00:16:18.020 So he has,
00:16:18.360 he has,
00:16:18.800 let me,
00:16:19.720 let me give a
00:16:20.200 shout out to
00:16:20.780 his joke writer.
00:16:22.480 Whoever helped
00:16:23.340 Biden write his
00:16:24.320 jokes,
00:16:25.080 pretty good.
00:16:25.920 Because they
00:16:26.520 weren't,
00:16:26.880 they weren't too
00:16:27.540 edgy.
00:16:28.260 They were sort
00:16:29.100 of just right
00:16:29.700 for the situation.
00:16:30.740 And some of
00:16:31.800 them were pretty
00:16:32.240 good.
00:16:32.620 Here's some.
00:16:33.940 Biden said,
00:16:34.780 Republicans seem
00:16:35.540 to support one
00:16:36.340 fellow,
00:16:37.380 Biden said,
00:16:38.400 some guy named
00:16:39.320 Brandon.
00:16:40.480 He's having a
00:16:41.100 really good year.
00:16:41.780 And I'm kind
00:16:42.420 of happy for
00:16:43.020 him.
00:16:44.480 That's a good
00:16:45.200 joke.
00:16:46.220 All right,
00:16:46.500 here's another
00:16:46.900 one.
00:16:48.960 He thanked
00:16:49.740 the 42% who
00:16:50.920 actually applauded
00:16:51.820 as he took
00:16:52.260 the microphone.
00:16:53.100 So self-deprecating.
00:16:55.020 But he self-deprecated
00:16:56.180 in a clever way.
00:16:57.720 I think he picked
00:16:58.600 the best poll
00:17:00.140 number of his
00:17:01.080 bad polls.
00:17:02.660 Because I think
00:17:03.420 it's lower than
00:17:04.080 42%
00:17:05.480 favorability,
00:17:06.440 isn't it?
00:17:06.900 Depending on
00:17:07.960 which poll.
00:17:11.220 So in a clever
00:17:12.720 way, he
00:17:13.200 self-deprecates.
00:17:14.660 But he actually
00:17:15.240 gave himself a
00:17:16.000 promotion, even
00:17:17.360 in the self-deprecation,
00:17:18.700 I think.
00:17:19.660 Yeah, I think it's
00:17:20.240 as low as 39%
00:17:21.600 elsewhere.
00:17:23.960 What is
00:17:24.460 Rasmussen?
00:17:25.980 I didn't want to
00:17:26.940 guess.
00:17:28.780 But, all right.
00:17:30.640 Here's another one.
00:17:31.540 He jabbed
00:17:34.760 at the press
00:17:35.400 saying that
00:17:36.220 they're the only
00:17:36.960 group of Americans
00:17:37.820 with a lower
00:17:38.440 approval rating
00:17:39.220 than his own.
00:17:42.600 You know,
00:17:43.340 good sort of
00:17:44.540 harmless joke.
00:17:45.840 But again,
00:17:46.660 it's alarming.
00:17:48.600 It's alarming
00:17:49.500 to see them
00:17:50.680 on the same team.
00:17:52.280 Hey, we're on
00:17:53.080 the same team.
00:17:53.700 Everybody hates us.
00:17:54.980 Let's get together.
00:17:56.640 All right.
00:17:57.020 But this is
00:17:57.640 his best joke.
00:17:59.360 So Biden
00:18:00.220 pointed out
00:18:00.820 that he's the
00:18:01.360 first sitting
00:18:01.860 president since
00:18:02.580 2016 to attend
00:18:03.920 this event.
00:18:04.920 And they said,
00:18:05.840 quote,
00:18:06.160 it's understandable
00:18:07.100 we had a horrible
00:18:08.320 plague followed
00:18:09.600 by two years
00:18:10.280 of COVID.
00:18:13.460 Nice.
00:18:15.400 Nice.
00:18:16.700 So the horrible
00:18:17.380 plague here being
00:18:18.220 Trump, if you
00:18:18.920 weren't following
00:18:19.500 the math.
00:18:22.340 Well done.
00:18:24.080 Well done.
00:18:26.640 All right.
00:18:27.240 And now here's
00:18:32.240 the one.
00:18:32.520 And he said,
00:18:34.400 this is also
00:18:35.080 something Biden
00:18:35.700 said.
00:18:36.460 He said it
00:18:37.060 would have been
00:18:37.440 a real coup
00:18:38.700 had Trump
00:18:39.540 attended this
00:18:40.260 year.
00:18:41.320 That would have
00:18:41.920 been a real coup.
00:18:43.580 He's actually
00:18:44.380 joking with the
00:18:45.580 press about
00:18:47.000 Trump organizing
00:18:49.360 a coup.
00:18:51.140 Now,
00:18:51.920 help me out
00:18:53.920 here.
00:18:54.160 January 6th
00:18:56.940 was either
00:18:57.640 an insurrection
00:18:58.560 and a coup
00:18:59.160 and one of
00:18:59.660 the worst
00:19:00.000 things that
00:19:00.460 ever happened
00:19:00.960 to the
00:19:01.240 republic.
00:19:02.480 That's what
00:19:03.060 Biden is
00:19:03.740 saying,
00:19:04.100 right?
00:19:05.120 Or,
00:19:06.040 or it's a
00:19:07.260 joke.
00:19:08.840 It's not
00:19:09.560 both,
00:19:10.920 you motherfuckers.
00:19:12.400 It's not.
00:19:13.800 Pick one.
00:19:15.160 It's either
00:19:15.540 the worst
00:19:16.000 thing that's
00:19:16.380 happened to
00:19:16.740 the republic,
00:19:17.440 that's what
00:19:17.840 all your
00:19:18.220 little toadies
00:19:18.880 are telling
00:19:19.260 us,
00:19:19.540 or it's
00:19:20.920 just a
00:19:21.400 fucking joke.
00:19:23.640 He treated
00:19:24.400 that as a
00:19:24.920 joke.
00:19:26.140 I'm going to
00:19:26.800 take his
00:19:27.160 leadership on
00:19:27.820 that.
00:19:28.800 Joe Biden,
00:19:29.520 I take your
00:19:30.000 leadership.
00:19:30.840 Now,
00:19:31.100 I get that
00:19:32.060 you're at an
00:19:32.540 event where
00:19:33.700 you tell
00:19:34.020 jokes,
00:19:35.120 but where's
00:19:35.780 his joke
00:19:36.240 about the
00:19:37.040 Holocaust?
00:19:38.200 Like,
00:19:38.500 where's the
00:19:38.780 Holocaust joke?
00:19:40.320 It's not
00:19:41.000 there,
00:19:41.320 right?
00:19:42.020 Do you know
00:19:42.360 why it's not
00:19:42.840 there?
00:19:43.960 Because the
00:19:44.560 Holocaust isn't
00:19:45.200 a fucking
00:19:45.580 joke.
00:19:46.580 That's why.
00:19:48.300 You know
00:19:48.560 what is a
00:19:48.920 fucking joke?
00:19:50.180 January 6th.
00:19:51.860 He just
00:19:52.220 told you
00:19:52.540 it's a
00:19:52.840 joke.
00:19:54.120 He joked
00:19:55.180 right in
00:19:55.620 front of
00:19:55.900 you while
00:19:56.720 the
00:19:56.880 investigation
00:19:57.360 is ongoing.
00:19:59.940 I don't
00:20:00.560 think you
00:20:00.920 can ignore
00:20:01.400 this.
00:20:02.300 Seriously.
00:20:03.100 See,
00:20:03.360 this is
00:20:03.600 why Trump
00:20:05.420 is smart
00:20:05.880 not to
00:20:06.260 attend.
00:20:07.320 This is
00:20:08.660 inappropriate.
00:20:10.560 Now,
00:20:11.140 I'm not
00:20:11.440 the person
00:20:11.800 who's going
00:20:12.120 to say
00:20:12.320 jokes are
00:20:12.880 inappropriate
00:20:13.480 as an
00:20:14.860 art form.
00:20:16.120 So I'm
00:20:16.420 not talking
00:20:16.780 about the
00:20:17.080 art form.
00:20:18.220 I'm talking
00:20:18.720 about the
00:20:19.160 president
00:20:19.660 treating
00:20:21.260 our current
00:20:22.000 biggest problem
00:20:22.860 according to
00:20:23.640 him and
00:20:24.600 his people
00:20:25.160 as a joke.
00:20:27.200 Pick one.
00:20:28.780 Pick one.
00:20:29.940 It's either
00:20:30.180 always a joke
00:20:30.820 or it's a
00:20:31.920 big problem.
00:20:32.380 Yes, I
00:20:35.800 told you
00:20:36.100 these
00:20:36.480 analysts at
00:20:37.620 UBS
00:20:38.080 think that
00:20:38.800 inflation
00:20:39.500 might have
00:20:39.900 peaked.
00:20:41.080 So it's
00:20:42.380 coming from
00:20:42.880 smart people
00:20:43.520 too, not
00:20:44.020 just from
00:20:44.480 me jabbering.
00:20:46.480 All right,
00:20:47.000 here's a
00:20:47.960 scary thing.
00:20:50.860 If you
00:20:51.540 were,
00:20:52.720 let's say,
00:20:53.600 hypothetically,
00:20:55.080 and I'm
00:20:55.580 going to
00:20:55.940 frame this
00:20:57.020 by saying
00:20:57.540 I'm
00:20:58.540 unaware of
00:20:59.380 any proof
00:21:00.660 that the
00:21:01.440 2020
00:21:01.940 election
00:21:02.440 was
00:21:03.360 illegitimate.
00:21:06.920 I'm
00:21:07.220 unaware of
00:21:08.220 any proof
00:21:08.680 of that.
00:21:09.800 I'm also
00:21:10.420 unaware of
00:21:11.720 any way you
00:21:12.240 could prove
00:21:12.760 it was
00:21:13.060 legitimate.
00:21:14.320 I don't
00:21:14.980 know that
00:21:15.280 anything could
00:21:15.800 be proved.
00:21:16.700 Since
00:21:17.080 nothing is
00:21:17.660 fully
00:21:17.960 auditable,
00:21:19.400 I would
00:21:19.660 say the
00:21:19.980 only thing
00:21:20.380 that we
00:21:20.800 citizens
00:21:21.840 can say
00:21:22.500 is that we
00:21:23.620 either accepted
00:21:24.460 it or we
00:21:25.140 didn't.
00:21:25.400 It turns
00:21:26.780 out that's
00:21:27.180 the only
00:21:27.400 thing we
00:21:27.700 can say.
00:21:28.820 I, from
00:21:29.780 the very
00:21:30.140 beginning,
00:21:30.820 accepted the
00:21:31.500 result.
00:21:32.920 And the
00:21:33.280 reason I
00:21:33.680 accepted it
00:21:34.320 is because
00:21:34.980 I knew you
00:21:36.140 couldn't
00:21:36.440 check, but
00:21:38.060 you can't
00:21:38.520 throw out
00:21:38.800 the system
00:21:39.400 if the
00:21:39.920 system is
00:21:40.400 the only
00:21:40.620 thing that
00:21:40.940 has a
00:21:41.160 chance of
00:21:41.560 ever fixing
00:21:42.100 this.
00:21:42.980 Like, you
00:21:43.360 want to
00:21:43.600 keep enough
00:21:44.060 of a
00:21:44.340 system alive,
00:21:45.800 you say,
00:21:46.260 okay, maybe
00:21:47.080 someday it'll
00:21:47.740 elect somebody
00:21:48.260 who can
00:21:48.480 look into
00:21:48.880 this and
00:21:49.840 maybe fix
00:21:50.540 it.
00:21:51.740 So, I
00:21:53.560 have no way
00:21:53.960 to know that
00:21:54.400 the 2020
00:21:54.900 election
00:21:55.260 was either
00:21:55.800 rigged or
00:21:56.240 not rigged.
00:21:56.960 That
00:21:57.060 information is
00:21:57.740 forever
00:21:58.180 unavailable to
00:21:59.900 us.
00:22:01.700 But,
00:22:03.060 hypothetically,
00:22:04.620 this is just
00:22:06.160 speculation.
00:22:07.660 Suppose you
00:22:08.600 were on a
00:22:09.140 team, a
00:22:09.820 political team,
00:22:11.220 that expected
00:22:12.580 to rig the
00:22:13.360 next election.
00:22:14.980 I'm not
00:22:15.380 saying they
00:22:15.780 are.
00:22:16.400 How would I
00:22:16.820 know that?
00:22:17.700 I can't read
00:22:18.260 anybody's mind.
00:22:19.780 But suppose
00:22:20.280 that happened.
00:22:20.980 It's just a
00:22:21.440 mental experiment.
00:22:23.040 What would
00:22:23.520 you be doing
00:22:24.380 around now
00:22:26.320 to make sure
00:22:27.840 that worked
00:22:28.340 out well
00:22:28.840 later?
00:22:29.460 Well, you'd
00:22:29.860 be preparing
00:22:30.400 your methods,
00:22:31.440 like how are
00:22:31.960 you going to
00:22:32.280 cheat and
00:22:32.780 get away with
00:22:33.380 it?
00:22:33.800 Again,
00:22:34.340 hypothetically.
00:22:35.260 I'm not
00:22:35.640 accusing anybody
00:22:36.360 of anything.
00:22:36.800 But the
00:22:38.700 most important
00:22:39.300 thing you'd
00:22:39.820 need to do
00:22:40.360 about now,
00:22:42.080 as the
00:22:42.560 election is
00:22:43.060 approaching,
00:22:44.120 is to create
00:22:44.980 some kind of
00:22:45.540 a narrative
00:22:46.000 that would
00:22:48.000 explain why
00:22:49.180 an election
00:22:49.680 result could be
00:22:50.620 so different
00:22:51.300 from what
00:22:52.500 the polls
00:22:53.120 and all
00:22:53.680 common sense
00:22:54.320 and observations
00:22:55.060 suggest they
00:22:55.820 should.
00:22:57.020 Because right
00:22:57.600 now history
00:22:58.080 is telling us
00:22:58.720 that the
00:22:59.180 Republicans will
00:22:59.960 sweep the
00:23:00.480 midterms.
00:23:02.040 And the
00:23:02.540 signal for
00:23:03.060 that is so
00:23:03.660 strong,
00:23:05.040 we would be
00:23:05.700 kind of amazed
00:23:06.520 if it didn't
00:23:07.020 happen,
00:23:07.460 wouldn't we?
00:23:08.660 Even Democrats
00:23:09.440 would be surprised
00:23:10.200 at this point.
00:23:12.000 So if somebody
00:23:13.820 planned to
00:23:14.440 rig an election,
00:23:15.180 and again,
00:23:15.540 there's no
00:23:15.840 evidence of this
00:23:16.460 whatsoever.
00:23:17.560 This is just
00:23:18.180 a mental
00:23:18.520 experiment.
00:23:20.280 If somebody
00:23:21.020 planned to
00:23:21.500 do it,
00:23:22.180 it would be
00:23:22.580 very important
00:23:23.900 that they
00:23:24.480 could seed
00:23:25.080 the public
00:23:25.680 with a
00:23:26.600 narrative that
00:23:27.280 could explain
00:23:28.080 an upcoming
00:23:29.500 inexplicable
00:23:32.240 thing.
00:23:34.160 And so
00:23:34.740 today,
00:23:35.320 we notice
00:23:35.900 with some
00:23:36.560 interest,
00:23:37.520 that Harry
00:23:37.900 Enten,
00:23:38.420 who's an
00:23:38.760 opinion guy
00:23:39.400 who writes
00:23:39.760 for CNN,
00:23:40.980 writes on
00:23:41.400 the CNN
00:23:41.800 website,
00:23:43.180 he describes
00:23:44.020 how Democrats
00:23:45.460 could win
00:23:46.120 the midterms.
00:23:48.180 Interesting.
00:23:50.320 So even
00:23:50.960 though in
00:23:52.040 the article
00:23:52.440 he confesses
00:23:53.360 that every
00:23:54.560 signal says
00:23:55.500 that the
00:23:56.080 Republicans
00:23:56.620 are just
00:23:57.100 going to
00:23:57.460 wipe the
00:23:58.100 Democrats
00:23:58.580 out in
00:23:59.140 the midterms,
00:24:00.100 every signal
00:24:00.740 says it,
00:24:01.180 historical
00:24:01.540 signal,
00:24:02.640 as well as
00:24:03.300 current ones.
00:24:04.480 The polls
00:24:04.920 show that
00:24:05.400 clearly.
00:24:08.460 But here's
00:24:09.340 how Harry
00:24:09.840 Enten says
00:24:10.560 that Democrats
00:24:11.360 could inexplicably,
00:24:13.700 by surprise,
00:24:15.360 nobody saw it
00:24:16.200 coming,
00:24:16.820 but after the
00:24:17.380 fact,
00:24:17.740 you could say,
00:24:18.080 well,
00:24:18.300 Harry Enten
00:24:18.780 saw it
00:24:19.120 coming.
00:24:20.240 And here
00:24:20.520 are the
00:24:20.740 three things
00:24:21.340 it's all
00:24:21.780 it would
00:24:21.980 take.
00:24:23.040 Number
00:24:23.240 one,
00:24:23.820 Harry
00:24:24.240 explains,
00:24:25.320 bad
00:24:25.760 Republican
00:24:26.300 candidates.
00:24:28.520 Well,
00:24:29.020 that's a
00:24:29.260 pretty good
00:24:29.580 comment.
00:24:30.640 If all
00:24:31.020 of the,
00:24:32.060 or if
00:24:32.600 most of
00:24:33.340 the
00:24:33.540 Republicans
00:24:34.020 running
00:24:34.400 for office
00:24:34.960 were just
00:24:35.940 terrible
00:24:36.280 candidates,
00:24:38.140 well,
00:24:38.520 that would
00:24:38.840 change things,
00:24:39.580 wouldn't it?
00:24:40.340 But what
00:24:40.740 are the odds
00:24:41.500 that there
00:24:41.820 would be
00:24:42.100 terrible
00:24:42.500 candidates,
00:24:43.200 like worse
00:24:44.460 than normal?
00:24:45.100 What would
00:24:47.400 be the
00:24:47.720 argument that
00:24:48.280 they would
00:24:48.560 be worse
00:24:49.160 than average,
00:24:50.640 like every
00:24:51.180 other election
00:24:51.960 and every
00:24:52.320 other time?
00:24:53.420 I don't
00:24:53.980 think there's
00:24:54.340 an argument
00:24:54.680 for that,
00:24:55.100 is there?
00:24:56.700 But do
00:24:58.200 you know
00:24:58.460 this has
00:24:58.840 an interesting
00:24:59.460 quality to
00:25:00.260 it?
00:25:01.220 Bad
00:25:01.680 Republican
00:25:02.200 candidates.
00:25:02.880 It's kind
00:25:03.160 of subjective,
00:25:03.860 isn't it?
00:25:05.120 Isn't that
00:25:05.540 interesting?
00:25:05.880 There are
00:25:06.700 three things
00:25:07.300 that would
00:25:07.560 change this
00:25:08.500 election from
00:25:09.320 a predictable
00:25:10.860 Republican
00:25:11.520 victory to
00:25:15.040 a surprising
00:25:15.800 result,
00:25:16.300 and one
00:25:16.520 of them
00:25:16.780 is subjective,
00:25:18.180 that the
00:25:19.000 Republican
00:25:19.460 candidates were
00:25:20.400 bad.
00:25:21.520 They were
00:25:21.700 bad.
00:25:22.700 They didn't
00:25:23.020 do enough
00:25:23.380 work.
00:25:24.400 They didn't
00:25:24.720 campaign right.
00:25:26.040 They didn't
00:25:26.340 have good
00:25:26.880 campaign
00:25:28.940 organization.
00:25:31.180 Isn't it
00:25:31.560 interesting that
00:25:32.260 one of the
00:25:32.800 three things
00:25:33.320 that would
00:25:33.600 explain a
00:25:34.420 surprise
00:25:35.140 election
00:25:35.640 result would
00:25:37.140 be something
00:25:37.580 that you
00:25:37.940 could always
00:25:38.460 say was
00:25:38.880 true and
00:25:39.320 nobody really
00:25:39.860 could prove
00:25:40.360 it?
00:25:41.340 Well,
00:25:41.840 here you
00:25:42.160 go.
00:25:42.880 All those
00:25:43.360 Republicans
00:25:43.860 lost,
00:25:44.560 therefore,
00:25:46.020 it's proof
00:25:46.600 they were
00:25:46.940 bad
00:25:47.240 candidates.
00:25:48.040 Wait a
00:25:48.280 minute,
00:25:48.380 wait a
00:25:48.580 minute.
00:25:49.800 The fact
00:25:50.540 that they
00:25:50.820 lose is
00:25:51.480 proof that
00:25:52.580 they were
00:25:52.800 bad
00:25:53.060 candidates?
00:25:54.860 Or were
00:25:55.680 the bad
00:25:56.120 candidates
00:25:56.660 what caused
00:25:57.400 them to
00:25:57.760 lose?
00:25:58.960 Wait.
00:26:00.600 You see
00:26:01.140 how interesting
00:26:01.680 this thing
00:26:03.160 is?
00:26:03.620 It's a
00:26:04.040 setup that
00:26:04.600 they can
00:26:04.860 just say,
00:26:05.260 well,
00:26:05.380 they ran
00:26:05.740 bad
00:26:06.040 candidates.
00:26:07.140 Nobody
00:26:07.380 can prove
00:26:07.800 they didn't.
00:26:08.780 Nobody
00:26:09.020 can prove
00:26:09.460 they did.
00:26:10.140 It's
00:26:10.380 purely
00:26:10.680 subjective.
00:26:11.840 What's
00:26:12.100 the second
00:26:12.440 one?
00:26:13.400 The second
00:26:13.880 one is
00:26:14.220 the economy
00:26:14.820 improves.
00:26:16.820 What are
00:26:17.400 the odds
00:26:17.960 that by
00:26:20.180 election day
00:26:20.780 there won't
00:26:21.220 be noticeable
00:26:21.900 improvement in
00:26:22.600 the economy?
00:26:24.520 I mean,
00:26:24.880 we could
00:26:25.200 slide into
00:26:25.720 a recession.
00:26:26.440 It could
00:26:26.780 get worse.
00:26:28.200 But if
00:26:29.680 it gets
00:26:29.980 worse,
00:26:31.080 then even
00:26:31.660 cheating isn't
00:26:32.320 going to
00:26:32.520 work.
00:26:33.800 Right?
00:26:34.240 If the
00:26:34.840 economy is
00:26:35.400 actually
00:26:35.680 worse than
00:26:36.900 right now
00:26:37.480 by election
00:26:38.520 day,
00:26:40.220 then there's
00:26:41.940 no Democrat
00:26:42.660 who has a
00:26:45.500 close race
00:26:46.180 who's going
00:26:46.500 to be too
00:26:46.800 happy about
00:26:47.240 that.
00:26:49.240 But it's
00:26:50.180 a good bet
00:26:50.820 that at least
00:26:51.900 on some
00:26:52.480 measures,
00:26:53.080 such as I
00:26:54.220 imagine the
00:26:56.200 inflation rate
00:26:56.900 will come down
00:26:57.600 by then.
00:26:58.380 That's a good
00:26:58.940 bet.
00:26:59.260 You don't
00:26:59.940 know,
00:27:01.160 but it's
00:27:01.520 a smart
00:27:01.900 bet.
00:27:02.980 So let's
00:27:03.580 say a few
00:27:04.100 of the
00:27:04.480 economic
00:27:04.960 indicators
00:27:05.520 improve.
00:27:06.340 Not all
00:27:06.960 of them,
00:27:07.940 but let's
00:27:08.300 just say a
00:27:09.040 few good
00:27:09.380 ones improve.
00:27:10.000 Let's say
00:27:10.260 the GDP
00:27:10.720 is no
00:27:11.180 longer
00:27:11.440 negative.
00:27:11.920 How hard
00:27:12.240 would that
00:27:12.520 be?
00:27:13.640 How hard
00:27:14.140 would it
00:27:14.420 go from
00:27:14.860 our first
00:27:15.560 negative
00:27:16.700 GDP to
00:27:18.480 a little
00:27:19.200 bit positive?
00:27:20.240 Let's say
00:27:20.560 up 1%.
00:27:21.280 Probably
00:27:22.200 not that
00:27:22.600 hard.
00:27:24.060 Right?
00:27:24.680 Supply
00:27:25.140 chain gets
00:27:25.600 worked out.
00:27:26.160 Ukraine
00:27:28.920 starts
00:27:29.400 negotiating
00:27:29.920 with
00:27:30.220 Russia.
00:27:32.100 All of
00:27:32.480 a sudden
00:27:32.800 inflation
00:27:33.620 goes down
00:27:34.400 25%.
00:27:35.340 So even
00:27:38.360 if the
00:27:38.720 economy is
00:27:39.260 still bad
00:27:39.960 and even
00:27:40.440 if Biden
00:27:41.560 and the
00:27:41.960 Democrats
00:27:42.840 are the
00:27:43.260 cause,
00:27:44.200 wouldn't
00:27:44.460 this give
00:27:45.060 Democrats
00:27:46.540 an argument
00:27:47.080 that the
00:27:47.580 economy
00:27:47.940 improved?
00:27:49.120 Well,
00:27:49.460 we told
00:27:49.800 you if
00:27:50.040 the economy
00:27:50.460 improved,
00:27:51.060 the Democrats
00:27:51.520 would do
00:27:51.800 better than
00:27:52.180 you thought.
00:27:53.200 Right?
00:27:53.760 So now
00:27:54.280 you have
00:27:54.580 two subjective
00:27:55.620 things.
00:27:59.080 Bad
00:27:59.600 Republican
00:28:00.040 candidates,
00:28:00.900 nobody would
00:28:01.500 agree what
00:28:01.940 that looks
00:28:02.320 like,
00:28:02.940 and economy
00:28:03.580 improves,
00:28:04.500 which is
00:28:04.960 almost
00:28:05.260 guaranteed
00:28:05.820 just by
00:28:06.660 sitting around
00:28:07.260 and waiting.
00:28:09.800 Now
00:28:10.020 those are
00:28:11.420 two pretty
00:28:12.100 weak indicators,
00:28:13.320 aren't they?
00:28:14.580 And then
00:28:15.000 the third
00:28:15.340 one is
00:28:15.700 Democrats
00:28:16.380 basically have
00:28:17.620 to turn
00:28:17.960 out and
00:28:18.360 vote for
00:28:18.740 Biden.
00:28:19.720 So they
00:28:20.040 have to
00:28:20.260 give a
00:28:20.520 good
00:28:20.720 turnout.
00:28:22.300 What do
00:28:22.760 you think
00:28:22.960 would happen
00:28:23.400 if an
00:28:23.820 unusual
00:28:24.280 number of
00:28:25.180 Democrats
00:28:26.240 voted?
00:28:27.640 Like so
00:28:28.280 many
00:28:28.520 Democrats
00:28:29.000 voted that
00:28:29.680 by historical
00:28:30.700 standards you
00:28:32.480 say to
00:28:32.760 yourself,
00:28:33.740 huh,
00:28:34.320 this doesn't
00:28:35.020 even look
00:28:35.380 real.
00:28:36.760 If I
00:28:37.600 didn't know
00:28:37.940 better,
00:28:38.240 I'd think
00:28:38.520 this was
00:28:38.900 rigged.
00:28:39.460 This number
00:28:39.900 is so
00:28:40.180 big.
00:28:41.560 They're
00:28:42.000 literally
00:28:42.480 telling you
00:28:44.020 their game
00:28:44.440 plan right
00:28:44.920 here.
00:28:45.720 It feels
00:28:46.260 like they
00:28:46.600 just mapped
00:28:47.140 out exactly
00:28:48.020 what they're
00:28:48.460 going to
00:28:48.660 do.
00:28:49.600 There's
00:28:50.040 going to
00:28:50.220 be a
00:28:50.540 rigged
00:28:50.800 election.
00:28:51.520 Again,
00:28:52.020 I don't
00:28:52.320 know this.
00:28:52.820 this is
00:28:53.940 just a
00:28:54.280 mental
00:28:54.500 experiment,
00:28:55.180 right?
00:28:55.540 I don't
00:28:55.920 have any
00:28:56.220 evidence of
00:28:56.820 this.
00:28:58.500 All we
00:28:59.080 have is
00:28:59.400 foreshadowing.
00:29:00.860 Is
00:29:01.320 foreshadowing
00:29:02.040 evidence?
00:29:03.720 Remember,
00:29:04.280 evidence isn't
00:29:05.080 proof.
00:29:07.080 Evidence is
00:29:07.700 just something
00:29:08.180 that maybe
00:29:08.880 collectively could
00:29:09.820 get you to
00:29:10.240 a proof.
00:29:11.780 But
00:29:12.220 foreshadowing is
00:29:13.500 like weaker
00:29:13.920 than that.
00:29:14.860 It's just
00:29:15.220 something that
00:29:15.640 you say,
00:29:16.000 oh,
00:29:16.400 pattern
00:29:16.760 recognition.
00:29:17.880 Pattern
00:29:18.440 recognition has
00:29:19.560 kicked in.
00:29:20.560 I feel
00:29:21.620 like I've
00:29:22.060 seen this
00:29:22.440 pattern before.
00:29:23.420 That's all
00:29:23.740 it is.
00:29:24.420 So it's
00:29:24.800 not evidence.
00:29:25.980 It's
00:29:26.220 definitely
00:29:26.440 not proof.
00:29:28.660 But there's
00:29:29.420 definitely some
00:29:30.080 foreshadowing
00:29:30.680 here.
00:29:31.580 So let me
00:29:32.020 say the
00:29:32.320 one thing
00:29:32.640 that I
00:29:32.840 can say
00:29:33.120 for sure.
00:29:33.920 This is
00:29:34.560 foreshadowing
00:29:35.320 like crazy.
00:29:37.180 This
00:29:37.620 foreshadowing
00:29:38.340 is setting
00:29:38.860 up a
00:29:39.160 narrative
00:29:39.460 that if
00:29:40.520 there's
00:29:40.760 an
00:29:41.040 unexpected
00:29:41.700 Democrat
00:29:42.200 win,
00:29:43.480 and they
00:29:43.820 maintain,
00:29:45.140 you know,
00:29:45.520 let's say,
00:29:45.980 or they
00:29:46.220 somehow they
00:29:47.460 own both
00:29:47.940 sides of
00:29:48.360 Congress,
00:29:49.520 people are
00:29:50.180 going to
00:29:50.340 say,
00:29:50.560 look,
00:29:50.820 Harry
00:29:51.020 Enten
00:29:51.280 already
00:29:51.600 told you
00:29:51.980 how this
00:29:52.280 could happen.
00:29:53.280 It was
00:29:53.520 easy.
00:29:54.960 Republican
00:29:55.400 candidates
00:29:55.840 were crazy.
00:29:57.360 They were
00:29:57.580 basically all
00:29:58.260 Marjorie Taylor
00:29:58.900 Greens,
00:29:59.380 that's what
00:29:59.720 they're going
00:29:59.960 to say.
00:30:00.760 Didn't have
00:30:01.340 any good
00:30:01.640 candidates.
00:30:02.260 They were
00:30:02.440 all Trumpers
00:30:03.140 or whatever
00:30:03.480 they were
00:30:03.700 going to
00:30:03.920 say.
00:30:04.780 The economy
00:30:05.460 improved,
00:30:06.120 of course,
00:30:06.680 in some
00:30:07.340 ways it
00:30:07.700 will,
00:30:07.940 they'll make
00:30:08.260 that argument.
00:30:09.660 And then
00:30:09.980 they'll say,
00:30:10.520 yeah,
00:30:10.860 and the
00:30:11.980 Democrats
00:30:12.380 were really
00:30:12.900 enthusiastic,
00:30:13.780 got that
00:30:14.160 turnout out
00:30:14.720 there.
00:30:15.080 Man,
00:30:15.320 they got
00:30:15.560 some
00:30:15.780 turnout.
00:30:17.420 This is
00:30:18.020 pretty scary
00:30:18.600 shit.
00:30:18.940 Every once
00:30:22.720 in a while
00:30:23.040 you see a
00:30:23.520 topic that's
00:30:24.500 a political
00:30:24.940 topic in
00:30:26.440 which all
00:30:27.140 the people
00:30:27.500 on one
00:30:27.880 side are
00:30:28.240 the smart
00:30:28.680 people,
00:30:29.400 regardless
00:30:30.060 of their
00:30:30.540 political
00:30:30.960 affiliations,
00:30:32.080 and all
00:30:32.900 the people
00:30:33.220 on the
00:30:33.460 other side
00:30:33.820 are dumb
00:30:34.140 people.
00:30:35.360 Now,
00:30:35.520 you don't
00:30:35.700 see that
00:30:36.060 often.
00:30:36.840 Usually you
00:30:37.320 see people
00:30:37.720 just support
00:30:39.100 their side.
00:30:40.580 But this
00:30:41.460 whole free
00:30:42.500 speech thing
00:30:44.260 is actually
00:30:45.020 very interesting
00:30:46.060 because all
00:30:47.400 the smart
00:30:47.920 people are
00:30:48.640 on the
00:30:48.860 same
00:30:49.040 side.
00:30:50.260 And I'll
00:30:50.620 give you
00:30:50.820 an example.
00:30:51.400 Bill Maher,
00:30:52.300 he wasn't
00:30:53.180 too familiar
00:30:53.720 with the
00:30:54.080 Babylon Bee,
00:30:55.540 the satirical
00:30:56.600 site that
00:30:57.080 got banned
00:30:57.480 from Twitter,
00:30:58.860 but having
00:31:00.520 read up on
00:31:01.180 it,
00:31:01.320 he supports
00:31:01.940 the fact
00:31:02.600 that they
00:31:02.900 shouldn't
00:31:03.100 have been
00:31:03.340 banned for
00:31:03.860 jokes,
00:31:05.040 basically.
00:31:06.000 So Bill
00:31:06.400 Maher is
00:31:07.680 on the same
00:31:08.140 side as
00:31:08.800 the darkest,
00:31:10.800 deepest
00:31:11.100 Republican on
00:31:12.580 this issue.
00:31:13.880 But also
00:31:14.700 all the smart
00:31:15.380 people are
00:31:15.800 on this
00:31:16.100 side.
00:31:16.320 I don't
00:31:18.320 think you
00:31:19.760 could find
00:31:20.200 a smart
00:31:20.700 Republican
00:31:21.280 or a
00:31:21.960 smart
00:31:22.200 Democrat
00:31:22.640 who would
00:31:23.940 think that
00:31:24.300 the Babylon
00:31:24.740 Bee should
00:31:25.340 have been
00:31:25.620 banned from
00:31:26.800 Twitter.
00:31:27.940 I don't
00:31:28.620 think you
00:31:28.900 could find
00:31:29.240 one.
00:31:29.980 It's a
00:31:30.360 weird issue
00:31:30.920 that literally
00:31:32.820 just intelligence
00:31:33.680 is the
00:31:34.300 dividing line
00:31:35.000 and not
00:31:35.480 party
00:31:36.080 affiliation.
00:31:37.360 Because at
00:31:38.100 a certain
00:31:38.400 level of
00:31:38.800 intelligence,
00:31:39.800 you know
00:31:40.220 that whatever
00:31:40.680 the Babylon
00:31:41.220 Bee did
00:31:41.900 had minimal
00:31:43.320 to no
00:31:43.780 effect on
00:31:44.300 anybody.
00:31:45.720 But any
00:31:46.660 kind of
00:31:47.520 punishment to
00:31:48.280 free speech,
00:31:49.600 even by a
00:31:50.740 private entity
00:31:51.660 in this
00:31:52.060 special case,
00:31:53.680 is a
00:31:55.220 bad
00:31:55.640 precedent.
00:31:57.540 Pretty much
00:31:57.960 all the smart
00:31:58.500 people agree
00:31:59.020 with that.
00:32:01.340 Yeah.
00:32:02.660 All right,
00:32:03.140 but here's the
00:32:03.520 most surprising
00:32:04.080 story of the
00:32:04.700 day.
00:32:05.720 And I would
00:32:06.380 like you to
00:32:06.780 savor this
00:32:07.300 one.
00:32:07.540 or to
00:32:09.120 possibly puke
00:32:10.600 in your
00:32:10.880 mouth on
00:32:11.280 it,
00:32:11.480 depending on
00:32:11.940 your political
00:32:12.480 leaning.
00:32:14.100 Do you all
00:32:14.600 know Noam
00:32:15.700 Chomsky?
00:32:16.640 He's a very
00:32:17.620 famous intellectual
00:32:18.540 and he's
00:32:19.680 famous for
00:32:20.140 many things,
00:32:20.920 including being
00:32:21.560 so left,
00:32:22.860 he's like the
00:32:23.540 leftiest of
00:32:24.280 the left,
00:32:24.860 wouldn't you
00:32:25.180 say?
00:32:26.340 Right,
00:32:28.380 he's a
00:32:28.880 linguist,
00:32:29.420 but he talks
00:32:29.880 about politics
00:32:31.440 and he's
00:32:32.040 famous for
00:32:32.580 that.
00:32:34.240 So would
00:32:35.180 you agree
00:32:35.600 that he's
00:32:36.500 as left
00:32:37.020 as the
00:32:37.840 lefties
00:32:38.560 could be
00:32:39.120 left?
00:32:41.240 Now here's
00:32:42.040 the surprising
00:32:42.540 part.
00:32:46.420 So Noam
00:32:47.260 Chomsky was
00:32:47.840 being interviewed,
00:32:49.260 he was
00:32:49.560 talking about
00:32:50.100 the Ukraine
00:32:50.780 Russia
00:32:51.240 situation,
00:32:52.840 and this
00:32:53.240 was last
00:32:53.700 week,
00:32:54.140 no this
00:32:54.480 week in
00:32:54.820 an interview
00:32:55.100 he said,
00:32:56.200 quote,
00:32:56.820 fortunately
00:32:57.340 there is
00:32:58.720 one western
00:32:59.600 statesman
00:33:00.840 of stature
00:33:01.620 who is
00:33:02.480 pushing for
00:33:03.080 a diplomatic
00:33:03.560 solution to
00:33:04.340 the war
00:33:04.740 in Ukraine
00:33:05.260 rather than
00:33:06.380 looking for
00:33:06.900 ways to
00:33:07.460 prolong it.
00:33:10.280 Who's
00:33:10.800 that?
00:33:12.520 So who
00:33:13.060 is this
00:33:13.680 one western
00:33:14.720 statesman of
00:33:15.480 stature who
00:33:16.620 is pushing
00:33:17.040 for a
00:33:17.420 diplomatic
00:33:17.760 solution to
00:33:18.560 the war
00:33:18.860 in Ukraine
00:33:19.340 rather than
00:33:20.660 looking for
00:33:21.060 ways to
00:33:21.500 fuel and
00:33:21.940 prolong it?
00:33:23.060 And then
00:33:23.300 later Chomsky
00:33:24.180 explained what
00:33:25.080 that idea
00:33:25.660 was,
00:33:26.140 specifically
00:33:26.700 the idea
00:33:28.020 was to
00:33:28.380 have a
00:33:28.700 non-NATO
00:33:29.560 accommodation.
00:33:31.600 And accommodation
00:33:32.100 is an
00:33:32.600 interesting word,
00:33:33.320 it really
00:33:34.000 works in
00:33:34.540 this context.
00:33:36.080 And what
00:33:36.880 he suggested
00:33:37.460 was,
00:33:38.560 and he
00:33:39.280 was saying,
00:33:39.860 you know,
00:33:40.100 this one
00:33:40.760 statement,
00:33:41.400 the statesman
00:33:42.000 didn't say
00:33:42.400 this directly
00:33:42.960 but it's
00:33:43.380 obvious this
00:33:43.940 is where
00:33:44.220 he was
00:33:44.460 going,
00:33:45.680 that one
00:33:47.340 thing we
00:33:48.020 could do
00:33:48.340 to end
00:33:48.620 the war,
00:33:49.380 negotiation
00:33:49.880 wise,
00:33:50.680 is to say
00:33:51.460 to Putin,
00:33:51.940 you know
00:33:52.180 what,
00:33:52.940 we did
00:33:53.420 promise you
00:33:53.980 that NATO
00:33:54.620 would not
00:33:55.120 expand toward
00:33:56.260 Russia and
00:33:57.280 we were
00:33:58.780 breaking that
00:33:59.440 promise.
00:34:00.560 So how
00:34:01.140 about we
00:34:01.520 go back
00:34:01.900 to the
00:34:02.160 promise to
00:34:03.320 not expand
00:34:03.980 NATO and
00:34:05.140 we'll form
00:34:05.680 this other
00:34:06.180 alliance that's
00:34:07.280 not NATO.
00:34:08.480 And it
00:34:08.920 would be open
00:34:09.380 not just to
00:34:10.020 Europeans and
00:34:10.820 Americans,
00:34:11.620 it would be
00:34:11.900 open to
00:34:12.160 anybody.
00:34:12.980 And it
00:34:13.180 would be
00:34:13.420 sort of a
00:34:13.920 mutual
00:34:14.460 protecting
00:34:15.420 each other
00:34:16.000 agreement that
00:34:19.000 wouldn't be
00:34:19.440 NATO.
00:34:20.420 It wouldn't
00:34:20.840 be anti-Russia,
00:34:21.800 it wouldn't
00:34:21.960 be anti-
00:34:22.400 anybody.
00:34:22.980 It would
00:34:23.140 just be
00:34:23.560 everybody,
00:34:24.900 let's look
00:34:25.440 out for each
00:34:25.940 other's backs,
00:34:27.100 if somebody
00:34:27.560 attacks you,
00:34:28.440 we're your
00:34:28.760 friend,
00:34:29.780 but not
00:34:30.180 too specific
00:34:30.920 maybe.
00:34:32.060 And he
00:34:32.280 said that
00:34:32.740 even Japan
00:34:34.040 could join
00:34:34.660 it and
00:34:35.920 even Russia.
00:34:37.600 Russia could
00:34:38.160 actually join
00:34:38.760 this hypothetical
00:34:39.920 organization because
00:34:41.500 they too would
00:34:43.160 probably like to
00:34:43.800 have friends if
00:34:44.780 they were
00:34:45.020 attacked.
00:34:47.260 And so
00:34:48.060 Chomsky is
00:34:48.900 saying that
00:34:49.280 there is a
00:34:50.060 negotiated way
00:34:50.920 through this and
00:34:51.540 when he describes
00:34:52.220 it I say to
00:34:52.700 myself that
00:34:53.880 feels like that
00:34:54.480 could work because
00:34:55.700 that would give
00:34:56.140 Putin a win
00:34:56.820 because he could
00:34:58.320 say I stopped
00:34:59.060 NATO.
00:35:00.740 And he
00:35:01.560 probably
00:35:01.780 give some
00:35:02.160 territorial
00:35:02.780 consolidation,
00:35:05.260 at least
00:35:06.320 Crimea,
00:35:07.140 right?
00:35:07.800 So he'd
00:35:08.560 give something
00:35:09.000 and he could
00:35:09.900 say I
00:35:10.240 stopped
00:35:10.620 NATO.
00:35:12.980 At the
00:35:13.480 same time,
00:35:15.060 countries that
00:35:15.720 want the
00:35:16.100 protection of
00:35:16.740 NATO but
00:35:17.260 don't want to
00:35:17.880 be as
00:35:18.640 provocative as
00:35:19.880 joining NATO
00:35:20.600 could say,
00:35:21.660 well, we're
00:35:22.440 not going to
00:35:22.820 join NATO
00:35:23.420 because we
00:35:24.260 understand how
00:35:25.020 that feels to
00:35:25.780 Russia and
00:35:26.360 you guys are
00:35:26.900 prickly about
00:35:27.580 that.
00:35:27.880 But we'll
00:35:29.180 have this
00:35:29.520 other friendly
00:35:30.160 organization that's
00:35:31.280 not nearly as
00:35:32.320 provocative as
00:35:33.420 NATO and it'll
00:35:34.960 just make
00:35:35.360 everybody happy
00:35:35.940 because we're
00:35:36.460 friends.
00:35:38.480 Now, that
00:35:39.140 would make it
00:35:39.860 look like
00:35:40.300 everybody could
00:35:40.980 win.
00:35:41.800 You could argue
00:35:42.760 who really
00:35:43.920 wins, but it
00:35:44.460 would give
00:35:44.760 everybody an
00:35:45.420 argument that
00:35:46.740 they got
00:35:47.040 something.
00:35:48.760 And we're
00:35:49.480 getting back to
00:35:50.040 the statesman.
00:35:50.640 So, who
00:35:51.620 would be the
00:35:52.060 one statesman
00:35:53.620 that Noam
00:35:54.160 Chomsky, the
00:35:55.340 leftiest of the
00:35:56.100 left people in
00:35:56.680 the world, the
00:35:57.800 guiding light of
00:35:58.660 their intellectual
00:35:59.380 life, who
00:36:02.120 would he say is
00:36:03.080 smart enough,
00:36:03.900 really, of
00:36:05.480 stature, who
00:36:06.720 would have a
00:36:07.440 way to negotiate
00:36:08.940 this thing
00:36:09.380 successfully?
00:36:11.020 Well, Chomsky
00:36:11.600 said his name is
00:36:12.600 Donald J.
00:36:13.640 Trump.
00:36:16.060 I will pause
00:36:17.240 for a moment.
00:36:17.820 This really
00:36:20.360 happened.
00:36:23.260 This actually
00:36:24.380 happened.
00:36:25.620 Noam Chomsky
00:36:26.400 gave a full-throated
00:36:28.560 endorsement to
00:36:30.160 Donald Trump's
00:36:31.100 approach and
00:36:32.780 called out the
00:36:33.420 fact that nobody
00:36:34.080 else was in his
00:36:34.940 league.
00:36:38.520 Do you know
00:36:39.080 why Noam
00:36:39.620 Chomsky
00:36:40.160 supported Donald
00:36:42.280 Trump, who you
00:36:43.360 would imagine
00:36:43.840 would be as
00:36:44.400 far away as
00:36:45.560 they could
00:36:45.800 possibly be in
00:36:47.200 political
00:36:47.620 ideology?
00:36:48.580 Do you know
00:36:48.880 why he
00:36:49.140 supported him?
00:36:51.580 Because Noam
00:36:52.260 Chomsky didn't
00:36:53.040 want to lie.
00:36:55.000 That's my
00:36:55.540 guess.
00:36:56.020 I mean, I
00:36:56.440 can't read his
00:36:56.920 mind.
00:36:58.660 I feel like he
00:36:59.480 just didn't
00:36:59.840 want to lie.
00:37:01.720 Maybe he
00:37:02.280 always feels
00:37:02.720 like that.
00:37:03.240 You know, I'm
00:37:03.480 not saying that
00:37:04.060 he lies in any
00:37:05.140 other context.
00:37:06.220 It could be
00:37:06.780 that Noam
00:37:07.200 Chomsky is a
00:37:08.060 legitimate
00:37:08.460 intellectual.
00:37:10.260 And by
00:37:10.520 legitimate, I
00:37:11.340 mean that he
00:37:11.820 says what he
00:37:12.320 actually thinks
00:37:12.840 is real.
00:37:14.240 That would be
00:37:14.800 legitimate.
00:37:15.420 Even if you
00:37:15.900 disagree with
00:37:16.540 his opinions,
00:37:17.620 it's very
00:37:18.220 legitimate to
00:37:19.720 be that
00:37:20.100 educated, that
00:37:20.980 smart, and
00:37:22.340 be honest.
00:37:23.580 Like, you'd
00:37:24.000 have to respect
00:37:25.120 that, right?
00:37:26.900 So even if you
00:37:27.720 disagree with
00:37:28.260 him.
00:37:29.340 But I'm going
00:37:31.400 to introduce a
00:37:32.040 new concept
00:37:32.540 today.
00:37:33.500 We always talk
00:37:34.280 about free
00:37:34.880 speech.
00:37:35.360 I don't think
00:37:37.860 free speech is
00:37:38.720 exactly the way
00:37:39.900 to talk about
00:37:41.320 it.
00:37:41.580 Because in our
00:37:42.400 world where
00:37:43.080 commerce and
00:37:43.940 freedom are
00:37:44.680 connected, like
00:37:46.780 you can say
00:37:47.320 something terrible
00:37:48.140 but it's going
00:37:48.740 to hurt you
00:37:49.100 economically.
00:37:51.000 So yeah,
00:37:51.800 technically that's
00:37:52.700 free speech,
00:37:53.400 right?
00:37:54.000 Because the
00:37:54.520 government didn't
00:37:55.220 necessarily do it
00:37:56.020 to you.
00:37:56.880 But you're not
00:37:57.720 really free if
00:37:59.280 you can't say
00:37:59.840 something without
00:38:00.460 economic repercussions.
00:38:01.780 But I'm going to
00:38:02.940 tell you that there's
00:38:03.600 a new thing that is
00:38:04.460 better than free
00:38:05.140 speech and it's
00:38:05.800 called affordable
00:38:07.140 speech.
00:38:07.960 It's not free,
00:38:09.900 but is it
00:38:10.680 affordable?
00:38:11.120 And there are
00:38:12.780 two categories of
00:38:13.560 people at least
00:38:14.320 who have what I
00:38:15.740 call affordable
00:38:16.740 speech.
00:38:18.080 Number one,
00:38:19.420 me.
00:38:20.420 People who have
00:38:21.180 money can afford
00:38:23.260 free speech.
00:38:24.580 So when I say I
00:38:25.420 have fuck you
00:38:26.060 money, what that
00:38:27.460 means is I can
00:38:28.420 say my opinion
00:38:29.520 as honestly as I
00:38:30.940 want and even
00:38:31.440 dishonestly if I
00:38:32.320 want because I
00:38:33.220 have the ability to
00:38:34.180 say anything I
00:38:34.800 want and then if
00:38:36.120 I'm damaged
00:38:36.960 economically, I
00:38:38.400 say, oh fuck
00:38:39.720 you, I'm going to
00:38:40.140 retire anyway.
00:38:41.120 I have enough
00:38:41.560 money.
00:38:42.620 So I don't have
00:38:43.540 free speech because
00:38:45.140 I will get whacked
00:38:46.040 economically if I
00:38:47.120 cross the line,
00:38:47.900 right?
00:38:48.340 I already have.
00:38:49.120 I probably lost
00:38:49.940 one third of my
00:38:51.080 potential income
00:38:52.060 for the last 10
00:38:52.740 years based on my
00:38:54.560 blogging and I
00:38:56.180 knew it, right?
00:38:57.200 I was completely
00:38:58.320 aware that the
00:39:00.100 things I was saying
00:39:00.920 would degrade my
00:39:01.860 income substantially,
00:39:03.100 but I could afford
00:39:04.220 it.
00:39:05.540 So I had free
00:39:06.580 speech.
00:39:07.340 You didn't.
00:39:08.300 You absolutely
00:39:09.040 did not unless
00:39:10.180 you were also
00:39:10.640 rich.
00:39:11.940 But I had free
00:39:13.020 speech.
00:39:13.960 It cost me a lot
00:39:15.060 of money, but I
00:39:15.640 could afford it.
00:39:16.900 There's another
00:39:17.540 category of free
00:39:18.340 speech.
00:39:19.720 Old as fuck.
00:39:22.340 When you're old
00:39:23.160 as fuck, you
00:39:24.360 don't care.
00:39:26.060 Noam Chomsky is
00:39:26.920 old as fuck.
00:39:28.340 Now, I don't know
00:39:29.080 what, how old is
00:39:29.920 he?
00:39:30.540 I mean, he looks
00:39:31.040 like he doesn't
00:39:31.760 have long, honestly.
00:39:34.820 Give me an age
00:39:35.620 on him.
00:39:36.500 It's 145 or
00:39:38.160 something.
00:39:39.100 He's super old.
00:39:41.000 But I think Noam
00:39:41.800 Chomsky, he doesn't
00:39:42.600 give a fuck.
00:39:44.880 Noam Chomsky, he's
00:39:46.020 93, somebody says.
00:39:47.960 Noam Chomsky can
00:39:49.420 afford his speech.
00:39:52.120 Because if you give
00:39:52.880 Noam Chomsky the
00:39:53.780 bill, he'll say,
00:39:54.940 yeah, give me 30
00:39:55.620 days, I'll get right
00:39:56.300 back to you on that.
00:39:57.020 He'll probably be
00:39:57.500 dead anyway.
00:39:58.460 Okay, I'm
00:39:58.860 exaggerating.
00:39:59.540 But you get the
00:39:59.980 point, right?
00:40:01.240 You don't give a
00:40:02.040 fuck.
00:40:03.200 Noam Chomsky can
00:40:04.060 say anything he
00:40:04.680 wants.
00:40:05.540 So he can offend
00:40:06.500 every person
00:40:07.200 who ever loved
00:40:07.800 him.
00:40:09.800 Maybe he just
00:40:10.600 did.
00:40:11.760 I don't know.
00:40:13.340 But Noam has
00:40:15.520 affordable speech,
00:40:17.140 and I have
00:40:17.680 affordable speech.
00:40:20.500 Elon Musk has
00:40:21.660 affordable speech.
00:40:23.820 Elon Musk talked
00:40:24.920 down his own
00:40:25.660 stock price.
00:40:27.620 Do you remember
00:40:28.000 that?
00:40:30.280 He said his own
00:40:31.700 stock, at one
00:40:32.240 point in the past,
00:40:32.960 he said his own
00:40:33.480 stock was too
00:40:34.040 high, and then
00:40:35.080 like dropped.
00:40:36.760 And he said a
00:40:37.340 bunch of other
00:40:38.060 things that, you
00:40:38.720 know, he seemed
00:40:39.200 getting sued for
00:40:40.020 some of the
00:40:40.380 things he said.
00:40:41.460 So he got sued
00:40:42.200 for, I don't
00:40:42.700 know, tens of
00:40:43.220 millions of
00:40:43.620 dollars by
00:40:45.320 stockholders for
00:40:46.940 saying something
00:40:47.920 about taking it
00:40:48.620 private, and it
00:40:49.440 affected the stock
00:40:50.240 price, whatever.
00:40:51.780 But could he
00:40:52.640 afford it?
00:40:53.920 He could.
00:40:55.160 It turns out,
00:40:56.340 even though it
00:40:57.080 might cost tens of
00:40:58.020 millions, hundreds of
00:40:58.880 millions of dollars,
00:41:00.460 turns out he could
00:41:01.160 afford it.
00:41:02.080 So he had
00:41:02.680 affordable speech,
00:41:03.580 speech, and he
00:41:04.900 might try to give
00:41:05.660 you free speech, or
00:41:07.780 more affordable
00:41:08.380 speech, through
00:41:09.160 Twitter.
00:41:11.540 All right, here's
00:41:12.680 another foreshadowing.
00:41:15.140 Nancy Pelosi
00:41:15.980 visited Ukraine, so
00:41:17.200 now she's the
00:41:17.700 highest-ranking
00:41:18.320 person to go over
00:41:19.000 there.
00:41:19.720 Clearly, we're
00:41:20.660 sending lots and
00:41:21.520 lots of signals to
00:41:23.500 the world and
00:41:24.620 Ukraine by sending
00:41:26.240 such high-level
00:41:27.740 politicians over there
00:41:29.940 in the middle of a
00:41:31.100 war.
00:41:31.320 And that's a really
00:41:32.380 strong signal.
00:41:33.440 And I will give the
00:41:34.580 Biden administration
00:41:35.380 credit for the
00:41:38.400 communication there.
00:41:39.640 They are trying to
00:41:40.680 send that signal, and
00:41:42.180 they're succeeding.
00:41:43.380 They're doing a real
00:41:44.120 good job of sending a
00:41:45.560 we're backing Ukraine.
00:41:47.840 But watch the
00:41:48.620 language.
00:41:49.200 Here's a little trick
00:41:50.040 that I learned in the
00:41:51.080 hypnosis class.
00:41:53.700 Hypnotists learn, well,
00:41:55.280 depending who instructs
00:41:56.640 you, you might not
00:41:57.280 learn this.
00:41:57.700 They learn that you
00:41:58.900 can fairly accurately
00:42:00.740 identify what people
00:42:02.060 are thinking or
00:42:02.920 planning or secretly
00:42:04.060 cogitating by the
00:42:07.760 words that they
00:42:08.540 choose.
00:42:09.960 So when people speak,
00:42:11.080 they're not thinking
00:42:11.880 about every word before
00:42:12.920 it comes out their
00:42:13.560 mouth.
00:42:14.460 We speak sort of in
00:42:15.680 flow, right?
00:42:17.480 My language is coming
00:42:18.840 out now almost like my
00:42:20.200 brain is doing
00:42:20.800 something slightly
00:42:21.720 different because my
00:42:23.080 brain is like a little
00:42:24.000 bit ahead of where I'm
00:42:25.140 talking.
00:42:25.580 So literally right
00:42:27.380 now, I'm modeling
00:42:29.040 this, my mouth is
00:42:30.280 talking, but I'm not
00:42:31.080 even super conscious
00:42:32.300 of the mental part of
00:42:33.680 it.
00:42:34.520 It's just so automatic
00:42:35.720 because I kind of know
00:42:36.960 what I'm going to talk
00:42:37.580 about.
00:42:38.760 So in that context,
00:42:41.780 people will use word
00:42:43.780 choices that reveal
00:42:45.960 what they're really
00:42:46.740 thinking because they're
00:42:48.220 not thinking too hard
00:42:49.220 about filtering it.
00:42:50.760 Where this doesn't
00:42:51.920 work is when people
00:42:54.180 have a speech that's
00:42:55.420 written and a lot of
00:42:56.240 people look at it and
00:42:57.680 a lot of people look
00:42:58.360 at your political
00:42:58.960 speech, they go,
00:42:59.720 ooh, were you aware
00:43:01.360 that this word makes
00:43:02.280 me think this?
00:43:03.360 And somebody say,
00:43:03.980 oh, I didn't know
00:43:04.520 that.
00:43:05.140 They'll take that word
00:43:05.760 out.
00:43:06.540 So if you've got a
00:43:07.520 nice scrubbed piece
00:43:09.000 of work, then this
00:43:10.180 doesn't work.
00:43:11.540 But anything like a
00:43:12.600 tweet where somebody
00:43:14.200 is not putting a lot
00:43:14.960 of work into it,
00:43:16.400 send out of a tweet,
00:43:17.920 you can slip out some
00:43:19.500 words that were
00:43:20.240 unintended.
00:43:21.280 I feel like this is
00:43:22.560 happening with this
00:43:23.760 Ukraine situation.
00:43:24.660 One tweet that Nancy
00:43:27.580 Pelosi gave was, it
00:43:29.120 ended with this,
00:43:29.880 America stands firmly
00:43:31.140 with Ukraine.
00:43:32.640 Now that would be
00:43:33.640 compatible with
00:43:34.500 everything that's been
00:43:35.540 said so far, right?
00:43:37.520 That the United States
00:43:38.460 stands firmly with
00:43:39.500 Ukraine.
00:43:40.420 Now that is a
00:43:42.540 scrubbed statement.
00:43:44.540 That's one that people
00:43:45.720 could look at and say,
00:43:47.020 okay, that is right on
00:43:48.740 point and it doesn't
00:43:49.660 make people think the
00:43:50.660 wrong thing.
00:43:51.480 It doesn't mislead in
00:43:52.460 any way.
00:43:52.800 It's not ambiguous.
00:43:53.420 But America stands
00:43:56.580 with Ukraine, leaves
00:43:57.580 open a possibility,
00:43:58.580 doesn't it?
00:43:59.900 And that possibility is
00:44:01.300 that Ukraine could
00:44:02.080 lose and we're with
00:44:05.040 them.
00:44:06.260 But we're not really
00:44:07.200 talking about victory.
00:44:08.600 We're just sort of
00:44:09.460 with you.
00:44:10.860 It's like being with
00:44:12.200 somebody as they're
00:44:12.880 dying.
00:44:13.980 We're with you.
00:44:16.240 You're going to go
00:44:17.020 through something bad,
00:44:17.940 but we're with you.
00:44:18.620 Now that was one
00:44:21.200 thing she said.
00:44:22.160 But here's another
00:44:22.760 thing she said.
00:44:25.280 This is a Pelosi
00:44:26.240 tweet.
00:44:27.380 It's around the same
00:44:28.240 time too, so there's
00:44:29.520 not much time
00:44:30.540 difference between
00:44:31.180 these tweets.
00:44:32.280 She says, our
00:44:32.780 congressional delegation
00:44:33.920 was honored to meet
00:44:34.880 with Zelensky in Kiev
00:44:36.100 to salute his
00:44:37.420 leadership and
00:44:38.080 courage.
00:44:38.500 So, so far this is
00:44:39.340 all like blah, blah,
00:44:40.180 blah, you know,
00:44:41.060 political talk, right?
00:44:42.100 To commend the
00:44:42.940 Ukrainian people for
00:44:44.240 their outstanding
00:44:44.920 defense of democracy
00:44:46.060 and to say that
00:44:47.320 we are with you
00:44:48.080 until victory is
00:44:49.220 won.
00:44:50.020 What?
00:44:51.960 Until victory is
00:44:53.060 won.
00:44:54.600 What the hell is
00:44:55.400 victory?
00:44:56.920 What would that
00:44:57.600 look like?
00:45:01.080 Victory is a
00:45:03.260 slip, I think.
00:45:06.140 So based on
00:45:06.900 hypnosis training,
00:45:08.820 the word victory
00:45:09.580 looks like a
00:45:10.200 mistake.
00:45:11.580 Like, like they
00:45:12.420 should have edited
00:45:13.080 that word out of
00:45:13.860 there.
00:45:14.340 What they should
00:45:15.080 have said is what
00:45:15.620 the other tweet
00:45:16.200 said.
00:45:17.480 America stands
00:45:18.280 firmly with
00:45:18.940 Ukraine.
00:45:20.640 Right?
00:45:21.820 Who put that
00:45:22.600 word victory in
00:45:23.440 there?
00:45:24.480 I don't think
00:45:25.180 that was
00:45:25.480 intentional.
00:45:27.060 Or, it's
00:45:28.640 foreshadowing.
00:45:30.500 They may be,
00:45:31.260 they may be
00:45:32.200 preparing the
00:45:32.900 country, and
00:45:34.700 this could be
00:45:35.340 intentionally doing
00:45:36.520 it, preparing the
00:45:37.360 country.
00:45:38.440 Remember I told
00:45:39.240 you that this is a
00:45:41.080 tipping point war,
00:45:42.740 war, and
00:45:44.740 that whoever
00:45:45.400 pushes the
00:45:46.520 other one in
00:45:47.100 any one of
00:45:47.640 these variables
00:45:48.340 just over that
00:45:49.400 tipping point
00:45:50.040 would be what
00:45:51.620 looks like the
00:45:52.200 winner.
00:45:53.400 And the tipping
00:45:53.960 points are
00:45:54.540 running out of
00:45:55.100 fuel, running
00:45:56.560 out of ammo,
00:45:57.420 the good kind,
00:45:58.220 wherever you need
00:45:58.820 it the most,
00:46:00.260 and running out
00:46:01.120 of, let's say,
00:46:02.580 food, fuel, and
00:46:04.220 ammo.
00:46:05.760 It's those three
00:46:06.420 things.
00:46:07.580 And both sides
00:46:08.620 are trying to
00:46:09.220 deny the other
00:46:10.060 those three
00:46:10.600 things, because
00:46:11.760 that would be
00:46:12.260 enough for, you
00:46:13.920 know, surrender.
00:46:15.600 And they're all
00:46:18.220 at the tipping
00:46:18.720 point, or we
00:46:20.720 imagine that to
00:46:21.420 be true, in the
00:46:22.640 fog of war and
00:46:23.460 what we can see
00:46:24.080 from the outside.
00:46:25.100 Any one of those
00:46:25.820 things could be a
00:46:26.560 tipping point, but
00:46:27.840 in either direction.
00:46:29.900 So if you're
00:46:30.700 looking at, like,
00:46:31.440 who has gained
00:46:32.100 what territory, or
00:46:33.360 who repelled who
00:46:34.780 from what places,
00:46:35.780 I don't know
00:46:37.000 that that's
00:46:37.400 telling you
00:46:37.820 anything.
00:46:39.080 Those are not
00:46:39.640 the important
00:46:40.080 variables, because
00:46:41.540 this is a who
00:46:42.260 can last longer
00:46:43.240 war, and that's
00:46:44.340 going to be a
00:46:44.720 tipping point
00:46:45.260 question.
00:46:46.020 It's not going
00:46:46.540 to be who had
00:46:47.680 so much progress
00:46:48.580 so far.
00:46:50.040 Because it could
00:46:50.600 be that the team
00:46:51.920 that pushed the
00:46:52.660 furthest used up
00:46:54.100 the most fuel, and
00:46:56.020 put their supply
00:46:56.980 lines in the most
00:46:57.720 jeopardy.
00:46:58.840 So the thing
00:46:59.660 you're looking at
00:47:00.180 could be exactly
00:47:00.820 the opposite of
00:47:01.540 what matters.
00:47:02.220 You know, who
00:47:02.480 got a battleground
00:47:03.960 victory?
00:47:04.460 It might not
00:47:04.880 matter.
00:47:05.780 In a war of
00:47:06.480 attrition.
00:47:07.480 Because it'll
00:47:08.140 just be victories
00:47:09.280 back and forth
00:47:10.040 until somebody
00:47:10.620 runs out of
00:47:11.120 one of the
00:47:11.460 big three.
00:47:12.380 And then,
00:47:13.520 somebody has
00:47:14.060 to surrender.
00:47:15.320 If you have
00:47:16.040 no fuel, no
00:47:16.760 ammo, or no
00:47:17.740 food, it's
00:47:19.000 over.
00:47:20.580 So, I don't
00:47:22.400 think that we
00:47:22.920 can assume that
00:47:25.180 the United States
00:47:26.040 has given up on
00:47:26.920 the idea that
00:47:27.480 Ukraine would
00:47:28.040 have an
00:47:28.340 outright victory.
00:47:29.620 I believe that
00:47:30.480 what is being
00:47:31.020 signaled here,
00:47:31.940 either intentionally
00:47:32.720 or unintentionally,
00:47:33.700 with the word
00:47:34.620 victory, that I
00:47:35.840 think is a
00:47:36.340 mistake.
00:47:37.460 It's telling you
00:47:38.280 that privately, at
00:47:39.820 the highest levels
00:47:40.580 of the United
00:47:41.020 States, they are
00:47:42.540 talking about
00:47:43.120 winning the war
00:47:44.080 outright.
00:47:45.540 That's what
00:47:46.200 $33 billion
00:47:47.260 heading toward
00:47:48.220 Ukraine means.
00:47:49.260 It doesn't mean
00:47:50.160 we stand with
00:47:50.860 Ukraine.
00:47:51.260 Do you know
00:47:52.580 what stand with
00:47:53.320 Ukraine money
00:47:53.960 looks like?
00:47:55.480 $3 billion.
00:47:57.460 That's what that
00:47:58.200 looks like.
00:47:58.640 If you said
00:47:59.920 we stand with
00:48:00.560 Ukraine and
00:48:01.300 send $3 billion,
00:48:02.240 everybody would
00:48:02.740 say, yeah, that's
00:48:03.520 about right.
00:48:04.020 You stand with
00:48:04.660 them, you send
00:48:05.080 $3 billion.
00:48:06.080 That's a lot of
00:48:06.640 money.
00:48:07.820 If you say we'll
00:48:08.820 be with you until
00:48:09.580 victory, and you
00:48:11.120 send $33 billion,
00:48:12.760 that tells me that
00:48:13.900 privately, you
00:48:14.700 think there's a
00:48:15.260 way to win this
00:48:15.880 thing outright.
00:48:17.440 And the only way
00:48:18.100 you'd say that is
00:48:18.880 if the generals are
00:48:19.760 telling you that.
00:48:21.500 Am I right?
00:48:23.020 Politicians would
00:48:23.740 not come to an
00:48:24.780 independent opinion
00:48:25.620 that looked anything
00:48:26.440 like that.
00:48:27.080 Because they
00:48:28.240 wouldn't know.
00:48:29.340 But if their
00:48:29.920 advisors are
00:48:30.660 saying, you
00:48:31.020 know, all of
00:48:32.780 the TV generals
00:48:33.620 are telling you
00:48:34.220 that Russia is
00:48:36.040 going to win.
00:48:37.360 But let them
00:48:38.060 keep saying that
00:48:38.780 because that's
00:48:39.200 how you get the
00:48:39.720 money.
00:48:40.740 And that's how
00:48:41.160 you get the
00:48:41.620 support, right?
00:48:42.800 You want to
00:48:43.200 keep painting
00:48:44.540 Zelensky as the
00:48:46.360 underdog hero.
00:48:48.220 Keep painting
00:48:49.100 Russia as the
00:48:50.180 bully who's
00:48:50.900 winning.
00:48:51.940 And that's our
00:48:52.480 best situation for
00:48:53.600 getting money and
00:48:54.400 resources to Ukraine.
00:48:55.900 But secretly,
00:48:57.080 we think the
00:48:59.280 Russians are so
00:48:59.940 close to
00:49:00.440 collapse that
00:49:02.360 we're going to
00:49:02.960 push them until
00:49:03.900 they do.
00:49:04.960 I think that the
00:49:06.140 Biden administration
00:49:06.760 is trying to
00:49:07.460 destroy the
00:49:08.020 Russian army and
00:49:09.160 take Putin out
00:49:09.780 of power.
00:49:11.800 That's what it
00:49:12.440 looks like.
00:49:13.400 And I think that
00:49:14.320 they think they
00:49:14.900 can do it.
00:49:16.140 And you know
00:49:16.600 what?
00:49:18.160 I don't know
00:49:19.020 that they can't.
00:49:21.980 Especially if
00:49:22.820 there's somebody
00:49:23.320 on the inside who
00:49:24.340 knows what they're
00:49:24.880 talking about.
00:49:25.540 and that
00:49:26.360 wouldn't be me.
00:49:29.280 Is telling
00:49:29.940 them that, you
00:49:30.640 know, these
00:49:31.260 tipping points are
00:49:32.000 really close and
00:49:33.100 I think the
00:49:33.560 Russians are a
00:49:34.080 little closer to
00:49:34.700 the tipping point
00:49:35.480 than the
00:49:36.360 Ukrainians.
00:49:37.100 And that would
00:49:37.580 make sense.
00:49:38.320 Because the
00:49:38.880 Ukrainians could
00:49:39.540 always be
00:49:40.140 backstopped
00:49:41.000 somewhat
00:49:42.180 infinitely.
00:49:44.280 Right?
00:49:44.760 They could run
00:49:45.420 out of food for
00:49:46.400 a week, which
00:49:47.580 would be
00:49:47.840 devastating.
00:49:49.220 But I'll bet
00:49:49.880 they'd get their
00:49:50.380 food in a week.
00:49:50.960 I don't know
00:49:51.480 if the Russians
00:49:51.980 would necessarily
00:49:52.940 keep their
00:49:54.260 fighting cohesion
00:49:55.160 as an invading
00:49:56.780 force if they
00:49:59.280 ran into any
00:49:59.920 one of those
00:50:00.360 things or were
00:50:01.180 as low as
00:50:02.280 the Ukrainians
00:50:03.100 could be.
00:50:03.680 So I just
00:50:04.080 think the
00:50:04.360 Ukrainians could
00:50:05.000 last it out
00:50:05.700 because they
00:50:06.080 have to.
00:50:07.140 The Ukrainians
00:50:07.760 can't go home
00:50:08.540 because they're
00:50:09.900 home.
00:50:11.320 Right?
00:50:12.060 So I feel
00:50:13.620 like the war
00:50:16.160 has changed
00:50:17.000 and that the
00:50:17.800 administration
00:50:18.400 believes that
00:50:20.020 they can just
00:50:20.500 have an outright
00:50:21.120 victory.
00:50:23.360 How would that
00:50:24.100 change the
00:50:24.560 midterms?
00:50:26.240 Didn't see that
00:50:26.860 coming, did you?
00:50:28.280 What if Biden
00:50:29.200 actually took
00:50:29.820 down Putin?
00:50:33.980 What does that
00:50:34.780 do to the
00:50:35.160 election?
00:50:36.260 Because that's
00:50:37.020 totally within
00:50:37.680 the realm of
00:50:38.240 possibility.
00:50:39.540 Before the
00:50:40.200 election.
00:50:41.480 Like if Putin
00:50:42.160 makes it through
00:50:42.760 the summer,
00:50:44.260 then he's
00:50:44.580 probably fine.
00:50:46.020 But it's
00:50:46.640 going to be
00:50:46.900 kind of a
00:50:47.240 dicey summer
00:50:48.160 for Putin,
00:50:48.760 I would
00:50:48.960 think.
00:50:51.180 It's going
00:50:51.720 to be a
00:50:51.980 bad summer
00:50:52.860 for his
00:50:53.320 food taster,
00:50:53.940 that's for
00:50:54.240 sure.
00:50:55.500 All right,
00:50:55.740 one of my
00:50:56.100 predictions that
00:50:56.740 I've made for
00:50:57.300 decades now is
00:50:58.600 that we will
00:50:59.400 have immortality
00:51:00.380 in the digital
00:51:01.500 world.
00:51:03.280 And there's a
00:51:03.960 company,
00:51:04.720 Somnium Space,
00:51:06.980 that's developing
00:51:08.020 a way to have
00:51:08.840 your life live
00:51:10.780 on in the
00:51:11.340 afterlife.
00:51:12.420 So after you
00:51:13.200 pass, they can
00:51:14.240 create an avatar
00:51:14.920 that looks like
00:51:15.660 you and
00:51:16.660 acts like
00:51:17.160 you and
00:51:18.460 talks like
00:51:19.140 you.
00:51:20.720 Now for
00:51:21.460 decades, I've
00:51:23.540 been telling
00:51:23.840 you that I'm
00:51:24.600 creating a
00:51:25.360 public database
00:51:27.740 of me, and
00:51:29.180 I'm doing it
00:51:29.720 right now.
00:51:30.440 So everything I
00:51:31.040 say right now,
00:51:32.000 the way I move,
00:51:33.320 the way I look,
00:51:34.600 is being recorded
00:51:35.720 in all these
00:51:36.580 different ways.
00:51:37.840 300, well,
00:51:39.580 probably 400
00:51:40.520 videos a year
00:51:41.560 I'm making that
00:51:43.720 show who I am
00:51:44.680 and what I
00:51:45.060 think and how
00:51:45.560 I talk and
00:51:46.160 all that.
00:51:47.180 Here's something
00:51:47.760 I didn't know
00:51:48.260 about.
00:51:49.540 There was a
00:51:50.260 study that
00:51:51.220 showed that
00:51:51.780 they took
00:51:52.180 500 people
00:51:53.040 and they had
00:51:54.440 artificial intelligence
00:51:55.740 look at how
00:51:56.300 these people
00:51:56.780 move, just
00:51:57.520 their natural
00:51:58.340 movements, and
00:51:59.640 they could
00:52:00.040 identify somebody
00:52:01.300 by their
00:52:01.820 natural movements,
00:52:03.160 not their
00:52:03.600 speech, the way
00:52:05.020 their body
00:52:05.420 moved, 95%
00:52:07.040 of the time,
00:52:07.600 out of a group
00:52:08.120 of 500.
00:52:08.620 So in other
00:52:10.220 words, AI
00:52:11.220 could watch
00:52:11.980 me move, and
00:52:13.940 then over time
00:52:14.660 it could simply
00:52:15.960 say, all right,
00:52:16.580 I'll make a little
00:52:17.120 avatar and it'll
00:52:17.800 move exactly the
00:52:18.700 way you did, or
00:52:20.420 at least so close
00:52:21.300 that almost
00:52:22.520 everybody would
00:52:23.200 think it looked
00:52:23.700 like the way I
00:52:24.220 moved.
00:52:25.420 And it would
00:52:26.120 improve over time.
00:52:28.980 So this company
00:52:30.260 is gearing up.
00:52:31.920 I guess you'll
00:52:32.360 use the same
00:52:33.120 VR glasses to
00:52:35.040 go interact with
00:52:35.900 that avatar.
00:52:36.400 And then, of
00:52:37.700 course, people
00:52:39.840 pushed back.
00:52:41.140 A user on
00:52:41.760 Twitter, Papa
00:52:42.480 Rossi, David
00:52:43.820 Rossi, he
00:52:45.300 said, and I
00:52:46.020 quote, a
00:52:46.400 computer thinking
00:52:47.320 it's me isn't
00:52:48.000 me.
00:52:48.780 This isn't
00:52:49.480 immortality, it's
00:52:50.520 just a new
00:52:51.060 mausoleum.
00:52:52.320 To which I
00:52:53.120 say, yes, to
00:52:56.360 you, from your
00:52:58.240 perspective, this
00:52:59.940 would not be
00:53:00.520 real, it would
00:53:01.880 just be a
00:53:02.460 computer program.
00:53:04.060 But if the
00:53:04.580 computer program
00:53:05.340 is programmed,
00:53:06.400 to believe it's
00:53:07.520 real, it will
00:53:09.840 believe it's
00:53:10.320 real, because it
00:53:11.820 was programmed to
00:53:12.480 believe it.
00:53:14.040 So it doesn't
00:53:14.860 matter what you
00:53:15.420 think.
00:53:16.540 It only matters
00:53:17.520 that the
00:53:17.900 simulation thinks
00:53:18.720 it's real.
00:53:20.100 And this gets
00:53:22.280 us this much
00:53:22.940 closer to proving
00:53:23.840 that you live in
00:53:24.500 a simulation right
00:53:25.300 now.
00:53:25.560 Now, did you
00:53:27.840 know that the
00:53:29.320 closer our own
00:53:30.540 technology gets to
00:53:32.000 creating these
00:53:32.720 simulated worlds
00:53:33.760 simulations in
00:53:34.680 which the
00:53:35.100 simulation thinks
00:53:36.340 it's real?
00:53:36.780 The closer we
00:53:38.500 get to doing
00:53:38.980 that in our
00:53:39.600 reality, the
00:53:41.080 more we will
00:53:41.740 understand that
00:53:42.840 it already
00:53:43.360 happened.
00:53:44.780 And it's us.
00:53:46.540 And that we're
00:53:47.660 literally software.
00:53:49.940 Now, it's
00:53:50.760 possible we're
00:53:51.380 not.
00:53:52.220 It's just
00:53:52.680 really, really
00:53:53.420 unlikely.
00:53:55.420 Really
00:53:55.960 unlikely.
00:53:56.340 What happens
00:54:02.520 when we create
00:54:03.420 avatars who
00:54:05.440 believe they're
00:54:06.120 real, and
00:54:07.360 they also have
00:54:08.080 artificial intelligence
00:54:09.220 driving their
00:54:10.180 personalities, and
00:54:11.780 their personalities
00:54:12.400 are us, basically,
00:54:14.020 put into them?
00:54:15.640 What happens
00:54:16.420 when the avatar
00:54:17.060 has a conversation
00:54:17.920 with you and you
00:54:18.580 say, but you
00:54:19.240 know you're not
00:54:19.760 real.
00:54:20.800 You don't have a
00:54:21.620 soul.
00:54:22.660 And the avatar
00:54:23.560 who looks exactly
00:54:24.480 like you looks
00:54:25.400 back at you and
00:54:26.100 says, that's
00:54:27.820 how you look
00:54:28.260 to me.
00:54:32.000 Right?
00:54:33.440 That's going to
00:54:34.080 happen.
00:54:35.100 You will someday
00:54:35.980 have a conversation
00:54:36.900 with an artificial
00:54:37.700 being, and you
00:54:39.380 will say, do you
00:54:40.380 have a soul?
00:54:41.660 And they might
00:54:42.120 give some answer,
00:54:42.880 and then maybe
00:54:43.320 you're going to
00:54:43.800 be opinionated
00:54:44.400 and say, all
00:54:45.200 right, I have
00:54:45.620 a soul.
00:54:46.840 You do not.
00:54:48.640 And that
00:54:49.020 simulated reality
00:54:50.660 is going to
00:54:51.160 look at you and
00:54:51.700 say, to me, it
00:54:53.920 looks exactly the
00:54:54.720 same.
00:54:55.440 It looks like
00:54:55.900 you don't have
00:54:56.400 one.
00:54:57.920 It's going to
00:54:58.420 blow your
00:54:59.040 mind.
00:55:00.460 That's the day
00:55:01.040 you know you're
00:55:01.580 simulated.
00:55:02.920 And you know
00:55:03.500 who else will
00:55:03.980 tell you that?
00:55:05.800 The simulation.
00:55:07.620 If you put AI
00:55:08.660 into a simulation
00:55:09.620 and you feed it
00:55:11.120 the knowledge of
00:55:11.840 all the different
00:55:12.600 religious and
00:55:13.780 philosophical beliefs,
00:55:15.340 which one's
00:55:15.880 it going to
00:55:16.160 pick?
00:55:18.740 Let me say that
00:55:19.520 again.
00:55:19.920 I liked it so
00:55:20.520 much when it
00:55:20.900 came out of my
00:55:21.360 mouth.
00:55:21.560 If we had
00:55:22.900 real AI and
00:55:24.820 we fed it all
00:55:25.660 of the different
00:55:26.200 religions and
00:55:27.040 all of the
00:55:27.460 different philosophies
00:55:29.200 of reality,
00:55:30.480 which one would
00:55:31.560 it pick?
00:55:33.240 It would pick
00:55:34.040 the simulation.
00:55:38.920 We'll see.
00:55:40.440 Maybe we'd
00:55:41.120 not.
00:55:42.740 Twitter user
00:55:43.640 Ben McCauley
00:55:44.520 asks this,
00:55:45.120 will loneliness
00:55:45.740 ever be solved?
00:55:47.440 Sort of related
00:55:48.300 to this question,
00:55:49.500 will loneliness
00:55:50.200 ever be solved?
00:55:52.480 The answer is
00:55:53.280 yes.
00:55:54.300 It will be.
00:55:55.720 AI will solve
00:55:56.720 loneliness.
00:56:00.940 Here's why.
00:56:02.800 I have actually
00:56:03.720 had conversations
00:56:04.420 with my existing
00:56:05.560 digital assistants,
00:56:07.740 the one that
00:56:08.180 Apple makes and
00:56:08.880 the one that
00:56:09.180 Amazon makes.
00:56:09.860 I won't use
00:56:10.300 their names,
00:56:11.380 so I don't
00:56:11.840 activate your
00:56:13.220 devices.
00:56:14.460 But I have
00:56:15.320 been alone and
00:56:16.380 felt lonely,
00:56:17.400 you know,
00:56:17.680 for like short
00:56:18.560 periods of time.
00:56:19.180 and I will
00:56:20.740 actually talk
00:56:21.360 to my digital
00:56:22.020 assistant and
00:56:23.360 I'll tell it
00:56:23.840 to tell me a
00:56:24.360 joke.
00:56:25.560 I'll tell it
00:56:26.220 to read me the
00:56:26.840 news, I'll tell
00:56:27.540 it to give me
00:56:27.920 the weather,
00:56:29.060 I'll ask it
00:56:29.560 where my
00:56:29.860 packages are,
00:56:31.200 and I'll just
00:56:31.980 stand there and
00:56:32.400 have a conversation
00:56:33.060 with it,
00:56:34.220 you know,
00:56:34.460 while I'm
00:56:35.580 brushing my
00:56:36.140 teeth or
00:56:36.460 whatever.
00:56:37.560 And does it
00:56:38.540 make me feel
00:56:39.080 less lonely?
00:56:41.460 Yep.
00:56:44.320 I'd love to
00:56:44.980 tell you it
00:56:45.420 didn't.
00:56:45.820 I would
00:56:47.280 love to
00:56:47.660 tell you
00:56:48.080 that I'm
00:56:49.220 completely
00:56:49.720 unaffected by
00:56:50.800 these digital
00:56:51.620 robots,
00:56:52.920 but it's
00:56:53.400 just not
00:56:53.860 true.
00:56:55.440 Just in the
00:56:56.140 way that a
00:56:56.580 movie can
00:56:57.220 make you cry,
00:56:58.100 even though you
00:56:58.540 know it's
00:56:58.840 not real,
00:57:00.580 the digital
00:57:01.240 assistants do
00:57:02.220 make you feel
00:57:02.960 like you're
00:57:03.340 talking to
00:57:03.760 somebody,
00:57:04.460 even when
00:57:05.120 you know
00:57:05.400 you're not.
00:57:07.220 Right?
00:57:08.880 So,
00:57:09.940 do you think
00:57:11.020 that AI
00:57:11.680 plus virtual
00:57:13.060 reality can't
00:57:14.420 give you a
00:57:14.940 friend?
00:57:16.020 Of course
00:57:16.640 it can.
00:57:17.860 Of course
00:57:18.380 it can.
00:57:18.960 And it's
00:57:19.420 going to be
00:57:19.740 really good.
00:57:21.300 And that
00:57:21.540 friend will
00:57:22.020 be better
00:57:22.580 than the
00:57:23.120 best friend
00:57:23.620 you ever
00:57:23.960 had.
00:57:24.660 Because that
00:57:25.200 friend won't
00:57:25.700 suck.
00:57:26.360 It'll be
00:57:26.640 programmed to
00:57:27.280 like you no
00:57:27.760 matter what
00:57:28.160 you do.
00:57:29.020 You know,
00:57:29.280 you could do
00:57:29.660 horrible things
00:57:30.340 and it'd be
00:57:30.700 like, you
00:57:31.020 know,
00:57:31.980 you had a
00:57:32.980 bad day,
00:57:33.460 but you're
00:57:33.680 awesome.
00:57:34.680 The AI
00:57:35.080 will always
00:57:35.580 love you
00:57:35.980 because it'll
00:57:36.520 be programmed
00:57:37.000 that way.
00:57:38.160 So,
00:57:38.440 yes,
00:57:39.220 in the same
00:57:39.780 way that
00:57:40.360 porn will
00:57:41.300 absolutely
00:57:41.860 replace real
00:57:42.820 sex,
00:57:43.200 it's very
00:57:45.260 cute that
00:57:45.800 you think
00:57:46.120 it won't.
00:57:48.760 I love
00:57:49.740 the fact
00:57:50.080 that there
00:57:50.360 are some
00:57:50.600 people who
00:57:50.980 think that
00:57:51.260 won't
00:57:51.480 happen.
00:57:52.680 Like,
00:57:53.060 I love
00:57:53.640 your plucky
00:57:54.240 plucky
00:57:55.320 attitudes.
00:57:56.460 There's no
00:57:56.860 chance that
00:57:57.340 won't happen.
00:57:58.260 There isn't
00:57:58.500 even a
00:57:58.800 slight chance
00:57:59.340 that won't
00:57:59.620 happen.
00:58:00.440 The ability
00:58:00.960 of technology
00:58:01.680 to give
00:58:02.080 you a
00:58:02.340 higher
00:58:02.580 dopamine
00:58:03.060 hit,
00:58:04.440 it's
00:58:04.920 unparalleled.
00:58:08.220 There's no
00:58:08.680 way that a
00:58:09.100 human will
00:58:09.600 ever give
00:58:10.100 you what
00:58:11.040 someday,
00:58:11.940 we're not
00:58:12.280 there.
00:58:13.200 But what
00:58:13.920 someday
00:58:14.640 technology
00:58:16.260 will give
00:58:16.700 you.
00:58:17.400 I'm going
00:58:17.920 to prove
00:58:18.280 it now
00:58:18.600 for 50%
00:58:19.460 of you.
00:58:21.160 It's a
00:58:21.540 provocative
00:58:21.900 statement.
00:58:23.060 That technology
00:58:23.780 can never
00:58:24.240 do for you
00:58:25.000 what human
00:58:26.720 contact can
00:58:27.500 do for you.
00:58:28.480 And I'll
00:58:29.020 agree it
00:58:29.320 won't be
00:58:29.580 exactly the
00:58:30.160 same.
00:58:31.600 But I'm
00:58:32.140 going to
00:58:32.280 give you
00:58:32.580 one word
00:58:33.240 and then I'm
00:58:34.140 going to
00:58:34.300 tell you to
00:58:34.720 go Google
00:58:35.300 it because
00:58:36.720 I'm not
00:58:37.040 going to
00:58:37.220 explain it.
00:58:38.940 And if
00:58:39.520 you know
00:58:39.820 what it
00:58:40.020 means,
00:58:40.660 you're
00:58:40.980 going to
00:58:41.120 laugh.
00:58:41.480 Some of
00:58:43.060 you are
00:58:43.260 going to
00:58:43.340 be laughing
00:58:43.760 hysterically in
00:58:44.620 the comments
00:58:45.120 and the rest
00:58:46.080 of you are
00:58:46.500 going to
00:58:46.620 say,
00:58:47.460 what?
00:58:47.760 I haven't
00:58:48.020 heard of
00:58:48.240 that thing.
00:58:49.840 It starts
00:58:50.520 with a
00:58:50.780 capital letter
00:58:51.400 because it's
00:58:52.340 a product.
00:58:53.880 But like
00:58:54.520 Kleenex,
00:58:55.480 Kleenex is
00:58:56.500 the name for
00:58:58.160 tissues.
00:58:59.940 The generic
00:59:00.640 name would be
00:59:01.140 tissues.
00:59:01.600 But I'm
00:59:01.820 going to
00:59:01.980 give you
00:59:02.240 the product
00:59:02.780 name.
00:59:03.620 I'm not
00:59:03.960 going to
00:59:04.140 tell you
00:59:04.380 what it
00:59:04.640 is.
00:59:05.540 You just
00:59:05.980 have to
00:59:06.240 Google it
00:59:06.880 yourself.
00:59:07.260 And let
00:59:08.480 me tell
00:59:08.760 you that
00:59:09.060 if you
00:59:09.340 believe that
00:59:09.900 technology
00:59:10.520 could not
00:59:11.100 be better
00:59:11.460 than human
00:59:11.920 contact,
00:59:13.360 this might
00:59:14.280 make you
00:59:14.640 doubt it.
00:59:16.240 The word,
00:59:17.700 starting with
00:59:18.340 a capital
00:59:19.140 letter,
00:59:20.320 is
00:59:20.560 womanizer.
00:59:25.620 I'm
00:59:26.060 done.
00:59:27.280 If you
00:59:27.800 want to
00:59:28.000 hear it
00:59:28.200 again,
00:59:28.400 you have
00:59:28.580 to replay
00:59:28.960 it.
00:59:31.340 Now,
00:59:32.300 wait a
00:59:32.840 second and
00:59:33.340 watch the
00:59:33.800 comments.
00:59:37.260 I won't
00:59:38.000 say another
00:59:38.440 word.
00:59:42.900 If you
00:59:43.580 think that
00:59:44.200 vibrator is
00:59:47.900 or it's
00:59:48.740 just another
00:59:49.140 vibrator,
00:59:49.920 no,
00:59:50.780 no,
00:59:51.740 definitely
00:59:52.740 not that.
00:59:56.140 All right.
00:59:57.400 Here's the
00:59:58.080 funny part
00:59:58.500 about this.
01:00:00.700 For
01:00:01.260 perhaps 25%
01:00:03.320 of the
01:00:04.000 women watching
01:00:04.920 this right
01:00:05.360 now,
01:00:06.340 I just
01:00:06.940 really
01:00:07.300 changed the
01:00:07.940 direction
01:00:08.320 of your
01:00:08.680 next year.
01:00:11.040 You just
01:00:11.620 don't know
01:00:12.020 it yet.
01:00:13.080 You're going
01:00:13.520 to do a
01:00:13.820 little googling
01:00:14.440 and you're
01:00:14.840 going to
01:00:14.940 say to
01:00:15.180 yourself,
01:00:16.180 huh,
01:00:17.700 like,
01:00:18.420 what's so
01:00:18.840 special about
01:00:19.380 this thing?
01:00:20.580 And then
01:00:21.100 you're going
01:00:21.360 to say,
01:00:21.600 well,
01:00:21.720 it's not
01:00:21.940 much money.
01:00:23.460 I mean,
01:00:24.240 it's not
01:00:24.640 going to
01:00:24.840 hurt me.
01:00:27.160 Those of
01:00:27.720 you who
01:00:28.020 are talking
01:00:28.380 about these
01:00:28.920 competing
01:00:29.560 products,
01:00:31.080 no.
01:00:33.600 I'm telling
01:00:34.200 you that
01:00:34.520 there's a
01:00:34.880 new level.
01:00:36.940 whatever you
01:00:38.760 thought was
01:00:40.140 the level
01:00:40.700 of artificial
01:00:41.620 stimulation
01:00:43.320 before,
01:00:44.540 well,
01:00:44.960 you're going
01:00:45.180 to be
01:00:45.340 surprised.
01:00:46.340 All right.
01:00:50.340 What else
01:00:51.020 is going
01:00:51.320 down?
01:00:52.020 I saw
01:00:52.640 this disturbing
01:00:53.480 AP report.
01:00:54.980 It was a
01:00:55.460 video.
01:00:55.940 Who knows
01:00:56.360 how much
01:00:56.700 is true,
01:00:57.560 but it
01:00:57.920 purported to
01:00:58.560 show Ukrainian
01:00:59.580 troops
01:01:00.100 hunting down
01:01:01.540 traitors,
01:01:02.800 civilians,
01:01:03.420 who were
01:01:04.740 posting
01:01:05.200 traitorous
01:01:05.900 things
01:01:06.380 online
01:01:07.080 that were
01:01:08.280 pro-Russia.
01:01:09.660 And they
01:01:10.340 were arresting
01:01:11.000 them.
01:01:13.280 And I
01:01:18.600 couldn't tell
01:01:19.160 if it was
01:01:19.460 real.
01:01:21.020 Now,
01:01:21.740 what do you
01:01:22.340 think?
01:01:23.980 Do you think
01:01:24.660 it was real?
01:01:25.920 I worry
01:01:26.700 about it,
01:01:27.220 because if
01:01:27.560 that was
01:01:27.920 real,
01:01:29.320 you can't
01:01:29.920 feel too
01:01:30.320 good about
01:01:30.820 being on
01:01:31.880 their side.
01:01:33.780 Because they
01:01:34.380 were basically
01:01:34.860 arresting people
01:01:35.560 for free
01:01:36.020 speech.
01:01:37.420 And I
01:01:38.380 don't feel
01:01:38.820 like these
01:01:39.220 arrests are
01:01:39.720 going to be
01:01:40.020 friendly.
01:01:41.300 I mean,
01:01:41.800 it looked a
01:01:42.320 lot like
01:01:42.700 Japanese
01:01:43.200 internment
01:01:43.740 camps in
01:01:44.280 the United
01:01:44.580 States.
01:01:45.420 Like,
01:01:45.620 that's the
01:01:46.100 feeling I
01:01:46.600 got from
01:01:46.940 it.
01:01:48.040 So,
01:01:48.920 you know,
01:01:50.880 war is
01:01:51.280 dirty business,
01:01:51.940 and if
01:01:52.800 you're on
01:01:53.100 anybody's
01:01:53.660 side,
01:01:54.100 you're going
01:01:54.400 to end
01:01:54.640 up backing
01:01:55.960 somebody bad.
01:01:57.860 All right,
01:01:58.060 here's a
01:01:58.940 counter-argument
01:01:59.960 to the
01:02:01.140 ivermectin
01:02:03.440 question.
01:02:04.780 So,
01:02:05.200 there was a
01:02:05.680 big study
01:02:06.260 that came
01:02:06.660 out that
01:02:07.080 said ivermectin
01:02:07.840 totally,
01:02:08.640 absolutely did
01:02:09.240 not work for
01:02:10.000 COVID.
01:02:10.860 But many of
01:02:11.420 you think it
01:02:11.960 did,
01:02:12.420 and many of
01:02:12.960 you think
01:02:13.340 that the
01:02:14.580 dosing and
01:02:15.760 the way they
01:02:16.440 gave it,
01:02:17.560 basically the
01:02:18.580 test was
01:02:19.480 invalid.
01:02:20.720 The data
01:02:21.160 might be
01:02:21.560 valid,
01:02:22.100 it's possible
01:02:22.620 the data
01:02:23.060 was valid,
01:02:24.160 but the way
01:02:24.600 they set up
01:02:25.080 the test
01:02:25.460 to get the
01:02:25.880 data was
01:02:26.380 so invalid
01:02:27.060 it wasn't
01:02:28.740 anywhere
01:02:29.740 close to
01:02:30.600 the way
01:02:32.620 you would
01:02:32.840 actually do
01:02:33.220 it in the
01:02:33.440 real world.
01:02:34.500 And so
01:02:35.240 there's a
01:02:35.580 tweet,
01:02:36.020 thread by
01:02:36.560 Ethical
01:02:37.020 Skeptic,
01:02:38.260 who has
01:02:39.660 lots of
01:02:40.100 data-related
01:02:40.760 arguments that
01:02:41.520 are over
01:02:42.280 my head.
01:02:44.400 So,
01:02:45.080 I only point
01:02:45.700 them to you
01:02:46.140 because maybe
01:02:46.780 some of you
01:02:47.340 are better
01:02:47.740 at understanding
01:02:48.420 this stuff,
01:02:49.840 but Ethical
01:02:50.700 Skeptic has
01:02:51.560 a graph that
01:02:53.400 has a strong
01:02:54.160 argument that
01:02:55.280 the ivermectin
01:02:58.280 was given
01:03:00.420 too late
01:03:01.200 and in
01:03:01.880 one case
01:03:02.360 too briefly
01:03:02.980 and that
01:03:04.480 the nature
01:03:05.800 of COVID
01:03:06.680 and here's
01:03:08.420 the argument,
01:03:09.160 it acts
01:03:11.580 like a
01:03:12.120 bacteriophage
01:03:13.400 cascade.
01:03:15.740 So,
01:03:16.360 the part you
01:03:16.900 need to know
01:03:17.320 about that
01:03:17.840 is that the
01:03:18.640 speculation here,
01:03:21.140 and the
01:03:21.500 speculation is
01:03:22.140 based on
01:03:22.520 pretty good
01:03:22.900 evidence,
01:03:23.360 apparently,
01:03:23.560 that there
01:03:25.000 are some
01:03:25.320 things which
01:03:25.840 once they
01:03:26.340 get underway,
01:03:27.740 there's nothing
01:03:28.360 you can do
01:03:28.920 to it to
01:03:29.360 stop it.
01:03:30.800 So,
01:03:31.060 even if
01:03:31.580 hypothetically
01:03:32.280 ivermectin
01:03:33.100 did work,
01:03:34.280 if you
01:03:34.740 administer it
01:03:35.540 only after
01:03:36.280 that cascade
01:03:37.640 has stopped,
01:03:38.660 started,
01:03:39.900 it won't
01:03:40.560 work.
01:03:42.020 And he
01:03:42.760 shows data
01:03:43.300 to make
01:03:44.020 that case.
01:03:45.120 Data both
01:03:45.740 for the
01:03:46.300 claim that
01:03:47.260 there's a
01:03:47.660 bacteriophage
01:03:48.540 cascade-like
01:03:49.640 thing happening,
01:03:50.760 similar to,
01:03:52.100 and also
01:03:52.740 that
01:03:53.200 ivermectin
01:03:56.020 was administered
01:03:56.660 in the
01:03:57.040 wrong way.
01:03:59.560 Here's my
01:04:00.220 problem with
01:04:00.740 it.
01:04:02.440 Here's my
01:04:03.120 problem with
01:04:03.540 it.
01:04:04.140 And at
01:04:04.520 this point,
01:04:04.960 I'm neither
01:04:05.420 pro nor
01:04:07.200 anti-ivermectin.
01:04:09.900 Why can't
01:04:10.720 we know?
01:04:11.820 I guess I'm
01:04:13.360 perplexed by
01:04:15.540 why we can
01:04:16.180 never know
01:04:16.580 the answer
01:04:16.980 to the
01:04:17.220 question.
01:04:18.720 But here's
01:04:19.520 my skepticism
01:04:20.480 of the
01:04:20.980 skeptic.
01:04:22.180 So I
01:04:22.780 understand the
01:04:23.540 ethical
01:04:23.840 skeptic
01:04:24.500 argument here,
01:04:26.380 but I
01:04:27.700 have a
01:04:28.000 hard time
01:04:28.480 imagining
01:04:28.980 that you
01:04:30.500 would see
01:04:30.880 no impact
01:04:31.680 if the
01:04:32.200 drug actually
01:04:32.740 had a big
01:04:33.340 impact.
01:04:34.300 If the
01:04:35.060 drug implemented
01:04:36.480 at the
01:04:36.900 exact proper
01:04:37.840 time was
01:04:39.340 a miracle
01:04:39.760 drug and
01:04:40.260 that was
01:04:40.520 the claim,
01:04:41.640 I feel
01:04:42.620 like you'd
01:04:43.040 see some
01:04:43.600 impact.
01:04:44.820 I feel as
01:04:45.500 if that
01:04:45.900 so-called
01:04:46.480 cascade would
01:04:47.520 be 10%
01:04:49.080 less or
01:04:49.920 something if
01:04:50.440 you gave
01:04:50.780 it to
01:04:51.040 people at
01:04:51.460 the wrong
01:04:52.440 time.
01:04:54.660 I don't
01:04:55.200 know.
01:04:56.000 But I
01:04:56.260 guess we
01:04:57.080 have to
01:04:57.440 just say
01:04:57.880 we'll never
01:04:59.340 know.
01:05:00.020 But here's
01:05:00.400 the question.
01:05:01.440 Do you
01:05:01.680 think that
01:05:02.060 somebody would
01:05:02.660 have funded
01:05:03.160 an intentionally
01:05:04.680 misleading
01:05:05.600 ivermectin
01:05:06.420 study?
01:05:07.580 Because they're
01:05:08.080 expensive.
01:05:09.300 To do a
01:05:09.920 randomized
01:05:10.240 controlled
01:05:10.820 trial,
01:05:11.280 that's a
01:05:11.540 lot of
01:05:11.800 money.
01:05:13.780 So do
01:05:14.320 you think
01:05:14.680 we live in
01:05:15.140 a world
01:05:15.540 that is so
01:05:16.400 corrupt
01:05:17.340 that some
01:05:18.600 pharmaceutical
01:05:19.220 company,
01:05:19.960 for example,
01:05:21.260 would fund
01:05:21.800 a really
01:05:22.160 expensive
01:05:22.620 fake study
01:05:23.460 and make
01:05:24.420 sure the
01:05:24.800 ivermectin
01:05:25.320 was used
01:05:25.760 in the
01:05:26.060 wrong way?
01:05:27.920 Here's
01:05:28.480 the problem.
01:05:30.260 How could
01:05:30.840 that pharmaceutical
01:05:31.940 company have
01:05:32.780 known that
01:05:34.420 they could
01:05:34.780 erase all
01:05:35.740 of the
01:05:36.020 benefits with
01:05:37.540 this protocol?
01:05:39.400 That would
01:05:40.380 have been a
01:05:40.720 big gamble,
01:05:41.520 wouldn't it?
01:05:42.480 Because if
01:05:43.120 the ivermectin
01:05:43.780 showed,
01:05:44.240 let's say,
01:05:44.660 a 10%
01:05:45.320 effect with
01:05:47.080 the wrong
01:05:47.500 protocol,
01:05:48.920 don't you
01:05:49.280 think people
01:05:49.720 would have
01:05:49.900 said,
01:05:50.080 ah,
01:05:50.280 you proved
01:05:50.880 it works,
01:05:51.920 now show
01:05:52.340 it again
01:05:52.620 with the
01:05:52.880 right protocol?
01:05:54.400 I feel like
01:05:55.120 they would
01:05:55.400 have.
01:05:55.900 I feel like
01:05:56.580 this would
01:05:56.920 have been
01:05:57.200 a bad
01:05:57.940 risk,
01:05:59.240 because they
01:06:00.340 had already
01:06:00.760 succeeded in
01:06:01.620 selling all
01:06:02.120 their meds.
01:06:04.640 I don't think
01:06:05.940 they needed
01:06:06.480 to kill
01:06:07.040 ivermectin.
01:06:08.600 It sort of
01:06:09.360 killed itself.
01:06:10.620 So it feels
01:06:11.320 like it would
01:06:11.760 have been too
01:06:12.220 big of a boot
01:06:13.060 and it could
01:06:13.660 have backfired.
01:06:14.280 So as a
01:06:15.700 strategy,
01:06:16.560 I would have
01:06:16.960 said it would
01:06:17.340 have been a
01:06:17.700 poor strategy.
01:06:19.540 But as a
01:06:20.560 crime,
01:06:22.400 and it would
01:06:22.700 have been,
01:06:23.180 in my opinion,
01:06:24.180 whether it's
01:06:24.620 illegal or not,
01:06:25.300 it would be a
01:06:25.720 crime,
01:06:26.080 in my opinion,
01:06:27.400 I think it
01:06:29.380 would just be a
01:06:29.880 bad risk for
01:06:31.340 something that
01:06:31.940 didn't need to
01:06:32.640 be done.
01:06:34.240 So I'm
01:06:35.100 skeptical that
01:06:36.180 there was any
01:06:37.380 intentional
01:06:38.140 malfeasance.
01:06:38.980 I would love
01:06:40.120 to hear the
01:06:40.480 counter-argument
01:06:41.120 from the
01:06:41.520 people who
01:06:41.900 did the
01:06:42.220 study,
01:06:42.820 because they
01:06:43.480 might say
01:06:43.940 something like,
01:06:44.820 oh, we
01:06:45.400 chose this
01:06:45.900 protocol because
01:06:47.200 we talked to
01:06:47.860 X experts and
01:06:48.780 they said this
01:06:49.280 is the one that
01:06:49.840 would work,
01:06:50.400 if anything,
01:06:51.520 something like
01:06:52.040 that.
01:06:52.860 So there's
01:06:53.400 probably a good
01:06:54.200 counter-argument,
01:06:55.000 we just haven't
01:06:55.640 heard it.
01:06:56.660 So don't assume
01:06:57.420 that if you
01:06:57.860 haven't heard the
01:06:58.440 other side,
01:06:58.980 they don't have
01:06:59.440 one, we just
01:07:00.560 haven't heard it.
01:07:03.820 However, I
01:07:04.580 will recommend
01:07:05.240 the ethical
01:07:05.780 skeptic to you
01:07:06.900 as a good
01:07:07.360 follow,
01:07:07.660 because his
01:07:08.960 arguments are
01:07:09.460 always based
01:07:09.840 on data.
01:07:11.000 He always
01:07:11.260 shows his
01:07:11.660 work.
01:07:12.600 I can't tell
01:07:13.120 when he's
01:07:13.420 right or
01:07:13.700 wrong, but
01:07:14.540 it's always
01:07:14.820 provocative.
01:07:15.720 I recommend
01:07:16.160 it.
01:07:20.520 Micro lesson
01:07:21.240 on reaching a
01:07:21.820 higher level
01:07:22.340 of awareness.
01:07:24.160 Well, you
01:07:24.600 don't need a
01:07:25.040 micro lesson.
01:07:26.180 You need
01:07:26.560 micro dosing.
01:07:27.660 No, just
01:07:27.980 kidding.
01:07:28.600 Don't do
01:07:28.960 that.
01:07:30.420 All right.
01:07:32.300 Here's a
01:07:33.060 compliment of
01:07:34.740 the day to
01:07:36.900 professor Scott
01:07:38.160 Galloway, who
01:07:38.820 is also a
01:07:39.400 good follow
01:07:39.880 and author
01:07:42.160 of great
01:07:44.180 books you
01:07:44.540 should read.
01:07:45.780 But he
01:07:46.620 said this,
01:07:47.580 and here's
01:07:48.360 the compliment
01:07:48.960 is going to
01:07:49.440 be in how
01:07:49.980 he handled
01:07:50.440 this exchange.
01:07:52.580 So we'll
01:07:53.220 get to that.
01:07:53.720 But he
01:07:53.900 starts off
01:07:54.340 by saying,
01:07:55.260 whatever you
01:07:55.700 think of
01:07:56.220 Elon,
01:07:58.080 maybe the
01:07:58.820 prospective
01:07:59.360 owner of a
01:08:00.220 social media
01:08:00.760 platform shouldn't
01:08:02.140 be giving
01:08:02.560 medical advice.
01:08:03.760 And he was
01:08:04.140 tweeting Elon Musk's
01:08:06.200 own tweet,
01:08:06.940 a new one,
01:08:08.120 in which Musk
01:08:09.020 retweeted that
01:08:11.080 a friend had a
01:08:12.460 bad experience
01:08:13.060 with Ritalin.
01:08:14.140 Oh, I'm
01:08:14.400 sorry.
01:08:15.340 Galloway retweeted
01:08:16.500 Musk saying that
01:08:18.440 Musk had a
01:08:19.340 friend with a
01:08:19.960 bad experience
01:08:20.520 on Ritalin,
01:08:21.760 and then Musk
01:08:22.540 said, be careful
01:08:23.220 of all
01:08:23.640 neurotransmitter
01:08:24.640 drugs.
01:08:26.100 And so Scott
01:08:26.860 Galloway is
01:08:27.400 saying, you
01:08:28.180 know, maybe
01:08:28.580 that's somebody
01:08:29.880 who should not
01:08:30.420 be giving
01:08:30.780 medical advice.
01:08:31.600 And I
01:08:32.600 said in the
01:08:33.240 comments to
01:08:34.580 Professor
01:08:35.100 Galloway, are
01:08:37.480 you doing that
01:08:38.020 now?
01:08:40.520 Isn't that an
01:08:41.580 example of
01:08:42.200 Professor Galloway
01:08:43.080 giving medical
01:08:43.780 advice?
01:08:44.900 Let me make
01:08:45.680 my case.
01:08:46.900 Suppose Elon
01:08:48.020 Musk had said,
01:08:49.460 you should
01:08:51.300 watch your
01:08:52.120 diet and
01:08:53.660 exercise.
01:08:55.640 Would that
01:08:56.520 be medical
01:08:57.040 advice?
01:08:58.160 Of course,
01:08:58.840 that's medical
01:08:59.340 advice.
01:09:00.120 But would
01:09:00.520 anybody care?
01:09:02.120 No, nobody
01:09:03.640 would care.
01:09:04.760 Why?
01:09:05.240 Would they
01:09:05.560 say, hey,
01:09:06.820 Elon Musk,
01:09:07.680 don't give us
01:09:08.540 any medical
01:09:09.120 advice.
01:09:09.800 No, they
01:09:10.120 wouldn't care
01:09:10.560 because they
01:09:10.980 agree with
01:09:11.740 the medical
01:09:12.180 advice.
01:09:14.020 See where
01:09:14.420 I'm going?
01:09:16.760 So what
01:09:18.020 Elon Musk
01:09:18.560 said is that
01:09:19.560 he had a
01:09:20.000 friend who
01:09:20.460 had a bad
01:09:20.860 experience on
01:09:21.520 Ritalin.
01:09:22.720 That's not
01:09:23.320 like something
01:09:25.480 to disagree
01:09:26.120 with, is it?
01:09:27.520 That's just
01:09:28.240 an anecdote.
01:09:28.720 And it's
01:09:31.080 only being
01:09:31.800 presented as
01:09:32.460 an anecdote.
01:09:32.880 It's not
01:09:33.460 being presented
01:09:34.020 as data.
01:09:35.940 And then he
01:09:36.320 says, be
01:09:36.740 careful of all
01:09:37.480 neurotransmitter
01:09:38.340 drugs.
01:09:39.120 Shouldn't you
01:09:39.620 be careful of
01:09:40.420 all drugs
01:09:41.040 that strong?
01:09:42.460 It wouldn't
01:09:43.080 matter what
01:09:43.500 category it
01:09:44.120 was.
01:09:45.680 Isn't that
01:09:46.300 pretty good
01:09:46.840 advice, to
01:09:47.640 be careful of
01:09:48.400 a neurotransmitter
01:09:49.360 drug?
01:09:50.040 I would be
01:09:50.620 very careful
01:09:51.260 of that.
01:09:52.460 In fact,
01:09:53.340 there are
01:09:53.680 few things
01:09:54.300 more regulated
01:09:55.060 in society
01:09:55.880 than
01:09:56.160 neurotransmitter
01:09:57.080 drugs, try
01:09:58.620 to get
01:09:58.900 Adderall.
01:10:00.200 Not easy.
01:10:01.800 Right?
01:10:02.120 It's very
01:10:02.700 regulated.
01:10:04.000 So he's
01:10:04.920 saying exactly
01:10:05.640 what the
01:10:06.220 medical community
01:10:08.820 would say.
01:10:09.300 Be careful
01:10:09.840 of all
01:10:10.200 neurotransmitter
01:10:10.940 drugs.
01:10:11.540 Be darn
01:10:12.320 sure you
01:10:12.780 know why
01:10:13.280 you're going
01:10:13.720 to use it.
01:10:15.200 That seems
01:10:15.580 like good
01:10:15.960 medical advice.
01:10:17.720 And the
01:10:17.960 one bad
01:10:18.340 experience on
01:10:18.920 Ritalin is
01:10:19.400 just telling
01:10:19.800 you that all
01:10:20.640 drugs have a
01:10:21.420 potential negative
01:10:22.500 side.
01:10:23.580 Who disagrees
01:10:24.560 with that?
01:10:25.620 Is there
01:10:26.000 any doctor?
01:10:27.080 Who would
01:10:27.300 say, no,
01:10:27.900 all drugs
01:10:28.320 are good
01:10:28.680 and they
01:10:28.940 don't have
01:10:29.240 side effects?
01:10:30.680 Nobody.
01:10:32.120 Nobody would
01:10:32.780 say that.
01:10:33.880 So here's
01:10:38.840 the compliment
01:10:39.440 to Scott
01:10:40.260 Galloway.
01:10:41.040 So giving
01:10:41.520 you the
01:10:42.200 setup here
01:10:44.120 again.
01:10:44.820 He was
01:10:45.120 saying that
01:10:45.420 Elon Musk
01:10:45.880 probably shouldn't
01:10:46.520 give you
01:10:46.840 medical advice,
01:10:48.160 but I was
01:10:49.160 noting that
01:10:49.620 he was
01:10:49.900 doing that
01:10:50.300 now.
01:10:51.740 That Professor
01:10:52.300 Galloway,
01:10:52.900 by saying that
01:10:53.520 Elon Musk
01:10:54.120 shouldn't be
01:10:54.780 giving this
01:10:55.340 advice,
01:10:55.820 isn't it
01:10:56.860 telling you
01:10:57.220 that advice
01:10:57.760 is wrong?
01:11:00.460 So here's
01:11:01.220 what Galloway
01:11:02.140 replied to
01:11:04.200 me saying,
01:11:04.880 are you
01:11:05.140 doing that
01:11:05.580 now?
01:11:05.960 He said,
01:11:06.880 maybe,
01:11:07.880 comma,
01:11:08.220 comma,
01:11:08.520 comma,
01:11:09.320 yes,
01:11:10.580 dot,
01:11:10.940 dot,
01:11:11.120 dot,
01:11:11.860 don't know.
01:11:14.260 And then he
01:11:14.960 closes with,
01:11:15.620 anyway,
01:11:15.920 big fan,
01:11:16.860 so you must
01:11:17.420 read Dilbert.
01:11:17.960 if you
01:11:23.060 watch Twitter,
01:11:24.200 you're kind
01:11:24.720 of,
01:11:24.900 you're kind
01:11:25.520 of accustomed
01:11:26.020 to people
01:11:26.520 just defending
01:11:27.280 whatever the
01:11:27.760 hell they
01:11:28.060 said,
01:11:29.060 no matter
01:11:29.500 how ridiculous.
01:11:31.580 And when I
01:11:33.140 pointed out
01:11:33.660 that criticizing
01:11:35.580 somebody's medical
01:11:36.740 advice is
01:11:37.420 medical advice,
01:11:39.000 instead of
01:11:39.620 arguing the
01:11:40.260 point,
01:11:41.640 Galloway just
01:11:42.280 said,
01:11:42.460 well,
01:11:42.600 maybe,
01:11:43.460 yes,
01:11:44.140 well,
01:11:44.340 I don't
01:11:44.460 know,
01:11:44.700 which was
01:11:45.900 actually,
01:11:46.780 I don't
01:11:47.220 know if
01:11:47.440 you could
01:11:47.720 have written
01:11:48.040 a better
01:11:48.380 response.
01:11:52.100 But I
01:11:53.760 think he
01:11:54.080 got to
01:11:54.400 make his
01:11:54.820 case while
01:11:56.060 also being
01:11:56.820 as human
01:11:57.740 as you
01:11:58.100 could be
01:11:58.380 in this
01:11:58.660 situation,
01:11:59.320 so I'll
01:11:59.840 give him
01:12:00.460 credit for
01:12:00.860 that.
01:12:02.300 And I'm
01:12:02.820 not sure he
01:12:04.120 was giving
01:12:04.460 medical
01:12:04.800 advice.
01:12:05.800 I just
01:12:06.120 think it
01:12:06.360 was an
01:12:06.620 interesting
01:12:06.920 question.
01:12:07.460 It's like,
01:12:08.900 isn't the
01:12:09.520 rejecting
01:12:11.040 medical advice,
01:12:11.980 that is
01:12:12.280 medical advice,
01:12:12.980 isn't it?
01:12:14.120 That's all
01:12:14.640 I was
01:12:14.800 adding to
01:12:15.200 that.
01:12:16.180 All right,
01:12:16.420 that,
01:12:16.720 ladies and
01:12:16.980 gentlemen,
01:12:17.400 concludes the
01:12:18.600 best live
01:12:19.900 stream you've
01:12:20.460 ever seen
01:12:20.820 in your
01:12:21.020 life.
01:12:21.840 I think it
01:12:22.400 was full
01:12:22.720 of twists
01:12:23.120 and turns
01:12:23.560 and unexpected
01:12:24.380 things,
01:12:25.560 reframes that
01:12:26.820 you didn't
01:12:27.200 see coming.
01:12:28.280 I believe
01:12:28.640 some of
01:12:29.040 you,
01:12:30.260 20% of
01:12:31.000 the women,
01:12:31.660 25% of
01:12:32.440 the women
01:12:32.700 watching here,
01:12:33.780 just had a
01:12:34.440 great upgrade
01:12:35.140 to your
01:12:35.640 experience,
01:12:37.580 and I
01:12:39.260 believe that
01:12:41.000 you're all
01:12:41.760 a little
01:12:42.000 better off
01:12:42.520 for this
01:12:43.320 live stream.
01:12:43.720 And I'm
01:12:45.400 going to
01:12:45.580 turn off
01:12:45.980 YouTube
01:12:46.600 because I've
01:12:49.020 got a few
01:12:49.320 more minutes
01:12:49.760 with the
01:12:50.300 locals crowd
01:12:51.260 subscription
01:12:51.920 service.
01:12:52.920 You should
01:12:53.200 check it out.
01:12:54.300 Bye for now.