Episode 1729 Scott Adams: Stories That Involve Elon Musk, Which Means Pretty Much Everything
Episode Stats
Length
1 hour and 7 minutes
Words per Minute
150.41655
Summary
It's always the darkest before the dawn, but it's the thirstiest before the sip, right? Join me for the dopamine hit of the day, as we discuss the latest in the Trump saga, including the latest reports that a grand jury is gathering evidence against the President.
Transcript
00:00:00.000
I think so, yes. I feel the golden age is upon us. It's a little bit disguised. I'll grant you that.
00:00:10.120
But it's always the darkest before the... that's right, the dawn. And it's always the thirstiest
00:00:17.080
before the sip. That's not a saying yet, but it will be. It will be, damn it, if I have anything
00:00:24.500
to say about it. And so, are you prepared? Are you ready for an experience which will connect all
00:00:31.400
people almost as if we are a global mind? Almost as if we form a superintelligence collectively
00:00:39.000
being channeled through me? So between coffee and this is shared experience, let's do something
00:00:48.380
amazing. All right. And all you need is a cup or mug or a glass of tanker shells aside in the canteen
00:00:53.960
junk of flask, a vessel of any kind. Fill it with your favorite liquid. I like coffee.
00:01:00.740
And join me now for the dopamine hit of the day. It's the tingle on the back of your neck. It's
00:01:07.700
the thing that makes you feel alive. That's right. Simultaneous sip. Go.
00:01:17.160
Oh, that's a good container of beverage right there. I hope the container of beverage you
00:01:22.180
just ingested was as good as the container of beverage I just ingested. And if that doesn't
00:01:28.960
get you going, nothing will. Well, the walls are closing in on Trump. His legal woes continue to,
00:01:36.960
oh, actually, no, nothing's happening. So it turns out that the latest rumor, so unconfirmed,
00:01:43.900
but it looks like all the Manhattan charges or the grand jury that was going to look into all the
00:01:51.040
Trump financial dealings, they've been looking and looking. They've been talking to people. They've
00:01:56.120
been investigating. They've demanded and they've received documents. And after months and months of
00:02:02.820
the grand jury stuff, the foreshadowing, not yet confirmed, is that it seems increasingly unlikely
00:02:11.600
there will be any indictments coming out of this.
00:02:14.700
Have we ever seen any President Trump witch hunts before? I feel as if it was nothing but witch
00:02:29.580
hunts. You know, I'm not going to say that Trump was an angel all of his life. And the reason I'm not
00:02:38.100
going to say it is because he told us that directly. He literally said in public, I'm, you know, I'm no
00:02:44.240
angel. But then he would tell you why, you know, he could help you and as President. So if you expected
00:02:50.620
him to be an angel in however you want to define that, you shouldn't have been surprised if he
00:02:59.000
talks about grabbing people by the whatever, because he kind of signaled that as directly as you possibly
00:03:05.040
could. And I always thought that it immunized him. A good way to immunize people is to tell
00:03:10.940
people that you have the flaw that you're worried they're going to blame you of. Because if you say
00:03:15.160
it first, it just takes all the fun out of it. You know, if Trump had said, you know, I say horrible
00:03:20.800
things in private, you should know that. And then you find out he said a horrible thing in private,
00:03:26.880
you're like, hey, you said a horrible thing. Okay, he did tell us that. And it just takes all the energy
00:03:31.600
out of it. So may I admit to you right now, I'd like to confess, I say horrible things in private.
00:03:41.460
Horrible things. Just terrible things. If any of it were presented to you out of context,
00:03:48.520
you would say to yourself, well, that's the worst person I think I've ever heard of in the world.
00:03:53.820
But here's the context. Do you have a friend like this? Now, this won't apply to all of you,
00:03:59.720
but some of you do. Do you have a friend who, if you're just alone, could be somebody you've
00:04:05.520
known a long time, usually it is, that the funniest thing you can do is to say the most inappropriate
00:04:12.020
things. Whatever is the most absolutely uncivilized thing you could say, something you would never say
00:04:21.040
in front of someone else. And the fun is how awful it is. Am I right? So I often think,
00:04:29.720
if somehow, you know, my digital devices are recording every word I say, and, you know,
00:04:35.380
somehow it all came back to me, and you played these bits, they would sound worse than anything
00:04:40.240
you've ever heard in your life. Like, you think you've heard people say bad things on hidden audio?
00:04:47.260
You should see mine. Well, I don't know, you should hear them, if such things exist. But yeah, I'll make
00:04:56.340
your head explode. But I wouldn't say it in public, right? The whole reason it's funny is because you
00:05:02.480
wouldn't say it in public. That's the entire energy of it, it's inappropriate. So it's tough to see stuff
00:05:09.600
out of context, is what I'm saying. So here's another big, gigantic story that has been haunting us
00:05:15.760
forever, this whole Manhattan possible indictments of Trump for financial chicanery or whatever,
00:05:22.980
that apparently none of it happened. None of it happened. What would happen if everybody saw
00:05:28.680
Trump's tax returns, and they were just clean? That would be the funniest thing, wouldn't it?
00:05:37.720
After all this time, like, let's just say the entire tax returns became public,
00:05:42.580
and everybody was like, oh, this is going to be good. This is going to be good. And everybody's
00:05:48.060
like salivating over it, and they're like, huh. Okay, there's nothing there. Because apparently
00:05:55.040
that's what happened with these Manhattan indictments. There was nothing there. It's what
00:05:59.200
happened with Russia collusion. You know, Russia collusion, not there. The closest they could find
00:06:06.600
is try to confuse us after the fact that Russian interference in the election was the same as or
00:06:14.280
somehow adjacent to Russia collusion with somebody running for president. Very different. Very
00:06:20.620
different. But that was the closest thing you'd get to making that stick, is talking about an
00:06:25.620
unrelated topic. That's the closest thing you get. So if Trump were to run for re-election,
00:06:35.700
he would be the most vetted person of all time. I don't think I would ever worry again that he
00:06:43.980
would be caught in some illegality or blackmailed. He might be the least blackmailable president now
00:06:51.260
of all time. Am I right? I mean, then you add to that his age, right? At some age, you stop worrying
00:07:00.640
about getting blackmailed. I was thinking about this the other day. At my current age, I was
00:07:08.060
saying, what if somebody blackmailed me? Like they really had the goods, whatever that was. And they
00:07:13.680
said, okay, you're going to be so terribly embarrassed. Your career will be destroyed if I release this
00:07:19.060
information. Well, if I were 30, that would be pretty scary, wouldn't it? Because you're like, oh, my
00:07:26.080
whole life's ahead of me. They're going to let this stuff out. If it happened now, I feel like I
00:07:32.840
think it was funny. I'm not positive. And I suppose that would depend on what it is, I guess. But I feel
00:07:40.500
like I would just laugh. Because it's hard. Nobody's going to take my money away from me. Or at least
00:07:47.680
nobody but the government, I guess. So I'm not sure what I would have to lose. It would just make me seem
00:07:53.200
more interesting, even in a bad way. And I'd say, well, okay, I'll take that trade off. If it makes
00:07:57.840
me seem more interesting. And by the way, if you ever hear bad things about me, I encourage you to
00:08:04.720
believe all of them, except the illegal ones. If you're any, if you're I did anything illegal,
00:08:09.540
totally did not do that. Because I actually, I do try pretty hard to avoid illegal stuff.
00:08:14.800
But if you hear anything that's just like wildly provocative, I would encourage you to believe
00:08:23.100
it, even though there's a very unlike, it's very unlikely it's true. But if it's fun, you should
00:08:29.160
believe it. If you enjoy it. So now there's a story about, I guess, Hannity was exchanging a whole bunch
00:08:37.540
of messages, over 80 messages with Mark Meadows about the January 6th situation. And the big scandal is
00:08:43.620
that Hannity was giving advice to the administration and the president through Mark Meadows. And
00:08:52.520
I'm watching this, I'm thinking, remind me why this is a story? What is the part that's news?
00:09:01.160
Is the news that Hannity and Trump were friends? Because they both talked about that publicly all
00:09:08.960
the time. Everybody knew that. Was, did anybody think that Trump doesn't listen to people who are
00:09:16.740
exactly the right person to give you exactly the right kind of advice? Who would you want advice
00:09:23.740
from if you were a Republican president in a tight spot? Who could give you the best possible advice?
00:09:33.140
Well, you know, Hannity would be near the top, I would think, right? You don't have to agree with
00:09:40.740
Hannity's opinions on anything to, for me to make this point. I'm just saying that Hannity's talent
00:09:47.160
stack, as I've pointed out before, uh, Hannity's just a perfect example of a talent stack. Somebody who,
00:09:53.920
if you looked at any individual thing he has a talent at, you know, speaking in public, knowing about
00:09:59.740
politics, whatever, you'd say, oh, that's, that's good. It's like, and sometimes really, really good.
00:10:05.720
But there's not one of those things that stands out as the best anybody's been at that thing, right?
00:10:11.400
He's got a look, he talks right, he's got the energy. He just has everything. So his magic is there are no
00:10:18.640
gaps. He just has everything. So it makes him, you know, very effective. So if you were going to give
00:10:24.020
somebody's advice in this exact topic, which is how do you handle the public opinion of something,
00:10:31.300
I would go to somebody who is one of the best people on the planet in managing public opinions.
00:10:37.660
Hannity is exactly whose advice I'd want to hear. And how, and what kind of advice did he give?
00:10:42.600
One of them is that he said that Trump should announce he will lead, he said after the six,
00:10:53.740
this is what Hannity said, that Trump should announce he will lead a nationwide effort to
00:10:57.740
reform voting integrity. Go to Florida and watch Joe mess up daily. Stay engaged. When he speaks,
00:11:05.740
people will listen. And I thought to myself, okay, that's really good advice, isn't it?
00:11:12.600
That's about as good as, if you were going to get advice, that's about as good as you could get.
00:11:20.640
Now, Trump didn't take this advice, right? Trump decided to be Trump, and maybe there's nothing
00:11:28.600
wrong with that because he's made it work so far. And so Trump decided to be, you know, fully
00:11:33.300
combative. But if I read between the lines, I think Hannity's approach was to basically shift
00:11:43.160
the argument and become the champion of election integrity, which nobody could disagree with.
00:11:49.000
Basically, it's a high ground maneuver. Have I told you that the high ground maneuver wins every
00:11:54.580
argument? It's the one that always wins. And as soon as you hear it, you're like, oh, okay,
00:11:59.740
damn it, the argument's over. That's the high ground. The high ground is not whether the election
00:12:05.300
was rigged or not rigged. That's the low ground. The high ground is what Hannity showed him. The
00:12:11.200
high ground is I'm going to lead a national effort to make sure that the next time this happens,
00:12:16.380
we're all comfortable with the outcome. National hero, right? Trump could have easily transformed
00:12:23.740
transformed this from maybe the biggest stain on his presidency. Not maybe. The biggest stain on his
00:12:30.740
presidency. He could have easily done taking Hannity's advice and turned it into, all right,
00:12:38.000
I guess we'll never know what happened in 2020. I have my suspicions. And people would say, okay,
00:12:43.860
that's fair. You have your suspicions. And it's fair that we'll never know. Yeah, okay. Not everything
00:12:50.420
was audited. Can't get into the, you know, the technology part of it especially. He would have
00:12:56.720
been a national hero. And probably when the next election rolled around, unless people thought he
00:13:03.420
rigged the election, I suppose they'd spin it that way, people would say, all right, let's run this
00:13:10.460
movie again. And we'll see if the election reform actually changes the outcome. Let's see who gets 81
00:13:16.500
million votes this time with election reform. Now, even if he lost, it would still be legendary
00:13:23.160
because people wanted election reform as like, you know, the basic, most fundamental thing to protect
00:13:30.100
the republic. So the fact that there's a story that Hannity was giving advice to Mark Meadows to give to
00:13:39.200
Trump, the story should have been, why wasn't Trump listening to it? That would have been the better
00:13:45.980
story. Because this is damn good advice. In my opinion, it's damn good advice.
00:13:53.720
Maria Bartiromo is getting some similar kind of pushback because apparently she shared some of the
00:13:59.200
questions that she was going to ask the president after January 6th with Mark Meadows, I guess. And
00:14:07.620
here's the first thing you should know about that. That's not unusual. It's not unusual for an
00:14:14.200
interview guest to get questions in advance. Because it's more about the topic. And it's more about
00:14:21.420
preparing somebody to have a good show. I can tell you that in many cases when I'm interviewed on
00:14:27.400
politics, I get the questions in advance. And, you know, there's nothing unusual about that.
00:14:35.940
The reason you do it is to make the show snappy. What you don't want is a show where
00:14:40.860
somebody asks a question, and then the guest says, uh, you know, I hadn't really thought about that.
00:14:48.180
You don't want that. So you want to say, I'm going to ask you some tough questions,
00:14:52.080
or not, but you tell them what they're going to be. And then the person has thought about it,
00:14:56.960
and they give a good, quick response, as short as possible. It's good for the audience. It's good
00:15:02.140
for the show. But it's also good for the interviewee. Everybody looks good.
00:15:05.440
Now, here's a question. In this case, is Maria Bartiroma an opinion person like Hannity,
00:15:16.260
where I think Hannity is perfectly transparent, that he's an opinion person, he's friends with
00:15:22.320
the president, they talk a lot. Perfectly transparent. But do you see Maria Bartiroma as opinion or news?
00:15:30.620
I'm just going to see what your opinion of her is. Some say news, some say opinion. Okay, that's the problem.
00:15:40.260
Both. Yeah, see, that's the problem. Because now this one gets a little more murky. But apparently,
00:15:48.100
apparently she didn't use exactly the questions that she broadcast. Because I don't know if you're aware of this,
00:15:58.580
well, even when the person asking the question has a set of questions that are on their notes,
00:16:05.220
that's mostly just so they don't forget a question or, you know, don't have any questions left.
00:16:12.700
But they kind of ask what they think is a good question when they actually get there.
00:16:16.260
So somebody like Maria Bartiroma isn't going to ask the exact question. It's just an indication
00:16:23.260
she's going to be in that area, basically. And that's what happened. She asked, you know,
00:16:28.380
some versions of the questions. So I don't think there's anything wrong with that, necessarily.
00:16:35.760
And I wouldn't be bothered. And I'm being consistent here.
00:16:40.360
Because when Chris Cuomo was accused of softball treatment of his brother, the governor,
00:16:51.340
if you recall, I also defended Chris Cuomo. Because it's transparent. As long as it's transparent,
00:16:59.600
I don't know if there's a higher standard. If you know it's his brother,
00:17:04.060
are you going to be surprised if a brother gave a brother advice in any context? That should have been
00:17:12.980
the least surprising news and shouldn't have affected anybody, really. So I just want to be
00:17:19.700
consistent. People should be able to talk to anybody they want and get news, get advice from anybody they
00:17:24.780
want. As soon as you make, oh, this one can't talk to this one unless you've told us. No. No,
00:17:30.880
anybody can talk to anybody about anything. That's by standard. And they don't necessarily have to
00:17:37.540
tell you. You can talk to anybody about anything. And they don't necessarily have to disclose it.
00:17:43.620
But it's nice when they do. All right. Have you noticed that every story is about Elon Musk?
00:17:51.800
We'll give you some examples. So AOC tweeted this sort of long, ambiguous tweet to which Musk
00:18:00.460
responded. So AOC tweets, tired of having to collectively stress out about what explosion of
00:18:07.180
hate crimes is happening because some billionaire with an ego problem unilaterally controls a massive
00:18:13.260
communication platform and skews it because Tucker Carlson or Peter Thiel took him to dinner and made
00:18:19.680
him feel special. Now, when I read that, I thought she was talking about Elon Musk buying Twitter.
00:18:27.220
Elon Musk must have thought the same because he tweeted back hilariously, quote,
00:18:38.720
OK. Now, if you see this outside of the realm of Twitter, which a lot of people will, they'll
00:18:47.600
just see this, say, reported in a news item or something. You don't really appreciate how perfectly
00:18:54.000
Twitter like his response is. Right. His response would be maybe inappropriate in almost any other
00:19:02.580
domain. In any other domain. It wouldn't be a little weird. But in this specific one of
00:19:10.020
Twitter, it's exactly right on point. He's hitting the target right on the head. Poink.
00:19:16.940
It's a Twitter response. So I've told you before and keep watching for this because it's fun to watch
00:19:24.000
that Elon Musk is very clear about what matters and what doesn't. And when things don't matter,
00:19:29.920
he mocks them. And when things do matter, like saving the planet or going to Mars or something,
00:19:35.720
he somehow makes that happen. So I've never seen anybody who's more clear about what's silly
00:19:40.700
and what's not. That's just one of his best qualities. And so, you know, he just makes fun of it.
00:19:47.540
And then apparently AOC tweeted, but quickly deleted. I was talking about Zuckerberg, but OK.
00:19:56.600
And then everybody had to debate whether she deleted it because it wasn't funny enough or didn't want to
00:20:04.180
engage or was it because it really wasn't about Zuckerberg or who knows. But apparently there's a
00:20:13.620
Zuckerberg version of meeting with at least Peter Thiel and there's a speculated Musk version in which
00:20:20.940
he probably met with Peter Thiel or did or something as part of, you know, deciding about Twitter.
00:20:27.160
And so I just love this little exchange. But so Elon Musk is in every part of the news. We'll keep
00:20:35.780
going on this. But first, so this really happened. A reporter for The Guardian decided to do a story
00:20:47.340
about virtual reality and so went into one of the virtual reality worlds. And I'm not sure which one
00:20:53.180
it was. I don't know if it was meta or just a virtual reality world. It doesn't matter to the
00:21:00.180
story. What matters to the story is that she was immediately assaulted with racism and actually was
00:21:09.400
groped in the VR environment. Basically sexually assaulted in virtual reality. Now, of course,
00:21:15.980
she was quick to point out that she never lost touch with the fact that it wasn't the real world.
00:21:20.820
But the problem with or the or the feature of virtual reality is it makes you feel the same
00:21:27.300
way as the real world. Knowing it's not the real world doesn't help you nearly as much as it should.
00:21:33.900
I've told you some of the story about my VR experience. I put on the glasses and I walked up to a cliff
00:21:40.680
edge so that in the virtual world, if I if I stepped off, it looks like I would fall to my death.
00:21:46.620
In reality, I knew I was just in a room in my house and had no danger whatsoever. I couldn't make
00:21:52.780
my legs move. I couldn't walk over the cliff in the virtual reality. Couldn't make my legs move.
00:22:01.640
My brain would say, move your legs. You're perfectly safe. I would even take the glasses off to make sure
00:22:07.900
I was still in the real room. Put them back in and say, all right, no, couldn't do it.
00:22:13.740
And so when she says she was actually assaulted, and like, I guess they cornered her and they were
00:22:20.240
doing stuff with their hands and stuff, that she felt actually assaulted. And I think that's real.
00:22:27.820
That's completely real. And so what are you going to do about that? Do you end up having all the same
00:22:35.700
laws in the virtual reality? Because the virtual reality just becomes your reality? Well, just to make
00:22:41.920
it more weird, there's a new invention that allows you to feel things while you're in virtual reality,
00:22:51.160
specifically on your mouth. And so Gizmodo had an article about this that I'm failing to find in my
00:23:00.100
notes. But apparently there's a little haptic response thing. And I saw a picture of it. You've got
00:23:08.300
the VR goggles on. And then there's some kind of sensors, or I don't know if they blow air or what
00:23:13.780
they do, hanging from the bottom of the goggles. And so they're directed at your lips and your mouth.
00:23:21.160
And the claim is that these little devices that are not touching your mouth, but I think they
00:23:27.920
might direct air or something at your mouth, they'll make you feel as if you're actually kissing
00:23:32.620
somebody, if you're kissing somebody in the VR world. Now, it did go on to say that you might be
00:23:41.260
able to feel it even internal to your mouth, such as if you had your mouth open. I'd imagine you'd feel
00:23:48.560
something on your tongue or the inside of your mouth, because that's where the haptic sensors would
00:23:54.040
be sensing. You kind of know where this is going, don't you? All you need to do is put those haptic
00:24:03.520
sensors in your belt, you know, one on your goggles and one on your belt, both shooting down, if you know
00:24:10.580
what I mean, if you know what I mean, if you could feel it, just like it's in the real world.
00:24:17.440
We're in big trouble. Big, big trouble. Big trouble. So much so that I tweeted, and people didn't
00:24:26.260
understand, that sometimes when you think you know somebody who's socially awkward, and they don't
00:24:32.740
have much of a social life, and they never seem to go out. And that's how that's your opinion of them.
00:24:38.260
So, oh, this is somebody with a bad social life. They don't like to go out. But I would suggest
00:24:43.400
to you there's one other possibility, that that is somebody who's really, really good at
00:24:48.580
masturbating. Like, so good, they can do it for hours, and it never gets old. To them, going out
00:24:56.720
might be the least fun thing they could possibly do. Imagine if you were bad at masturbating,
00:25:03.040
and somebody said, hey, there's a party. Or you can do this thing that's tons of fun. It's
00:25:09.180
going to last you two minutes. Two minutes of great fun. Or you can go to this party. Well,
00:25:14.780
the party sounds pretty good, doesn't it? Relative to two minutes of a good time, you'd have fun
00:25:21.660
all night. But suppose somebody was really good at it. They could keep themselves at a place for
00:25:30.580
hours at a time. Does the party look as good to them? What happens when virtual reality makes the
00:25:41.440
staying home alone just way better than going to a party? And I think probably you're already at the
00:25:49.040
point where for some people going out is the least fun thing they could possibly do. And where is that
00:25:55.980
going to take us? Because we're already there. I mean, we're knocking on that door.
00:26:02.380
I blocked Kathy Griffin today for being a racist, because she called Elon Musk a white supremacist.
00:26:08.520
And I thought, okay. You know, I'm certainly willing to put up with anything that she thinks is a joke.
00:26:18.600
Like, you know, I supported her with the severed head of Trump under, you know, under the rules of
00:26:25.800
parody and free speech and blah, blah, blah. So I didn't think that she should lose anything because
00:26:30.860
of that. I was very strongly supportive of that as a free speech thing. And as, you know, humorists take
00:26:37.580
chances, they don't always work. But you don't want them to not take chances. But calling Elon Musk a
00:26:47.740
white supremacist, I feel like that's just because the people on the right like him. I feel like that's
00:26:54.460
a little too close to home. Because I like him too. And I have no indication of anything like that.
00:26:59.960
That just feels so bigoted, essentially against white guys, basically, that I can't pretend that
00:27:11.060
somebody else is the bigot in this story. It just feels like she's the bigot in this story. Am
00:27:16.020
I right? It just feels like it's just an anti-white male thing.
00:27:21.960
All right. Here are the other stories that Elon Musk is associated with, right? Just listen to the whole list
00:27:37.840
and ask yourself, how is this even possible? Okay, I get that he's, you know, a richest person and bought
00:27:45.960
Twitter. So I get you a lot of attention. But look at all the stories that these topics that he's
00:27:51.640
directly involved in. Ukraine, right? He sent the Starlink stuff over there.
00:28:01.520
The Amber Turd story, because they're talking about, you know, his dating her. So the Johnny Depp
00:28:08.840
story is even that he's attached to. Anything about elections, fake news, Twitter, because they're
00:28:15.160
all sort of collectively one story now. Anything about income inequality, he's in that. Anything
00:28:20.160
about taxes of the rich and progressive taxes, he's in that story. Anything about free speech,
00:28:25.720
because of Twitter. Now he's weighing in on mental health. We'll talk about that. Tweeting about it.
00:28:30.240
So he's in a story about Adderall. He's in any story about space, climate change, how the country
00:28:37.040
is getting more divided, also because of the Twitter thing. Self-driving cars.
00:28:40.920
I mean, and he believes that the simulation is real. It's my theory that people who believe the
00:28:53.340
simulation is real can author it. And the more sure you are that the simulation is actually just
00:28:59.940
software, the more clearly you can see the machinery, and the more clearly you can tweak it.
00:29:05.580
It sure looks to me like he knows he lives in the simulation, and he's just playing it like a
00:29:12.020
game. It looks like he's playing it like somebody who's just a good gamer. You know, if you found
00:29:19.700
out tomorrow that this is all a game, and that we're just characters in it who temporarily don't know
00:29:24.320
what we are, and we think we're real, and you woke up and found out that Elon Musk was only,
00:29:29.760
only, the best gamer in another dimension. And he was just an avatar. But he was an avatar of the
00:29:37.800
best gamer for the game. So he became the richest person, and he became in every story. He had seven
00:29:44.880
kids, and God knows what kind of fun he has when he's alone. Does it look exactly like just a really
00:29:52.900
good gamer playing the game really well? It's weird how much it looks like that. You know, I've looked at
00:29:58.880
my own life, because I believe I live in a simulation as well. Like, literally. Like, actually, literally,
00:30:05.200
no joke. It's the most likely possibility. I can't say for sure about that or anything else. But I treat
00:30:13.760
it like it's not real. And my experiences just don't, they just don't seem like they could possibly
00:30:20.980
be coincidental. It just doesn't seem like it. I don't know what to think of that. But those people who
00:30:28.540
have said to me, this is just anecdotal, of course, that as soon as they feel they're in the simulation,
00:30:35.300
and they start using things like affirmations to author the simulation, they report that they get
00:30:41.140
good results. But of course, that's anecdotal. So Elon Musk weighed in on this thread. So somebody
00:30:51.480
named Michael Kersey tweeted this. This is just the second part of his tweet. He said,
00:30:57.920
pharmacological dark matter, and he's talking about a phenomenon among important people.
00:31:07.640
So he's just giving it a name. Pharmacological dark matter. So basically, the stuff we don't know
00:31:12.260
is invisible heavy amphetamine and other drug use among people playing significant roles in our society.
00:31:18.600
Now, you've heard me say that, right? The people who are doing the most moving and shaping of
00:31:27.260
civilization, mostly on drugs. It wouldn't happen otherwise. And, you know, don't take drugs.
00:31:38.820
Like, and I mean it, don't take drugs. I mean, and, you know, the only time you should is if you're
00:31:45.720
under a doctor's care, and even then it's probably too much. So I'm not encouraging it. I'm just saying
00:31:51.520
it's a fact, and ignoring it feels stupid. It just feels stupid to ignore it. The fact is
00:31:58.720
that some people, and here's the dangerous part, everybody responds to chemistry differently.
00:32:05.560
So there might be a drug that simply makes one person rich, because it just makes them perform
00:32:10.840
better, and they never have a side effect, or one they care about. And then another person,
00:32:15.920
it just kills them. It just freaking kills them, or it makes them crazy, or it ruins their life one way
00:32:21.660
or the other. So don't take drugs, because you don't know which one you are. You don't know if you're
00:32:27.360
the one the drug is going to kill, or the drug is going to help you. We're just not that smart,
00:32:32.640
because, you know, we're all different. So anyway, so Michael Kersey weighs in on this about the
00:32:40.540
significant role, and then that Mark Andreessen, one of the most important voices in the tech world,
00:32:52.300
tweeted this. He said, everyone thinks our present society was caused by social media. I'm wondering
00:33:00.240
whether Adderall plus ubiquitous Google searches have bigger effects. Now, I don't know about the
00:33:05.380
Google search part, but here's one of the most connected people in Silicon Valley, and then the
00:33:12.960
tech world, who would personally know the most important people in society. So, you know, this is
00:33:19.280
somebody who's been in the room and has the phone number to text pretty much anybody famous, I think,
00:33:25.720
at this point. And he's telling you that he thinks Adderall may be shaping civilization or things like
00:33:35.020
it. I'm not saying Adderall specifically, but you should take that as things like it, you know,
00:33:40.340
amphetamines. And then Alex Cohen tweeted this. He said, prescribed psychedelics will replace
00:33:47.640
amphetamines and SSRIs over the next decade, although I hope it's sooner than that, he said.
00:33:54.120
And then Musk gets into this, and he tweets, I've talked to many people who were helped by psychedelics
00:34:03.740
and ketamine, more people who were helped by psychedelics and ketamine than SSRIs and amphetamines.
00:34:11.580
And then he added, related to this, he said, and this is not me talking, so I'm just reporting what
00:34:18.720
he said, right? So these are not my words. He said, Adderall is an anger amplifier, avoided
00:34:24.280
all costs. Now, here's the interesting thing. As Jeff Pilkington pointed out in some tweets,
00:34:33.740
everybody's different. That's what I said earlier, right? I'm pretty sure that Adderall has saved
00:34:38.820
lives, but I'm pretty sure it's caused some problems. I think both of those can be true.
00:34:46.720
So it's a little bit, let's say it's a really good example of free speech, both its negatives
00:34:57.300
and its positives. You don't want to be getting medical advice from Elon Musk. That might be
00:35:04.120
his weakest category. I mean, he's insanely smart on a whole variety of things that allow him
00:35:13.220
to do what he does. But I think the medical part might be the part where you go, yeah,
00:35:17.980
you get a second opinion there. Usually when Musk says something, I'm usually done. It's like,
00:35:24.140
okay, I agree with that. There's nothing else to say. But I think in the medical domain,
00:35:31.240
let's be glad that there is free speech, because he can say this. If you looked at the comments,
00:35:36.080
there would be a whole bunch of people pushing back. And I say, okay, that's a really good example
00:35:41.880
of free speech working. Somebody's asking if I'm on Adderall. I'm not. So I've never been on any kind
00:35:49.740
of long-term stimulant, except coffee. Or sativa, I guess. But I will, I've said this before. But in college,
00:36:00.740
I had a few experiences with stimulants. And I wrote my entire senior thesis in like, I think it was like
00:36:11.300
mostly over a weekend. And then I was done with the semester. And it was actually easy. And I enjoyed it.
00:36:21.460
Think about that. I did a semester of work in four days. Got a reasonably good grade on it. I think B plus
00:36:28.400
or something. And I did it in four days. And I liked it. It was completely pleasant.
00:36:35.160
It was a horrible job. Like, just the most boring. It was like in economics. So it was a senior thesis
00:36:44.300
in economics. Do you know how boring that was? And I enjoyed the whole thing. Now, how many unicorn
00:36:52.800
companies have been created by people who had a little stimulant going on? How many? Probably a lot.
00:37:03.220
Right? Probably a lot. And it is one of the great untold stories. Anyway.
00:37:15.980
Michael Schellenberger, who is running for governor as an independent in California. I understand there
00:37:22.040
was some issue in terms of the paperwork. I've got to look into that a little bit. But how would you
00:37:30.440
like to be running for governor and part of your accomplishments just happened, which is Schellenberger
00:37:38.360
was, I think, the loudest, most effective voice for trying to save the Diablo Canyon nuclear power plant
00:37:47.320
in California? And the reasoning was it's already there. It's cost effective. And we don't have any green way
00:37:55.200
to replace it. And we'll just run out of energy if we don't keep it open. A pretty straightforward
00:38:02.080
argument. And apparently, he, the great weight of public opinion moved Schellenberger's way.
00:38:13.120
Is that a coincidence? Or did he cause it? Because remember, he was testifying to Congress. He was writing
00:38:19.200
books and articles on it. He was the most famous voice on this topic. I think he caused it. I think
00:38:26.740
he caused it. So he's running for governor, while the current governor is saying that he's going to
00:38:32.860
look to get some money from the federal funds that was allocated by Biden, and I'll give Biden credit
00:38:38.520
for this, for keeping nuclear power plants open longer than they would have been.
00:38:43.120
So the governor is looking at implementing the plan that the governor didn't want to do,
00:38:51.900
but Schellenberger convinced the entire country that they need to do it, and now he kind of has
00:38:57.680
to do it. I would hate to be running against Schellenberger in this situation. Honestly, I've never seen a more
00:39:07.440
capable politician in terms of, you know, competence about the actual topics that matter to the state.
00:39:16.620
It's sort of breathtaking, because we've never seen it before. I'm pretty sure we've never seen this
00:39:22.080
before. We've had some presidents who were a little wonky, you know, like Jimmy Carter and stuff,
00:39:27.700
but not like this. Like Carter wasn't, didn't have this kind of mastery over the exact topic that the
00:39:35.440
state cared about, and several of them, from homelessness to drug addiction to, you know,
00:39:41.460
he's talked about forest management. Basically, everything we care about, he has the better
00:39:47.320
solution for. All right, here's the most provocative thing that I've read lately.
00:39:52.160
We all assume at this point that Russia will have its way with Ukraine. In the comments,
00:40:00.000
where are you feeling as of today? Like, what do you feel today? Is Russia going to have its way?
00:40:10.560
All right, so I'm seeing mostly yeses. Some people saying they'll just take the east of Ukraine,
00:40:17.140
blah, blah, blah. All right. Now, I will remind you that in war, nothing is predictable.
00:40:24.120
So, among the unpredictable things that could happen, I tweeted this, and I feel terrible because
00:40:31.740
I want to mention the author and the publication, because I liked it so much, but I can't find it
00:40:37.240
in my own tweets for some reason. Maybe I imagined I tweeted it. Yeah, maybe I imagined it.
00:40:43.240
But if somebody sees it, maybe you can tell me in the comments if I did tweet it and you see it.
00:40:50.600
It was an article by somebody who definitely looked like they knew what they're doing. It was
00:40:54.060
somebody who had predicted in writing and could show the receipts that the Moskva, that ship would
00:41:02.020
be sunk by the Ukrainian anti-ship missiles. So, that's a pretty good prediction. Somebody had that
00:41:08.260
specific prediction and it happened. And then when I read the article, it seemed to know quite a bit
00:41:13.460
about the whole situation. So, I don't know how to judge military experts, because we've seen so
00:41:19.960
many of them being so wrong about so many things. But I'll give you his argument, and I'd love to see
00:41:26.480
how wrong it is. It goes like this. We're all focusing on the fact that the fighting isn't this
00:41:32.960
Donbass and the other name starts with L that I can't remember. And we're taking for granted
00:41:40.860
that Russia already owns Crimea, because they got annexed in 2014, something like that. So, yeah,
00:41:49.520
so Luhansk is the other region. So, we're ignoring Crimea, right? Because that's a done deal.
00:41:55.620
Russia already owns it. Here's the part I didn't know about. They barely own it. They do own it.
00:42:04.060
Their military is theirs. But apparently, if you're looking at it from a military perspective,
00:42:11.080
their connection to Crimea is one bridge and one poorly defended area. And if you take out the bridge,
00:42:20.940
which seems totally doable, right? Ukraine can take out a bridge. They've got missiles,
00:42:27.620
drones and whatnot. And then, apparently, the Ukraine military is actually pretty good.
00:42:33.820
They have better equipment. According to this expert, they win their firefights.
00:42:39.480
So, when it's something like a fair fight, the Ukrainians almost always win, because they're
00:42:45.860
better trained and better equipped. It's only when they're, you know, overpowered by artillery or
00:42:50.260
something that they get crushed. So, the thinking is that the Ukrainian military could
00:42:58.540
take out not only the thin connection between the main Russian forces and Crimea, they could
00:43:06.460
take out the bridge, and then they could just take Crimea, because it would probably be poorly
00:43:13.080
defended, because the strongest defenders are where the fighting is. And the fighting is not
00:43:18.860
happening in Crimea. And I'm thinking, that would be a shocker. What if Ukraine just tried to hold the
00:43:27.140
line, and just keep the main Russian army pinned down, and just take Crimea back? What would that do
00:43:36.440
to the Russian psychology? It would look like Russia lost the war. I mean, the war wouldn't be over.
00:43:43.540
I mean, Russia could maybe just take it back. I mean, it wouldn't be easy. But what would that do to the
00:43:51.080
whole balance of everything? I mean, that would be a great, at least a diversion. I mean, at the very
00:43:57.520
least, it would be an interesting diversion. And the idea is that Russia would have trouble equipping
00:44:02.700
and resupply in Crimea, because they could get cut off, but that Ukraine wouldn't, you know, because
00:44:09.280
they have a border there and stuff. They'd break out the tactical nuke if that happens, somebody says.
00:44:17.240
Would they? Because it seems to me that a tactical nuke would guarantee that Putin is taken out of office.
00:44:25.520
What do you think? I would say that right now, it looks like he might be, you know, at great risk,
00:44:33.420
maybe 50-50 proposition. But if he used a tactical nuke, I think that's the end. Am I wrong? There's no way
00:44:43.280
he could survive it, politically, if not his life. Because don't you think that there's somebody who's
00:44:51.440
like number two or three in the potential takeover chain that the CIA has already said, you know,
00:45:00.380
I know you're only number fourth or fifth in the potential chain of command. But if the other people
00:45:05.900
disappeared, and we made sure that you were well treated, do you think you'd like to step up to the
00:45:12.580
number one job? I don't know. I think that if a tactical nuke got used, it would be easy to turn people
00:45:19.200
within Russia really close to Putin against him. Because they would say, okay, this is a clean
00:45:24.940
win. If I take Putin out under these conditions, I'll even have internal support. Because I think
00:45:32.120
you could overthrow Putin and still have public support in Russia if he used a nuke. Or the other
00:45:40.280
possibility is that Russia is now so pro-Putin that they'd say, oh, that's strong. You know, we like it
00:45:48.140
that you were strong. Maybe. I don't know how to read the Russian mind. But it would certainly
00:45:54.820
turn enough people against Putin that he'd be in physical danger much more than he is now.
00:46:01.440
Now, here, let me throw something into the mix that feels different. So you know the Biden
00:46:08.800
administration is asking for, correct me if I'm wrong, $33 billion more for Ukraine, which
00:46:14.880
is a gigantic step up in military support. Do you know what that sounds like to me?
00:46:24.000
It sounds like the administration is playing to win, as in win the war. It doesn't look like
00:46:32.140
they're playing for a stalemate. It doesn't even look like they're playing to just, you know,
00:46:37.560
push them out of the East or something. I feel as if the Biden administration feels like
00:46:42.880
they could win. Because every day that the Russians don't, you know, have some crushing
00:46:52.740
victory on the East is a day that you say to yourself, huh, maybe that Ukrainian defense
00:46:59.380
is better than we thought. And they're not going to, the Ukrainian defense isn't going
00:47:03.140
to quit now because the good equipment is just coming in. And apparently they don't have
00:47:07.640
any manpower shortage. They had, they had more volunteers to fight than they had equipment
00:47:12.700
to give them. So if Ukraine doesn't run out of people to fight and they're getting more
00:47:20.220
equipment, not less, does the $33 billion signal that the Biden administration thinks they could
00:47:27.660
win outright? Because I feel like it does. Now, I don't think that means that they think it's
00:47:33.780
a sure thing or anything like that. But I feel like they went from, you know, let's see
00:47:38.860
how we can make this painful for Putin to let's finish off the Russian army. I think they want
00:47:45.600
to finish the Russian army. Or at least, you know, take it down by 50%. Because Putin's
00:47:52.560
down 25%, right? I never understood that if that's his entire military 25% degraded or that
00:47:59.700
was just what's in that area. I've not, I've not heard that clarified. So about 20 billion to
00:48:11.400
replace the weapons we gave them from the army. Okay. Oh, I see. So we'd be replacing our own
00:48:18.520
weapons with the $33 billion. Okay, that does look a little different. Those numbers are wrong.
00:48:24.980
Which numbers are wrong? The $33 billion? If all it is, is replacing our own equipment,
00:48:30.080
then it doesn't look like a step up, does it? So give me a, give me a fact check on that.
00:48:36.020
That replacing our own military equipment would not look like a step up.
00:48:43.520
Somebody says not true. Are we giving it directly to Ukraine? All right. Well, I guess we'll look
00:48:48.920
into this. But the point is, if it's a major escalation, if that's what that budget is telling
00:48:54.660
us, then I think the Biden administration actually feels they could win. Define win. Put Putin
00:49:02.260
out of power. Or, and or, get the Russian military to completely get out of Ukraine. Completely.
00:49:13.760
Now, of course, nobody wins because everybody's, everybody lost in this situation. But
00:49:19.040
that would look like a win to me. All right. I was asked to tell you about how we're entering
00:49:31.400
the golden age. Well, someone has asked me why I didn't have kids and do I regret it. Let
00:49:42.500
me answer that one first. I don't regret it. But, and the reason was, I just didn't want
00:49:54.580
to bring somebody into the world. Because there are plenty of people in the world. And I never
00:50:00.780
felt a need to propagate my DNA. Now, does that make me weird? Because I feel like there's
00:50:10.420
some kind of basic human impulse that people have to propagate their DNA. But I don't have
00:50:21.260
that. Because I didn't enjoy my childhood enough that I would take any chance that I could put
00:50:29.280
anybody through that again. But if there's somebody who's already having a tough childhood, and I
00:50:36.040
could help them have a better one, that feels like that's an easy win. Like, to me, that
00:50:40.880
would be satisfying. But to bring somebody into the world and then have them have a bad
00:50:45.320
life, I couldn't handle that. Like, my, I'm just not built that I could ever handle that.
00:50:51.120
So I take the sure thing of, you know, definitely helping humanity extend the light of consciousness,
00:51:01.560
I guess. So he says, what a load. There's somebody who thinks that I'm either lying, do you think
00:51:09.660
I'm lying to you, or lying to myself? Could be the one, right? I guess there's no way for
00:51:15.640
you to know, is there. But the thing I thought about the most is that I would rather support
00:51:24.600
the people who are already here. That's how I thought of it. It could be that that's a
00:51:29.120
rationalization of some sort. That would be pretty normal. But your child wouldn't have
00:51:36.700
a tough life. Yeah, no, I don't think, we're not talking about a tough life in terms of economics.
00:51:42.560
If it were just economics, I wouldn't have worried. But in terms of, you know, mental health
00:51:49.300
and that sort of thing. I'm not so worried about climate change ruining the world. I just
00:51:56.940
look around, I don't see a lot of happy people. Now, here's the golden age part. I'm pretty
00:52:02.920
sure that we're about to wake up, or we are waking up now, to this whole what drugs to put
00:52:10.420
in your body and whatnot. And maybe the pandemic helped with that. Because it sort of changed
00:52:16.000
how we see the medical community, you know, changed our opinion about maybe having to make
00:52:21.040
our own decisions and not depend on the experts so much, which could be good or bad. I guess
00:52:26.200
that could go both ways. But yeah, I think a big, the, I would go with Elon Musk's thing
00:52:36.040
that the ketamine, maybe. And of course, don't take any recommendations from me. I'm just saying
00:52:43.840
that these are things that are being talked about. I'm not recommending them at all. And
00:52:47.900
the psychedelics, I do think have the potential to be civilization changing. Absolutely. And
00:52:56.200
because they don't cost much, it's, you know. Let me ask you this. Imagine everything that
00:53:04.080
we do now to make the world a safer place. Right? We've got the United Nations and, you know,
00:53:11.020
we've got the hotline to the Kremlin. So we've got all these systems and things to make things
00:53:17.740
safer and avoid war. But we still have the damn wars. Right? So I guess those systems are better
00:53:24.320
than not having them. But they're not, they're not really getting it done. But imagine this.
00:53:30.300
This is purely speculative. And it's just, it's just a mental experiment. Imagine that if instead
00:53:37.280
of all that stuff, whenever there was a dispute, or even before there was one, the heads of
00:53:43.860
state would get together and do mushrooms together. That's it. And then replace everything else
00:53:51.000
with just that. Now, of course that will never happen. Of course it will never happen. I'm
00:53:57.000
not suggesting that's even remotely possible. I'm just saying if it did, it would probably
00:54:03.040
end war. Do you think, I mean, honestly, if you just imagine, you know, it doesn't work
00:54:13.920
if your leader is 100 years old. So forget about Trump. He won't take a drug. And forget
00:54:19.380
about Biden. He's too old. But imagine DeSantis, who I imagine is anti-drug. So again, this wouldn't
00:54:26.600
happen. But at least he's young enough. So imagine a President DeSantis someday, doing
00:54:32.540
mushrooms with a President Putin who's, you know, 70 or so, and not that old. And they
00:54:40.760
just like bond and see the world differently. And the next thing you know, it just war is
00:54:47.260
a lot harder. Like it doesn't make sense suddenly. Because it would be so much easier to say, you
00:54:53.600
know, wait a minute. Are you saying that if Russia and the United States simply had to
00:54:59.280
just make friends, that the same way that Germany benefited by being friends with the US and Japan
00:55:06.300
benefited, and basically everybody, every country that said, hey, can we be your friend, has benefited
00:55:14.540
substantially. Can you imagine sitting in a room and having the right kind of, you know, right
00:55:20.760
kind of, let's say, chemical incentive? And you just look at Putin and say, why are we
00:55:24.940
even doing this? Do you want to be the greatest leader that Russia ever had? And Putin would
00:55:30.920
be, well, I thought I already was. You say, okay, okay. But do you want to continue being
00:55:35.540
the greatest leader that there ever was? And Putin would say, how do you do that? He'd say,
00:55:41.440
end war. End war and go to space with us. How about that? How about end war and help
00:55:49.760
us go to space? And we'll get some good asteroids. We'll mine some stuff. We'll free your economy
00:55:56.180
to do what it can do. We'll share our technology with you. It'll be awesome. You'll be the best
00:56:01.380
leader that Russia ever had. You'll quadruple your GDP. There'll be statues of you everywhere.
00:56:06.940
And you will have ended war. I'm telling you, the golden age, you just have to accept it. It's
00:56:17.480
right there. It's right there. The hard part is getting people to just pick it up. It's like,
00:56:22.560
here it is. Here's all of your solutions. All of your solutions. We have all the answers now.
00:56:27.660
Here they are. People are like, I don't know. I don't trust those solutions.
00:56:36.640
So here's where the golden age, I think, is going to happen. I think that the energy shock
00:56:42.040
will cause us to be pro-nuclear in a way that we had to be. So our energy thing is now on a course
00:56:49.480
for full correction. It's going to be slow. But it's now on a very definite course toward full
00:56:56.320
correction. It's going to be nuclear. Energy will go nuclear. And through Tesla-like activities,
00:57:05.520
it will be also electric. I'm saying solar. So it's going to be solar and pretty much just solar
00:57:12.460
and nuclear. So I would say those two things solve climate change, or at least they're likely to be
00:57:19.220
enough to mitigate the worst problems as long as we're also mitigating things as we go. It's not the
00:57:25.600
only stuff. So that's better. What about the pandemic? Horrible, horrible thing, right? But don't you
00:57:32.600
think we got a lot better at handling the next pandemic? Like a lot better? And I feel like even
00:57:39.860
though you think these new vaccinations are killing people, some of you think that, I feel like what we
00:57:45.700
learned from that could create a platform for everything from vaccinations for cancer to vaccinations
00:57:51.440
for all kinds of stuff. Or we'll find out there was some problem with it. Can't rule that out.
00:57:57.980
I'm giving you the optimist view. Then look at the Ukraine-Russia war. Does it forever end the idea
00:58:07.540
that it's a good idea to attack your neighbor with tanks? Apparently not. Well, I mean, you know, up till now,
00:58:15.660
I guess Putin still thought it was a good idea to attack his neighbor with tanks. And at this
00:58:22.120
point, don't you think this will always be the cautionary tale? It's like, okay, all right, it
00:58:27.260
doesn't work. I think Afghanistan is bad of a situation as it was. At least it will always remind
00:58:35.200
us of what kinds of things not to do again. But Ukraine didn't look like Afghanistan. They looked
00:58:42.620
like such different places that maybe the lesson didn't transfer. But now you've got an industrial
00:58:49.040
country. You've got, you know, a backwards country. And neither of them could be conquered by the
00:58:55.260
Soviet Union or Russia. That should mean something. If you couldn't conquer either kind of country and
00:59:01.020
they're so different, maybe the whole country conquering thing isn't for you. All right. What else is
00:59:09.320
good? So we're going to solve energy. I think we're safer from pandemics.
00:59:19.080
And I think war looks less likely. I think the biggest problem is inflation.
00:59:25.880
All right. Let me give you the ultimate economic safety thought. Are you ready? Yeah. Most of you
00:59:34.260
have some concern, I would think, about inflation and GDP going down and maybe food shortages and
00:59:43.000
everything else. Most of you are starting to have a little anxiety about that, right? And gas prices.
00:59:47.360
Not a little. Maybe a lot. So some of you are having a lot of anxiety. I'm having anxiety about it.
00:59:55.640
And I'm rich. You know, relatively speaking. I can't even imagine how this would feel if I was just
01:00:03.820
squeaking by. This would feel like insanely bad. But let me give you the one positive thing that I
01:00:13.120
can give you. I've told you before that economics is real complicated stuff. I have a degree in economics
01:00:19.620
and I'm confused half the time. But there was one rule that I always look to that tells me where
01:00:25.760
things are headed. There's just one metric. If you get that one metric right, all the other things can
01:00:32.400
work themselves out. But if you get the one thing wrong, nothing can work itself out. Do you know what
01:00:38.840
the one thing is? Employment. Employment rate. Not even raises, not even cost of living adjustments, not unions.
01:00:51.000
Just the pure number of people who have jobs compared to the numbers who want them. And the fact that we have
01:00:56.880
labor shortages now. So we could actually use more workers. Even immigrants are in demand. Apparently, with all the
01:01:05.680
immigration we have, it's still hard to get the harvest picked, I think. So we're actually
01:01:10.920
understaffed. I don't know of any situation where you had close to full employment where you couldn't
01:01:18.280
work out the other stuff. You know what I mean? And here's the math of why that is so important.
01:01:25.040
The difference between an unemployed person and an employed person is a gigantic drag on the rest of
01:01:30.820
the people, right? If somebody is employed and they're taking care of themselves and maybe adding something, you know,
01:01:37.460
to the taxes, then that's taken care of. But if they're unemployed, not only are they not adding, but they're
01:01:44.680
subtracting. You have to pay them to live or they die. So one unemployed person is really, really expensive compared to
01:01:53.340
almost any other problem. So if you get that one thing right, the other stuff can be really painful
01:02:00.500
for a while, but the odds of it working out in the long run are real good. So when you're seeing people
01:02:06.780
who seem to be the most knowledgeable and experienced about economics, and they see all the things you see,
01:02:13.080
you know, they're seeing the inflation, they're seeing the, you know, it might get worse. They're seeing
01:02:17.520
everything. Supply chain problems, you know, China problems, blah, blah, blah. They see all that,
01:02:24.760
but they don't look like they're panicking, do they? Have you noticed that? There's no tone of panic,
01:02:31.820
even though all the metrics are sort of awful. Because that one thing is right, the jobs.
01:02:39.400
And I think the people who know the most are sort of just looking at that one and saying,
01:02:43.460
you know, probably much like I do, say, okay, as long as that one thing's okay, at least our
01:02:50.560
foundation is in good shape, right? The foundation is strong. Then you can work out the rest.
01:03:01.340
Mushrooms is tulips. Yeah, I see what you're saying.
01:03:08.580
Labor participation. Yes, that's a big one as well.
01:03:13.460
Total employment is still below pre-pandemic, but it's still good.
01:03:24.740
How do we fix BlackRock buying all the single-family homes?
01:03:29.540
The solution for housing is better housing, not building the same kind of homes and reselling them
01:03:36.520
over and over again. There is definitely a way to build a home for 10% of what it costs to build a home.
01:03:43.460
So I think that's going to be another part of the golden age.
01:03:50.220
I think that if you got rid of zoning and you turned it into a kit,
01:03:55.480
I've talked about this before, but imagine designing homes in which all the parts are an even amount.
01:04:05.140
In other words, a room in this house, house of the future, could be 10 by 12, but it could never be 10 by 12 feet and 2 inches.
01:04:16.520
And the reason is so you'd never have to cut anything.
01:04:20.360
So if you're putting it in the floor, you buy one-foot squares, and you put in as many as you need for the squares,
01:04:26.480
and nothing gets cut, and maybe it's a kit for them so everything snaps together,
01:04:30.940
so you could unsnap it and move a wall if you needed to, etc.
01:04:38.960
It's just that nobody has a business model to make money from making that.
01:04:43.380
The old Sears kits, so I know there's some historical examples,
01:04:49.980
but those kits I believe you still had to cut, didn't you?
01:04:55.240
I think those were, if you imagine what we could do now compared to what they could do in those days,
01:05:02.420
I imagine that we could make a kit that would be way better, just way better.
01:05:09.860
And then if you make the homes with, so here's how I do it.
01:05:14.660
I'd design perfect rooms, and then you could design a house that used as many of those pre-set rooms
01:05:26.620
And then you could build almost any kind of a house, but you never have to cut anything.
01:05:34.240
The tough part would be, I mean, I think you could even do the plumbing.
01:05:40.660
In fact, I'll bet you could make a house that's self-aware.
01:05:44.420
Imagine a house that comes as a kit, and each part has a camera in it.
01:05:49.560
So every part you put becomes alive, and it can see around itself, and then it attaches to some brain.
01:05:58.380
So the house could see a leak in your wall before you knew it was there.
01:06:02.540
But it could also tell if you'd assembled it correctly.
01:06:05.640
So you put the new component on, and it becomes alive, because it's like electrically connected to everything else.
01:06:12.900
And suddenly the house knows if you put the thing in the right place.
01:06:16.680
So you put the new piece on, and the house goes beep, beep, beep.
01:06:27.940
So if you made all the components alive, the house would help you assemble itself.
01:06:35.420
It'd be watching for any defects, and it would warn you every time.
01:06:41.280
That, ladies and gentlemen, is the best show ever.
01:06:47.640
Probably some of you are about to start one of the best days ever.
01:06:51.700
And I think that you would agree that today is the beginning of the rest of your life.
01:07:00.600
I read it in a greeting card once, so I know it's true.
01:07:03.860
And that is all I have to say to you on YouTube.