Timcast IRL - Tim Pool - April 10, 2023


Timcast IRL - Anheuser Busch Market Cap Drops BILLIONS As Boycott WORSENS w-Peter Boghossian


Episode Stats

Length

2 hours and 3 minutes

Words per Minute

197.74442

Word Count

24,372

Sentence Count

1,806

Misogynist Sentences

26

Hate Speech Sentences

19


Summary

On this week's show, we're joined by Peter Boghossian, founder of the Sokol Squared hoax and author of The Woke Papers, to talk about how he and his co-creator, Helen Pluck Rose, created the first "Woke Papers" and how they changed the way we think about feminism.


Transcript

00:00:00.000 you you
00:00:24.000 you it's a very interesting weekend
00:00:43.000 That's Luke, that's not me.
00:00:44.000 Serge, what are you doing?
00:00:45.000 Oh, it's just backwards.
00:00:47.000 It's backwards.
00:00:48.000 I'm used to pushing that top corner, so.
00:00:50.000 We're in a mobile studio setup in Austin, Texas, because we're here all week.
00:00:54.000 We've got a bunch of amazing guests, and we're doing a live show this Friday.
00:00:58.000 But let's talk about news.
00:00:58.000 That's what you're here for.
00:01:00.000 It was a crazy weekend.
00:01:00.000 Donald Trump tweeted just, World War III.
00:01:03.000 That's all he said.
00:01:03.000 And then he gets a whole bunch of news written about him.
00:01:06.000 And China says they're ready to fight.
00:01:08.000 Oh boy.
00:01:09.000 Pentagon leaked documents suggest Russia's actually winning, and surprise, surprise, they're lying to us, unless those leaked documents are actually just propaganda in and unto themselves.
00:01:19.000 But you know what?
00:01:20.000 I wanted to leave with a story that I think was kind of a white pill moment that would be inspiring for all of you, and that's Anheuser-Busch has seen a multi-billion dollar market cap drop off.
00:01:31.000 Since the start of this controversy when they hired Dylan Mulvaney, reportedly sales are dropping, rumors are circulating that they're not selling product, videos are popping up all over the internet showing beer sections where the Budweiser and the Bush products are all still there, and other companies are all sold out.
00:01:45.000 This could all be propaganda too, but you know what?
00:01:48.000 I want to at least believe that we can win if we organize.
00:01:52.000 And you know what is also true?
00:01:54.000 It's entirely possible the naysayers who are acting like this is not a big deal, Are actually the propagandists trying to say stop doing what you're doing.
00:02:01.000 When in reality, I look at the hard numbers.
00:02:04.000 Since the start of this controversy, their stock is down I think around 5% for a multi-billion dollar market cap drop.
00:02:09.000 So we can organize, we can vote with our dollars, and we can win.
00:02:13.000 So we got a lot of stuff to talk about.
00:02:14.000 But before we get into all that, let me try and slide forward in this awkward situation and Everyone head over to TimCast.com.
00:02:23.000 Become a member by clicking Join Us to join our members.
00:02:27.000 We're doing what we can here with this mobile studio, so thanks for hanging out.
00:02:31.000 But click Join Us, become a member, support our work as a member.
00:02:34.000 You'll get access to our members-only uncensored show.
00:02:37.000 We're going to have one of those up tonight at about 10, 10 p.m.
00:02:39.000 Eastern Time.
00:02:41.000 You can also join our discord server where you can chat with a whole bunch of like-minded people and after you've been a member for six months you get access to the VIP chat and voice chat where you can submit questions actually call in and ask us questions and you can jump the line if you want by signing up for 25 bucks per month and then of course we have the elite club which is more community building and we got a bunch of really cool programs every friday we're going to start shouting out members projects companies or otherwise because Y'all are actually the people sponsoring the show already, so Friday will be your day.
00:03:11.000 So smash that like button, subscribe to this channel, share the show with your friends.
00:03:15.000 Joining us tonight to talk about all of this and a whole lot more is Peter Boghossian.
00:03:19.000 Hey, thanks for having me.
00:03:20.000 Should I say Doctor Peter Boghossian?
00:03:22.000 No, Peter.
00:03:22.000 Peter's fine.
00:03:23.000 But you're our doctor, right?
00:03:24.000 Correct, but Peter's more... Peter's fine?
00:03:26.000 So don't listen to anybody because they're a doctor.
00:03:29.000 In fact, it probably means if they got their degree recently, it's probably a bad thing.
00:03:32.000 It probably is an indication that they're brainwashed.
00:03:35.000 But for people who aren't familiar, I mean, aside from a whole bunch of accolades, I think a lot of people might know you from the Sokol Squared hoax, where you, Helen Pluck Rose, and James Lindsay created these, what would you call them, woke?
00:03:49.000 Fake papers, woke papers that we played on the moral prejudices of journals and showed that the bodies of literature were corrupt.
00:03:57.000 The bodies of literature, by the way, that we're forming public policies upon.
00:04:01.000 So, simply put, You, in one instance, took a chapter from Mein Kampf and changed it to feminist talking points.
00:04:09.000 Correct.
00:04:09.000 You changed key pronouns to feminist nouns, and they accepted it.
00:04:14.000 Yeah, we had two papers.
00:04:15.000 one we just changed Jews for white men and the other one we...
00:04:21.000 That's brilliant.
00:04:22.000 Thank you.
00:04:23.000 And the other one we rewrote with that in mind.
00:04:26.000 So it's good to be here with everybody.
00:04:28.000 Yeah, and I think we have a lot to talk about considering what's going on with the Anheuser-Busch thing, wokeness.
00:04:34.000 Now there's talk about the Mario movie making a bunch of money, and so we'll talk about that stuff.
00:04:38.000 And then, of course, Riley Gaines is a big story.
00:04:40.000 University of San Francisco.
00:04:42.000 That's a big story in my world.
00:04:44.000 I think it's massive.
00:04:45.000 She got attacked.
00:04:45.000 She was chased by a mob and then locked in a room for, what, three hours.
00:04:50.000 So we'll get into all that stuff.
00:04:52.000 We'll get into it.
00:04:52.000 And what's really fascinating to me That's interesting, but the university's response is what we should drill down on.
00:04:59.000 Right on, absolutely.
00:05:00.000 And Luke Rutkowski's here!
00:05:01.000 Well, yeehaw!
00:05:02.000 Welcome back, beautiful and amazing human beings!
00:05:05.000 It's good to be back!
00:05:07.000 The dollar is collapsing, the petrodollar is being replaced, the proxy war is expanding, everything we told you was going to happen is happening, but it's...
00:05:15.000 happening on a bigger scale, so things are going to get a lot crazier.
00:05:18.000 It's good to be back.
00:05:19.000 Since we're in Texas, I decided to wear my Ron Paul, if I told you so, was a person shirt, which you could get on thebestpoliticalshirts.com.
00:05:29.000 Good to see you guys.
00:05:29.000 It's been a while.
00:05:30.000 It's the most cowboy you've ever been, man.
00:05:31.000 I know.
00:05:31.000 I love it.
00:05:32.000 You gotta fit in.
00:05:33.000 Yeehaw.
00:05:34.000 Hey, Peter, what's your doctorate in?
00:05:36.000 uh... education well i mean cross on the stick on
00:05:40.000 so it's just a bizarre this this first story right here we got surges out on
00:05:43.000 the money is on the camera so i can't have a chance to instance
00:05:47.000 all right let's uh... let's jump to this first story we got this in the daily
00:05:50.000 mail bud lights vp says she wanted to update the fratty and out
00:05:54.000 of touch branding with inclusivity
00:05:57.000 days before dylan will banish controversial partnership with beer was
00:06:00.000 unveiled And I'm gonna humble brag here.
00:06:03.000 I called this in my conversation with Vivek Ramaswamy when I said what's likely happening is these millennial women are getting promoted in these companies, they hold these views, then they start adopting these things.
00:06:14.000 It is not It is not like some 56-year-old Gen Xer or whatever running the company said, this is what we got to do.
00:06:20.000 It is a young millennial person taking these jobs.
00:06:23.000 But of course, what everyone really came here for is this tweet from DC Drano.
00:06:27.000 He says, on March 31st, Anheuser-Busch had a $132.38 billion market cap.
00:06:34.000 As of today, it is now $128.4 billion.
00:06:38.000 You know what that means?
00:06:39.000 The woke Bud Light campaign has already shaved off nearly $4 billion in company value.
00:06:45.000 Don't let Democrats lie to you.
00:06:47.000 Conservative boycotts work.
00:06:49.000 So, uh, white pill moment, I suppose?
00:06:51.000 Boycotts do work.
00:06:52.000 Voting with your dollar does work.
00:06:54.000 And this one is the easiest thing for people to do.
00:06:56.000 Because everyone's always like, what can I do?
00:06:59.000 I don't want to risk my family.
00:07:00.000 I don't know how I can affect the electoral system.
00:07:03.000 Who do I vote for?
00:07:03.000 I can only vote every few years.
00:07:05.000 This one's really easy.
00:07:06.000 Don't buy Bush.
00:07:07.000 Don't buy Anheuser products.
00:07:09.000 Don't buy Budweiser.
00:07:10.000 Don't buy Bud Light.
00:07:11.000 And if this can be kept up... That's the thing right there, right?
00:07:16.000 So it's only worked if it can be kept up.
00:07:20.000 So it's got to be kept up, and it's got to be sustained to a certain degree where you need one thing to happen.
00:07:27.000 Budweiser apologizes.
00:07:28.000 Yeah, that's it.
00:07:29.000 And until the apology, don't buy Bud.
00:07:32.000 Well, yeah, right.
00:07:33.000 When they do apologize, I say buy a bunch.
00:07:35.000 What market forecasters and everybody needs to see is that when they did this campaign with Dylan Mulvaney, their stock drops, their sales drop, and then as soon as they apologized, it goes right back up.
00:07:47.000 So I should go up even higher.
00:07:48.000 If they apologize, I'll buy a ton of beer.
00:07:50.000 I'm not going to drink it, but you know, I'll buy it.
00:07:52.000 I think the other thing to think about is exactly what you said.
00:07:56.000 What can you do?
00:07:57.000 And it's not that you can give money to cause you like what you can, but you can stop giving your money things.
00:08:03.000 For example, you can stop giving your money to your alma mater.
00:08:06.000 We have a project that we sponsored called Don't Donate, where we ask people the simplest ask in the world.
00:08:12.000 Don't donate to your alma mater.
00:08:13.000 It's not the same place.
00:08:15.000 You're pushing woke ideology.
00:08:18.000 You're literally giving it a kind of financial oxygen.
00:08:21.000 Same thing with Bud Light.
00:08:23.000 Don't buy it.
00:08:23.000 Any companies that you've been donating to or working with, check and see if they've been bought lately, because there's been a lot of corporate conglomeration in the last decade.
00:08:30.000 Right, like Anheuser-Busch merged with some... InBev, yeah.
00:08:34.000 Yeah, so now there's a bunch of other brands.
00:08:36.000 Anheuser-Busch actually really, really big.
00:08:38.000 Here's my favorite.
00:08:38.000 Marjorie Taylor Greene tweeted out a picture of buying Coors.
00:08:41.000 And then some leftist posted Coors woke campaign as well.
00:08:45.000 But here's my thing.
00:08:46.000 I don't care.
00:08:47.000 I do not care if Bud Light makes a pride can.
00:08:51.000 That's not what I care about.
00:08:53.000 What I care about is Dylan Mulvaney represents the algorithmic crisis that is causing teenage depression, that is expanding and creating wokeness, in my view, and then they reward it.
00:09:05.000 Not only that, but TikTok, I believe, is likely a Trojan horse.
00:09:10.000 You know, the story goes that in China they get more controlled and academic content and in the U.S.
00:09:15.000 it's all weird woke stuff and people who think they're frogs, you know, frog, frog self-pronouns and stuff like that.
00:09:20.000 Very horrible ideas that are corrupting people's brains and that, Dylan Mulvaney, is a perfect example of what those algorithms do.
00:09:28.000 So when Bud Light dumps money into that, promotes it, they're actually making the algorithmic crisis substantially worse.
00:09:35.000 So my thing is like, get away from it.
00:09:37.000 Don't fund that stuff.
00:09:38.000 There's a lot of interesting things at play here because there's also the ESG score.
00:09:42.000 There's also fifth generational warfare.
00:09:43.000 We could go deep down the rabbit hole when it comes to explaining what's really going on here.
00:09:47.000 But on a simple level, I think if we explain what just happened, it was a company trying to sell beer to, quote, what they deemed a younger audience, but mainly a part of just preaching to a woke religion that they were trying to serve here.
00:10:01.000 And it's pretty clear that they kind of forgot about their customer base.
00:10:05.000 And a lot of people are discussing, hey, what beer should I drink here?
00:10:07.000 What's next beer?
00:10:08.000 What's next beer?
00:10:09.000 I should move on.
00:10:10.000 I think it's also important to note here that even woke people won't be buying Bud Light because it's like you think we're that simple where you could just put some person on there and that's going to make me want to buy your your fluoride filled piss drink?
00:10:24.000 No thank you!
00:10:25.000 I don't work Not that bullcrap!
00:10:26.000 Anywhere near me, it's an endocrine disruptor, and I think a lot of people, the discussion that I would love to have here, to have a deeper discussion, maybe instead of replacing your beer, maybe let's stay away from the moob-creating endocrine disruptors that do, of course, have a very negative effect on you, and I think this is a perfect time to start that conversation being like, hey, don't support this bullcrap because it's bad for you.
00:10:46.000 I don't drink.
00:10:47.000 And how evil do you have to be?
00:10:48.000 Because the video of this VP came out and she's talking about, we have to appeal to a younger audience.
00:10:53.000 She has the rainbows behind her.
00:10:54.000 What kind of evil person do you have to be to say, we're going to get young kids hooked on beer and alcohol?
00:11:01.000 You are a low vibrational, like bad human being when you're such a corporatist where you're like, we got to get all these people hooked on alcohol.
00:11:09.000 You're a bad person.
00:11:10.000 I'm just sorry.
00:11:11.000 Young drinkers is how she phrased it.
00:11:13.000 Young drinkers, that's what she said.
00:11:15.000 Peppered with woke language throughout.
00:11:17.000 And so here's the thing.
00:11:18.000 They're targeting college kids.
00:11:19.000 They're targeting people who are under the legal age to drink.
00:11:22.000 Now, that being said, I think 18 should be the legal drinking age.
00:11:26.000 I think it's ridiculous they raised it to 21 in the first place.
00:11:28.000 And I actually think beer is actually pretty great.
00:11:31.000 I don't drink it.
00:11:32.000 The problem is abuse, not having some.
00:11:35.000 I might have a beer once or twice a year, to be completely honest, depending on the situation, but it's not going to be Bud Light, because Bud Light's just, it's like not really beer.
00:11:41.000 Newcastle.
00:11:43.000 There was even a new medical study that came out showing that even moderate drinking has very negative cognitive effects for your brain, and ages your brain, and stops blood flow to your brain.
00:11:53.000 The moop thing is real.
00:11:55.000 There's a reason Bill Gates has moops, okay?
00:11:58.000 And it's not accidental.
00:11:59.000 It's not a coincidence.
00:11:59.000 You're saying he drinks too much beer?
00:12:02.000 That and, of course, all the horrible pesticides and byproducts.
00:12:04.000 There's also a lot of glyphosate in, of course, the beer as well.
00:12:07.000 It's good to be back to get this off of my chest here.
00:12:09.000 I feel awesome to finally be able to explain a lot of this stuff because, you know, at the end of the day here, it is a corporation trying to sell a product, a product that does have a lot of negative consequences for people.
00:12:20.000 That should be the discussion here, I think.
00:12:21.000 So let me ask you this question.
00:12:23.000 Maybe we can all agree, I think everybody would be better off if they just drank water.
00:12:29.000 I have some water right here.
00:12:32.000 Plastic Ian!
00:12:33.000 Plastic water!
00:12:34.000 I'm on the road.
00:12:35.000 You know that we actually do have glasses over there.
00:12:37.000 Sorry to interrupt.
00:12:38.000 I didn't find them, so I'm using a measuring cup.
00:12:40.000 That's why I like you, Peter.
00:12:41.000 Hey, it holds water.
00:12:42.000 It's good enough.
00:12:43.000 Here's an important thing to consider, too, for this stuff.
00:12:46.000 The reason why I really, really am energized, enervated by this story is Bud Light deserves to lose money, not just because we disagree with the ideological position of say, you know, TikTok, Dylan Mulvaney, or whatever.
00:13:01.000 Bud Light deserves to lose money because this has to be one of the stupidest business decisions, irrespective of wokeness, in that You've got people from the ages of 21 to on average 79.3 years old.
00:13:18.000 Your market share is 50 plus years.
00:13:21.000 57, 58 years.
00:13:23.000 And what they've done is said, we've got to get rid of the frat culture and appeal to a younger audience.
00:13:28.000 So they're targeting people from the ages of 21 to 25.
00:13:33.000 In an effort to sell beer at the expense of 25 to 80.
00:13:37.000 And they've got to understand the amount of financial duress people are under that are from the age of 21 to 25.
00:13:43.000 Young people, it's a constant theme on this show and others that people don't have money, they don't have direction, the school is too expensive, they're in debt.
00:13:50.000 Well, they do have money, but the money's being devalued.
00:13:52.000 The money's being devalued.
00:13:53.000 But real quick, Ian, that actually makes them smart.
00:13:56.000 But they're targeting a segment of population that doesn't have money.
00:13:59.000 That's the weird thing.
00:14:00.000 That's actually a good point.
00:14:02.000 You know, it's funny because they say when times are good, people drink.
00:14:05.000 When times are bad, people drink more.
00:14:06.000 And so, right, and so maybe they're hoping, like, maybe the real gambit here is these young people are depressed and listless.
00:14:14.000 They're gonna get drunk.
00:14:17.000 Do you see it?
00:14:17.000 What better way to forget your sorrows?
00:14:19.000 What better way to live in your pod and eat the bugs?
00:14:21.000 And of course, if you remember during COVID, the alcohol industry skyrocketed.
00:14:26.000 They were allowed to have liquor stores that were allowed to be open.
00:14:29.000 Meanwhile, the mom and pop stores weren't allowed to be open as well.
00:14:32.000 So a lot of people drank a lot.
00:14:33.000 A lot of people who are depressed drink a lot.
00:14:35.000 And it makes a lot of emotional issues, a lot of emotional problems that much worse for a lot of individuals.
00:14:40.000 So when you're going after a key demographic that still has a part of their brain developing, and you're saying, let's put booze in there.
00:14:46.000 Come on.
00:14:47.000 Let's call it out for what it is here.
00:14:49.000 To me, a lot of this is just absolutely disingenuous.
00:14:52.000 I don't care.
00:14:53.000 I think there was just someone at Bud Light saying, we just need to make money.
00:14:56.000 Let's just make as much money as we can.
00:14:58.000 We don't care how we're going to do it.
00:15:00.000 And then they bring in the woke religion as a part to be subservient.
00:15:04.000 See, I was talking to Vivek Ramaswamy last week, and this is where the conversation was going.
00:15:08.000 And I said, I disagree.
00:15:10.000 What I think we're seeing at these companies, they hire a 24-year-old woman 10 years ago.
00:15:16.000 She's a marketing intern.
00:15:18.000 Two years later, she's full-time staff now.
00:15:21.000 She's a marketing producer or something like that.
00:15:25.000 It's now been a decade.
00:15:27.000 Her boss, who's 56, goes, what's her name?
00:15:32.000 Do you know her name, Luke?
00:15:33.000 Alyssa Hiners.
00:15:34.000 So here's what I'm imagining.
00:15:36.000 He goes, Alyssa!
00:15:37.000 It's about time I retired.
00:15:40.000 I'm about 60 some odd years old.
00:15:41.000 Let's say he's in his 60s and he's like, I think this is a job you can handle.
00:15:45.000 And she's like, thank you so much.
00:15:47.000 And then as soon as you walk outside, she goes, let's sponsor Dylan Mulvaney.
00:15:51.000 And that's what happened.
00:15:52.000 There was like, There are these people on the left that are saying they did all this market research.
00:15:56.000 They know exactly what they're doing.
00:15:57.000 I'm like, no, no, they don't.
00:15:58.000 The people who are woke, who are ideologically driven, are not thinking logically about numbers.
00:16:03.000 They're thinking, we are now wearing the institution like a skin suit.
00:16:07.000 Let's advance our cause.
00:16:08.000 I'm automatically thinking of that Vice News skit where they were taking adult toys and throwing them at the wall randomly.
00:16:15.000 That could be it, too.
00:16:16.000 You never know.
00:16:17.000 This woman clearly is woke, though.
00:16:18.000 We saw this video where she's like... She has pronouns in her bio.
00:16:21.000 She's peppering her language.
00:16:23.000 And I think that the key thing is what you said, Tim, it's that these people have intentionally trained themselves not to think.
00:16:31.000 All of their information comes through a funnel to promote wokeness.
00:16:36.000 But I guess what I was thinking about is, have you heard the expression, there's no such thing as bad press?
00:16:41.000 Yeah.
00:16:42.000 This seems to belie that.
00:16:44.000 There is bad press and bad press can harm.
00:16:49.000 I think the phrase there's another thing is bad press typically applies to an individual right I think someone like Ben Shapiro has actually greatly benefited from bad press but to be fair there's a double-edged sword that comes along with it in that The bad press for an individual and for a commentator, they become more prominent.
00:17:06.000 For someone like Ben Shapiro, or for any of us here, is that the more they talk about you, the more you become the subject and the figure to be talked about.
00:17:15.000 If you are Anheuser-Busch, you are on top of the mountain.
00:17:18.000 There's nowhere else to go.
00:17:19.000 The only press they're gonna get is gonna be negative for their stock.
00:17:23.000 So, bad press, bad for your stock.
00:17:24.000 Individuals, we don't have stock, so the more attention we get, the more we can capitalize on it.
00:17:28.000 For Bud Light, I'll put it this way.
00:17:31.000 They've got an opportunity when they bring politics into beer, which seems to make no sense in my opinion for the most part, but when they, this opportunity is, can we capture a large portion of one demographic?
00:17:41.000 Well, my attitude is how about you don't do any of it and you try and just capture as much of the entirety of people who drink beer by being a neutral beer company.
00:17:48.000 As soon as they enter the political fray, they segment off a large portion of their audience.
00:17:53.000 If you are a political commentator or a celebrity, you're not going to get a market cap that is the entire population of the planet.
00:18:00.000 Now imagine a company like Coca-Cola.
00:18:02.000 Coca-Cola's market cap literally is every human on earth.
00:18:05.000 Except for maybe like North Sentinel Island where they've never made contact with people or something like that.
00:18:09.000 So, they have no incentive whatsoever to be divisive.
00:18:12.000 And bad press is bad if your market cap is everyone.
00:18:16.000 Coke, Bud Light, whoever else, they want everyone on the planet drinking their product.
00:18:20.000 Well, go to war with one half of them and you'll only lose customers.
00:18:24.000 If you're a commentator, and you're a celebrity, so who cares if 80% of the world hates you?
00:18:28.000 20% likes you and you'll make money off them.
00:18:30.000 So, bad bad all around for Bud Light.
00:18:33.000 You said something, Peter.
00:18:34.000 I think you were alluding that they didn't seem to know what they were doing.
00:18:37.000 Is that what you said earlier?
00:18:38.000 Yeah, what I think is happening is that these folks, and Alyssa has pronouns in her bio, and she peppers all of her language, and I would encourage people to watch those videos where she's explaining why she did this.
00:18:51.000 They're not capable of independent, rational thought.
00:18:56.000 They view themselves as white knights who promote the ideology.
00:19:00.000 This is a way to promote her moral values.
00:19:03.000 And within that, then she convinces herself that, well, it will increase market share, will increase sales, but it's just not true.
00:19:10.000 I think it's the word inclusivity.
00:19:11.000 Do we have that video of Elissa?
00:19:13.000 Oh, that video is wonderful.
00:19:14.000 It's so great.
00:19:15.000 I don't know why she made it.
00:19:16.000 It's incredible.
00:19:16.000 Incredible that she inundated herself like that, but I guess everyone knew who she was on the team anyway
00:19:20.000 And the way she says we need this we're looking to do in I think it's the word inclusivity inclusivity
00:19:25.000 Yeah, something something something and it and an inclusive something something you're like why she used the word
00:19:29.000 inclusivity twice She doesn't know what it means. That's why well that video
00:19:32.000 that you mentioned was from March 30th And ever since I haven't seen much kind of any kind of
00:19:37.000 words from Bud Light afterwards after this. Oh, that's from okay
00:19:40.000 Yeah, so that's from an earlier video where she was explaining what she was doing
00:19:43.000 And I want to underscore something Tim said that if you really want to show that you can change and make a difference, it has to be sustainable.
00:19:51.000 Like you have to literally continue to not buy the product until they apologize.
00:19:56.000 And if you don't do that, then you're actually working against your interests because companies are going to keep doing this.
00:20:02.000 I'll be honest, I don't think an apology is enough.
00:20:05.000 Issue some crap they don't believe and then try and screw them.
00:20:07.000 The apology is everything.
00:20:09.000 The apology is everything.
00:20:10.000 I feel like making them go bankrupt is everything.
00:20:11.000 Because it's a signal.
00:20:12.000 It's a costly signal.
00:20:13.000 It means they will, if they issue an apology, that means Anheuser-Busch said, the risk to our business from the far left cult is less than the risk to our business from pissing off these people.
00:20:25.000 I see.
00:20:26.000 So you're not trying, that's a good point.
00:20:27.000 You're not trying to destroy the company.
00:20:28.000 You're trying to change the direction the company takes.
00:20:30.000 Yeah, and we're not going to be held hostage to woke maniacs.
00:20:34.000 We're just not going to be held hostage to these people.
00:20:36.000 Let me pull up this video here.
00:20:38.000 We have this video from Chuck Colesto.
00:20:40.000 It's a video of a guy, let's see, is the audio going to play?
00:20:44.000 So I don't know, is the audio coming through?
00:20:46.000 Give it to me, Chuck.
00:20:48.000 Nope, audio's not coming through at all.
00:20:50.000 But I'll just explain for you, simply put.
00:20:52.000 No, it's just not coming through for you.
00:20:53.000 I'm not seeing it on the... I'm not hearing it.
00:20:55.000 I'm not hearing it either.
00:20:56.000 And it's not coming up on the thing.
00:20:59.000 Yeah, so the audio is not playing, but that's fine, that's fine.
00:21:02.000 This is a video of a guy who's a distributor.
00:21:04.000 He works for an affiliate company, he says.
00:21:06.000 I want to make sure I clarify that.
00:21:08.000 He's got Bud Light on his arm at a liquor store.
00:21:11.000 Now this could be fake, you know, I just want to make sure that's clear, like, hard to verify, just a random video on the internet.
00:21:16.000 But this is just one of several videos I've seen of people in similar situations pointing out that the shelves are loaded with Anheuser-Busch products, and this guy says that he works for an affiliate company that only sells Bud and Bud Light products, and he's never seen sales this low, which worries him, because if we don't sell these products, I don't feed my family.
00:21:37.000 Like, this is my company, like, that I work for, we gotta do this.
00:21:40.000 And he goes, so, uh, thanks, Anheuser-Busch, maybe I won't feed my family.
00:21:45.000 I definitely think it's a little hyperbolic, but we've seen a bunch of these videos where people are reporting on the ground that sales are dropping.
00:21:53.000 We've seen the stock dropping, and I think it lines up.
00:21:55.000 And I think this, the left operates on lies and manipulation.
00:22:01.000 I don't want to say all the left.
00:22:02.000 I want to say the woke cult operates on lies and manipulation.
00:22:06.000 And so they will twist anything as a victory, and they will twist anything you say into something negative.
00:22:13.000 So here's what I have to say.
00:22:15.000 I'll be fully honest.
00:22:16.000 Maybe these videos are being faked by people who want to make it seem like we're winning, but I'll just take it.
00:22:21.000 I look at their stock prices.
00:22:23.000 Their stock prices are way down.
00:22:24.000 Yeah, so that's an independent metric, whereas the videos are not.
00:22:27.000 But you know what?
00:22:28.000 If the stocks are dropping, and there's rumors of sale decline, And then I see these videos, I'm more inclined to believe that these and other videos I've seen are probably true.
00:22:38.000 There's one video where it's just a guy standing in a supermarket and he just pans left and right, and then you see all the beer's gone, and then he pans over and all the Budweiser products still there.
00:22:48.000 Look, we're in Austin, at some point I'll go to a supermarket, why don't we just go to a supermarket and do it ourselves?
00:22:54.000 Yeah, because I wonder how pronounced it is in a place like Austin.
00:22:58.000 Yeah.
00:22:59.000 Right?
00:22:59.000 So here's the other thing we're hearing, the guitar player from The Offspring, he countered Kid Rock and he was like, we're going to stock all of our shows with Anheuser-Busch products to counter the bigots.
00:23:08.000 It's like, okay, I mean, this is...
00:23:11.000 It is a bit silly, I suppose, but hey, it's a culture war for a reason.
00:23:15.000 I mean, using a guy that claims to be transgender to sell your beer is a bit bigoted.
00:23:20.000 They're like using Dylan Mulvaney.
00:23:22.000 It's not like... It's even worse to use a guy to market tampons and sports bras.
00:23:29.000 Yeah, the sports bra thing was kind of... It's like using his identity.
00:23:31.000 It's so just underhanded dirtiness to sell product.
00:23:34.000 Here's what I think.
00:23:35.000 I'll tell you what I think.
00:23:37.000 TikTok, my personal opinion is, it's not real.
00:23:40.000 My opinion has never been that, my opinion has always been that it's not real.
00:23:45.000 And my opinion, TikTok, ByteDance, whatever.
00:23:49.000 So I said this a few years ago.
00:23:51.000 I explained how a new social media platform can emerge.
00:23:55.000 And I've actually talked to tech companies about this.
00:23:57.000 Here's what you do.
00:23:58.000 Create an app.
00:23:59.000 Then market that app on Instagram, Facebook, or wherever to teenagers.
00:24:05.000 Say, the new cool app to do this thing.
00:24:07.000 Hire a bunch of young people to do a fun thing like, you know, singing songs.
00:24:13.000 Then, when they download it, and they make a post, give them 150 likes.
00:24:17.000 Give them 50 new followers.
00:24:19.000 They're gonna respond to that positive reinforcement, and they're gonna be like, whoa, I got a bunch of new followers.
00:24:24.000 You give them fake followers and eventually all of these kids are gonna be like, you still use Instagram?
00:24:29.000 That's so lame.
00:24:30.000 Why aren't you using, you know, Globo?
00:24:32.000 And they're gonna be like, what's that?
00:24:34.000 You're not on Globo, dude.
00:24:35.000 I got 5,000 followers.
00:24:37.000 And then the friend is gonna go, you have 5,000 followers?
00:24:39.000 Yeah, I'm a big deal.
00:24:41.000 They're gonna feel real good, even though there's no real humans behind that number.
00:24:45.000 But that will convince their friends to get on the platform as well.
00:24:49.000 You know what's terrifying to me about that, by the way?
00:24:51.000 What's terrifying to me is I have thought over the last few years that so many things were conspiracy theories that turned out to be not conspiracy theories.
00:24:59.000 And now I'm listening.
00:25:03.000 Well, that's what happens when you're looking at it from the other end of history.
00:25:06.000 And now I'm hearing this and I'm thinking, you know, that's completely feasible.
00:25:10.000 Yeah, especially with quantum computing on the horizon and artificial intelligence, you could have 100,000 followers that aren't real people that are commenting actual things in response to what you're doing.
00:25:19.000 Yeah, the governments have been doing this for a very long time.
00:25:21.000 I think the Israeli government was the first one to come out about it publicly, where they described whole entire factories of people that were online digital troll armies that were specifically making sure that they were engaging in debate and conversations that made the Israeli government look good.
00:25:36.000 Joe Biden just announced a whole campaign for online influencers that are going to be his influence army for the next 2024 presidential election.
00:25:44.000 So obviously they're paying for PR using our tax dollars to have people shill for them who are just saying and doing anything for a buck.
00:25:51.000 Dylan Mulvaney was one of them.
00:25:53.000 Yes.
00:25:53.000 Wait, hold on.
00:25:56.000 I saw this story.
00:25:56.000 This is the 50 Cent Army for the Democrats.
00:25:58.000 Yes.
00:25:59.000 And Del Mulvaney is going to be... I don't know if he's a part of the 50 Cent Army.
00:26:02.000 He was.
00:26:03.000 He went to the White House.
00:26:05.000 But as we know, Hillary Clinton hired influencers, high-level influencers, even someone I don't think we should even... Casey Neistat?
00:26:11.000 Yes.
00:26:12.000 To specifically make endorsement.
00:26:14.000 Hold on.
00:26:14.000 Was he hired by them?
00:26:16.000 There's different ways that, of course, they conduct their businesses through PACs and different organizations, so we have to be careful with exactly how we're wording here, but I think it's fair to say that money exchanged hands for the support of this particular candidate, that money coming from, of course, either SuperPACs or our Connections organization.
00:26:32.000 That's a bold accusation.
00:26:33.000 But when you have an influencer... No, no, no, wait, wait, wait.
00:26:36.000 In terms of Casey Neistat, this one matters.
00:26:38.000 He's the biggest vlogger.
00:26:40.000 He's not really doing as much content as he used to.
00:26:42.000 He came out and endorsed Hillary Clinton.
00:26:44.000 It is my personal opinion that likely constituted a monetary exchange from someone or somehow.
00:26:50.000 Because I don't see how you get a dude who films drones and longboarding and talking about filmmaking, all of a sudden coming out and endorsing a politician in an election.
00:26:59.000 But it's entirely possible he's just going along with the crowd and that's why it ended up happening.
00:27:05.000 Be connected or not connected, but how much was the CNN deal?
00:27:08.000 That's also another one that also has a lot of people looking at the situation like, how much money did he make?
00:27:12.000 What did he do?
00:27:12.000 That whole thing didn't do anything!
00:27:15.000 So there's multiple ways of having people shill, but for the Biden administration to openly kind of discuss how they're going to be taking and giving money to people to shill for them, I mean, that's going to make everyone question like, wait, are you really in support of this person or are you just being paid to support this person?
00:27:34.000 When I go out into normal places, you know, if you're in D.C., you're gonna be around a bunch of people who are hyper-political, but when I go out to the movies, when I go to pool halls, when I go to the poker club or whatever, I have not met a single person who is pro-Biden.
00:27:51.000 Like, pro-Democrat.
00:27:52.000 I don't see them anywhere.
00:27:54.000 And, you know, of course, if I say something like that, I'm gonna get a bunch of leftists being like, haha, go touch grass, but like, dude, I've not seen it anywhere.
00:28:04.000 We've been going to Maryland Live, a casino, playing Hold'em 1-2.
00:28:09.000 Super, super inexpensive for those that are considering playing.
00:28:12.000 It's like you pay two bucks to play a hand, so it can be played cheaply.
00:28:16.000 Out of all the tables I've played at, and this is north of D.C., south of Baltimore, I have not met a single pro-Democrat person in this major urban center where it's just regular people hanging out from various ages, young guys to older guys.
00:28:30.000 Well, how do you know?
00:28:31.000 Does it come up at the tables?
00:28:32.000 Absolutely, every time.
00:28:33.000 Every time, without fail.
00:28:35.000 Someone mentions the cost of the game or the economy and then immediately, oh, Joe Biden, Joe Biden, he's bad.
00:28:40.000 Now, don't get me wrong.
00:28:41.000 I met a guy who said he thought the J6ers should be executed and he hated Donald Trump.
00:28:45.000 That I get, but I've not met anybody who was like, I like this guy.
00:28:49.000 Now that could be these people voted for him, and now they really, really do regret it.
00:28:54.000 Is it a blue area?
00:28:56.000 It's the bluest as blue can be.
00:28:58.000 I mean, we're talking about Baltimore and Maryland.
00:29:01.000 And this casino is, I think it's, I don't know what city it's in, but it's 30 minutes from downtown Baltimore.
00:29:09.000 It's like half an hour from D.C.
00:29:11.000 or Baltimore, right in between.
00:29:12.000 So everybody who's going there, it's people from the cities who are trying to play a game or whatever.
00:29:18.000 So I kind of feel like at this point, the reasoning for why Biden's building a 50 cent army is fairly obvious.
00:29:25.000 I do not see them being able to pull off another victory.
00:29:28.000 I'm seeing all these polls where they're like, Biden is beating Trump.
00:29:31.000 In these states, Biden is beating Trump.
00:29:32.000 And I'm like, don't care.
00:29:33.000 Don't believe it.
00:29:34.000 I mean, look, ballot harvesting may come to save the day, but It's clear that they're going to be ramping up what I would describe as unethical and malicious tactics in the election.
00:29:43.000 And the idea of a Democrat... So for those who don't know, the 50 cent army, what is it, in China?
00:29:49.000 Anybody who posts on social media pro-CCP propaganda, they get paid 50 cents every time they do or something like that.
00:29:55.000 So this is basically what we're going to start seeing with Democrats.
00:29:57.000 And that's probably what's already happening.
00:29:59.000 So we're talking about Anheuser-Busch, we're talking about Dylan Mulvaney and the algorithmic manipulation.
00:30:04.000 That is what the Democrats are riding that wave.
00:30:07.000 Yeah, the thing you didn't say, which I think needs to be said, which speaks to your point, is what this does to the integrity of the system overall.
00:30:15.000 This is what causes a legitimacy crisis, right?
00:30:18.000 So, we don't trust our institutions, we now think the people we like on social media, they're being paid certain... I've never taken a single penny for anything, my Twitter feed, anything.
00:30:27.000 I think that that compromises the integrity, but when you do that in a political level, it's terrifying to me.
00:30:33.000 Terrifying.
00:30:34.000 Absolutely, and especially when you connect it to something like the social credit score, which they do in China.
00:30:38.000 Because if you complain about the government in China, you of course get downranked with your social credit score.
00:30:43.000 You can't get high-speed internet.
00:30:44.000 You can't travel on airplanes.
00:30:45.000 You can't buy first-class tickets on trains.
00:30:49.000 And that's being tested in China, which they're trying to roll out and bring out here in the United States, as there's already a corporate social credit score that no one really likes to talk about.
00:30:57.000 It's not a government institution, but the corporate social credit score on everyone here already exists.
00:31:01.000 There is an ESG score with the corporations, but there's even another one with each citizen and civilian with all of your private records, especially from social media, being compiled together and then soon going to be used against you later.
00:31:13.000 What's that called?
00:31:14.000 Is that a program they're doing?
00:31:15.000 No, there's many different corporations keeping private records, and they sell all these private records about you and your itinerary, and what you do, and what you like, what you dislike, your views, your opinions.
00:31:23.000 All of that is databased.
00:31:25.000 So there's huge databases being built about particular individuals.
00:31:28.000 Those, again, are being used by the corporations.
00:31:32.000 They're being used by the government in China, and then that's where they're testing the 15-minute cities.
00:31:36.000 So if you say something bad, well, your 15-minute city might be a five-minute city.
00:31:40.000 And they're already locking down grids.
00:31:42.000 They're already testing it.
00:31:43.000 They're doing a trial run of that in the United Kingdom.
00:31:45.000 There's a lot of crazy stuff happening.
00:31:46.000 Hold on a little bit.
00:31:47.000 Explain the difference.
00:31:48.000 What does 15-minute, what does 5-minute mean?
00:31:50.000 Well, this is the restrictions that they're setting up, pretty much essentially prisons, pretty much essentially lockdowns.
00:31:57.000 They did this in China and they tested it during COVID.
00:32:00.000 They locked down entire communities because they said they had a positive case.
00:32:03.000 But what a lot of people are saying that this was most likely a larger test to see how these 15-minute cities would work, because they would lock down entire neighborhoods and say, You can't move between this street and this street.
00:32:14.000 You can't go between here and here.
00:32:16.000 They tested it out in Australia, where people were fined and arrested for going from one city block to another city block because they were going through an imaginary line in jurisdiction.
00:32:25.000 So a lot of people say that there was a dry testing of this larger concept, which is essentially prison grids, internal systems where you can't move around anywhere but 15 minutes within it, and you get everything you ever wanted.
00:32:36.000 Right, the idea is that within 15 minutes of your home, you can get access to anything you would need.
00:32:41.000 Correct.
00:32:42.000 Yeah, hospitals, supermarkets, entertainment, movies, whatever it is, it's all confined here.
00:32:47.000 And you better not go around.
00:32:49.000 This is again to de-incentivize people from driving, de-incentivize people from having personal liberty,
00:32:53.000 track, trace, and database of society, of total control grid of information,
00:32:58.000 knowing every little thing about you, and then that information being used against you
00:33:02.000 in every way.
00:33:03.000 And if you're not a part of the 50 Cent Army, well, your ability to move around
00:33:07.000 is going to be limited a lot more because the checkpoints are already there.
00:33:10.000 They're digital, and they have cameras and facial recognition, and they're setting it up.
00:33:16.000 It's pretty much the most care.
00:33:17.000 I just wanna, I wanna say something.
00:33:18.000 You know, there's a viral video we talked about.
00:33:20.000 It's really funny.
00:33:21.000 It's where a guy says he wants to show you a perfect example of a 15 minute scene, how it works.
00:33:26.000 And then he walks up this little shed and he says, let me show you a 50 cent city when he opens it up.
00:33:31.000 Chickens.
00:33:32.000 And he goes, welcome to a 15-minute city.
00:33:34.000 He walks inside and says, here, the citizens have everything they could hope for.
00:33:38.000 They can go outside if they want.
00:33:40.000 Look, the door's open.
00:33:41.000 They don't want to leave.
00:33:42.000 It's safe in here.
00:33:43.000 There's no danger.
00:33:44.000 They have food.
00:33:45.000 They have water.
00:33:46.000 It's always there for them.
00:33:47.000 And then he walks into another room, and he goes, see this?
00:33:50.000 He starts grabbing eggs.
00:33:51.000 And he goes, they make these, and I take them from them.
00:33:54.000 But that's OK.
00:33:55.000 They're happy, because they're safe in their 15-minute city.
00:33:58.000 So he made a really great point.
00:33:59.000 And I want to make a point on top of that.
00:34:01.000 Could you imagine if you went to your chicken coop one day to collect chickens and a bunch of roosters pulled out guns and pointed them at you and said, you're not taking the eggs this time?
00:34:10.000 You'd be like, what?
00:34:12.000 How did these chickens have guns?
00:34:14.000 And then you go back, we got to take these guns away.
00:34:17.000 Well, we can't.
00:34:18.000 The chickens have guns.
00:34:20.000 And you're like, I don't know how they're even using them.
00:34:22.000 And then you got to figure out how do you disarm the chickens?
00:34:25.000 If they're armed, and they can fight back, you can't take from them, you can't imprison, you can't control them, they're on equal footing with you.
00:34:31.000 Obviously, chickens don't have guns.
00:34:33.000 My point is, if someone is trying to control you, and you have the means of telling them no, they cannot implement their authoritarian, fascist takeovers, and that's what we've seen throughout history when despots and fascists try to take over, guns make it very difficult.
00:34:46.000 That was one of the impetuses of the Second Amendment.
00:34:49.000 You know, and that's why they're doing this in China.
00:34:51.000 They did this in China.
00:34:52.000 They're doing this in the United Kingdom.
00:34:54.000 They did this in Australia during COVID.
00:34:55.000 Right now in Oxford, specifically in the United Kingdom, they're doing the trial run for the 15-minute cities, blocking people from traveling one road to another, all in the name of sustainability.
00:35:06.000 It's not sustainability.
00:35:07.000 It's enslavement.
00:35:07.000 No, I was just going to say, if you look at the World Economic Forum, it's not framed as a conspiracy, it's not framed to manipulate, to control, it's framed to be sustainable.
00:35:18.000 That's the approach that they use.
00:35:21.000 It's sustainable to have farm animals, to treat human beings like they have no dignity, no individual human spirit, and they can't live on their own and they need to be dictated.
00:35:28.000 What they can and cannot do or believe in or think, and that's essentially this larger notion, this larger elitism of individuals thinking that they have the right.
00:35:36.000 And they're doing this in so many instances that are so bastardized, that are so disgusting.
00:35:41.000 They think that they have the right to control what goes into your head, to control what you see, what you hear.
00:35:47.000 And they're doing this with big tech social media already in so many different ways, trying to engineer people to be perfect lemmings, to be perfect slaves, to acquiesce, to bow down, and to take it.
00:35:57.000 And I think people who are paying attention had enough of this bullcrap.
00:36:00.000 All right, you guys ready for this one?
00:36:01.000 Let's jump to this next story from the Postmillennial.
00:36:04.000 FBI flags slang terms Chad, based, red-pilled, it's over, to target racially motivated violent extremists.
00:36:15.000 The FBI's domestic terrorism reference guide on involuntary celibate violent extremists offer a threat overview for incels that aims to identify them by the slang they use.
00:36:27.000 So, uh, you thought Biden's 50 cent army was bad.
00:36:30.000 Yeah.
00:36:31.000 You're gonna have a whole bunch of crackpot woke cultists getting paid to promote Democrats.
00:36:37.000 Now, the feds are gonna be targeting you if you say things like, based.
00:36:42.000 And that's one of the reasons your former guest, Vivek, wants to eliminate the FBI.
00:36:46.000 One of the many reasons he wants to eliminate the FBI.
00:36:48.000 I love that.
00:36:49.000 I'm like, wow, this guy's pretty hardcore.
00:36:52.000 Ron Paul was saying that many decades ago.
00:36:54.000 He was saying that 30 years ago.
00:36:55.000 And I think there was individuals even before that, at the very beginning of the FBI, they have a Horrible record, especially with J. Edgar Hoover, especially with what they did to JFK, what they did to MLK, what they did to Malcolm X, what they did with MKUltra.
00:37:09.000 Holy cow!
00:37:09.000 If you look at this agency, there needs to be some oversight.
00:37:12.000 There needs to be some accountability.
00:37:14.000 There was just A new disclosure happening with the Proud Boys trial with January 6th, and people just found out that out of the 13 Proud Boys, eight of them were paid FBI informants.
00:37:26.000 Five of them are being prosecuted.
00:37:29.000 So that's more federal agents that took part of January 6th, a part of the Proud Boys, than the actual Proud Boys, which is crazy.
00:37:37.000 Those are insane numbers.
00:37:38.000 And then they're hiding and making sure that they're spying on the defense.
00:37:42.000 I mean, what's happening with the January 6 case is just a notion of what's going to be happening to everyone unless we address this and stop it.
00:37:48.000 So this is a very postmodern notion.
00:37:51.000 The idea that That power is everywhere, and you can use words, and they've done so incredibly, they've been incredibly effective at this.
00:37:59.000 You can use words to change the culture.
00:38:02.000 You can use words, and, I mean, think about that.
00:38:06.000 Red pill, I mean, these are not the N word.
00:38:09.000 They're not even remotely close to it.
00:38:10.000 These are words that indicate a condition.
00:38:14.000 Oh, you know, from the Matrix.
00:38:15.000 But it's not that, it's a reference to being part of a subculture.
00:38:19.000 And if the FBI thinks Based?
00:38:22.000 Or Chad?
00:38:24.000 Make you a racially motivated extremist?
00:38:25.000 The left uses based, too, though.
00:38:27.000 Exactly.
00:38:27.000 There are seven words all together.
00:38:29.000 It was based, red-pilled, blue-pilled, normie, Chad, LARPing, and incel.
00:38:34.000 If that's the criteria that you're using, you're going to be going LARP to everyone.
00:38:39.000 LARPing is a leftist.
00:38:43.000 But leftists use LARPing substantially more than people on the right do to claim that the right is live-action role-playing their fantasies.
00:38:52.000 This is a big problem that with the FBI is having as they're downstream from social media from the technology and all this data can be spoofed.
00:38:59.000 Like if Facebook wants to tell the FBI that you said something you didn't say they can and they can make it look like you did in the software.
00:39:05.000 Even the idea that we're living in the United States and then this is an issue at all is insane.
00:39:10.000 That is another reason why there's a legitimacy crisis.
00:39:12.000 Nobody trusts the institutions anymore.
00:39:14.000 It's because it's a global community now.
00:39:16.000 Like, we live in the U.S., but this global governance, it's here now.
00:39:19.000 We can say, like, it's coming, but it's here now.
00:39:21.000 Like, it's the, it is one unigovernment that's trying to alter the shape of the world.
00:39:25.000 But the corruption is so in your face, though.
00:39:27.000 When you have a moment in time in history where you say, Chad, You're going to be investigated by the FBI, right?
00:39:33.000 If you say it online, you say incel, you say red pill, you go to Epstein Island.
00:39:37.000 You're on the client list of Mr. Epstein.
00:39:39.000 FBI's like, yeah, you're clear.
00:39:40.000 You're good.
00:39:41.000 Don't worry about that.
00:39:42.000 How in the world are we allowing the Federal Bureau of Investigations?
00:39:46.000 By the way, the DOJ, FBI had an investigation on Epstein.
00:39:49.000 They never released the findings of that investigation.
00:39:51.000 They have the client list.
00:39:53.000 They have the videos.
00:39:53.000 They have the pictures of very powerful people doing some very horrible things.
00:39:56.000 Well, why wasn't anyone punished?
00:39:58.000 I mean, if You know, if you were on some dude's island and then you were asked to investigate yourself, you wouldn't release the list either.
00:40:04.000 That's true.
00:40:05.000 Good point.
00:40:05.000 Very good point there.
00:40:08.000 I imagine that's likely the circumstance.
00:40:11.000 I think, uh, why are we letting people do this?
00:40:14.000 Because people haven't stepped up to build systems that are resilient to this kind of behavior.
00:40:18.000 So it's going to happen.
00:40:18.000 We need systems that are encrypted, where we control locally our own system.
00:40:23.000 We need to be able to govern ourselves locally.
00:40:26.000 Um, as best as possible.
00:40:28.000 When you start building systems like that that are not untrackable but challenging to track, I think you're in a better place where the code's readable.
00:40:34.000 But the issue is the FBI, the DOJ, have gone after anybody who makes encrypted communications.
00:40:40.000 I feel like they're afraid that there's gonna be some sort of revolt or like revolution in the states.
00:40:45.000 I hadn't heard that.
00:40:45.000 I encrypt everything.
00:40:47.000 Oh yeah, I mean, aside from the fact that they've probably cracked a whole lot of the encryption.
00:40:51.000 Telegram, the Russian government has cracked Telegram.
00:40:55.000 They're investing people who complain about the Russian war.
00:40:58.000 I'll just ask you something very simple.
00:40:59.000 Who do you think has more cyber resources?
00:41:02.000 A private company or a nation state?
00:41:05.000 It obviously depends on the nation state, right?
00:41:07.000 I think Chiquita Bananas had a good hold over, what was it?
00:41:10.000 What countries was it?
00:41:12.000 Nicaragua?
00:41:12.000 I don't think it was.
00:41:13.000 Costa Rica?
00:41:14.000 There were two countries.
00:41:16.000 But anyway, I digress.
00:41:17.000 We're talking about Russia and their military capabilities in terms of cyber warfare.
00:41:21.000 Pretty sure they're going to have more resources than Telegram.
00:41:24.000 And they're going to have a capability to either crack Telegram encryption or, simply put, go to the people who run it and then say, here's what you're going to do or else.
00:41:34.000 You know, maybe the CEO of some company goes out for his, uh, pick up his newspaper in front of his apartment when a black van pulls up.
00:41:40.000 Guys throw a bag over his head, throw him in.
00:41:42.000 They drive off.
00:41:43.000 Nobody knows it happened.
00:41:44.000 And then he wakes up in a dark room and they're like, you're going to give us a backdoor to all of your encryption or else.
00:41:49.000 And they go, whatever you say.
00:41:50.000 And that's exactly what happened to Lavabit and the founder of Lavabit.
00:41:54.000 I interviewed the guy.
00:41:55.000 It was Edward Snowden's email service that he was using.
00:41:58.000 The feds came to him and said, give us everything on Edward Snowden.
00:42:01.000 Or else.
00:42:01.000 And then he, I believe he printed out 100 pages of the password in very small font and then gave it to the FBI, legally And they had to manually put in the long password to the encryption.
00:42:14.000 And then when they were doing that, I think he deleted the service.
00:42:16.000 Shut the whole company down.
00:42:17.000 Shut the whole company down.
00:42:18.000 And then this is why, you know, this is like in China.
00:42:21.000 This is almost exactly what's happening in China.
00:42:24.000 But here we just have a different face on it.
00:42:25.000 And the more we don't understand this reality, the more we kind of hinder ourselves.
00:42:29.000 Because when we look at Facebook, when we look at YouTube, when we look at, you know, as we found out from Twitter, we're not looking at private entities and private corporations.
00:42:36.000 We're looking at government's hand inside of the business directing not only who gets to say whatever they want to say, but who gets to see it, why they get to see it, who gets promoted, who gets demoted.
00:42:47.000 That right there is the power to control people's minds, and that's the power that they're wielding every single day, and people need to realize it more than ever.
00:42:54.000 This story that came out a few weeks ago, or like a week and a half ago, that Elon Musk said they were qualifiers coded into the Twitter system that would identify you as a Democrat, a Republican, or a VIP.
00:43:05.000 Or Elon Musk, which was funny and weird that he had his own special category of like, this is Elon, he's tweeting.
00:43:11.000 But they would label you a Democrat or Republican.
00:43:13.000 Like they had that in their system and people were calling that out.
00:43:17.000 How much do you want to bet that if you had the Democrat tag, that was okay that you said that thing about those kids?
00:43:23.000 When that guy posts the picture of the wood chipper with blood coming out during the Covington kid incident, I wonder if he had one of those tags on it and they're like, he's okay.
00:43:31.000 So the thing that's been running through my mind through this conversation is let's say that somebody who doesn't listen to you tunes in, Tim.
00:43:41.000 It's so hard to break through to people when you show stories like this or when you talk about scores or when you talk about... because they'll just think you're a conspiracy theory, lab leak conspiracy.
00:43:54.000 So how do we, the thing that I've been thinking about for quite a while, is how do we reach these people?
00:43:59.000 Like, what is it that we can say?
00:44:01.000 Or how do we convey this in a way that makes them think, I'll put myself in this case, we are not conspiracy theorists when we talk about this.
00:44:11.000 Oh, this is funny.
00:44:12.000 I mean, look, we, we, I use NewsGuard certification for all the sources we use.
00:44:17.000 I know NewsGuard's not good.
00:44:18.000 I know they're biased.
00:44:19.000 They attacked us and then put a bunch of lies about TimCast.com.
00:44:23.000 We still ended up getting a near perfect score.
00:44:26.000 They gave us a strike out of like 10.
00:44:29.000 We got one because we quoted Donald Trump.
00:44:32.000 That's it.
00:44:33.000 That's all we did.
00:44:33.000 We ran a story, Donald Trump said this, and they went, oh, nope, nope, that's fake news.
00:44:36.000 And I'm like, all we're doing is quoting a guy.
00:44:38.000 But I digress.
00:44:39.000 The point is, when I try and talk to people who are, as I describe, uninitiated, I'm not gonna come out right away and start talking about Burisma and Ukraine and all this really crazy stuff.
00:44:48.000 I'll take it really, really light and be like, Yeah, that Trump impeachment, wasn't that, that was, um, what was it?
00:44:54.000 There's that video of Joe Biden saying that he was threatening to withhold congressionally approved aid, which was illegal or something like that.
00:45:00.000 Real light approach.
00:45:01.000 And then if they say, what?
00:45:02.000 I don't know anything about that.
00:45:03.000 I'll be like, oh, here, I'll show you.
00:45:05.000 And then you just pull up a NPR or some other news source and be like, I don't know, read it.
00:45:08.000 You tell me what you think.
00:45:09.000 I usually take that approach.
00:45:11.000 I'll say, here's the story.
00:45:12.000 What do you think?
00:45:13.000 Yeah.
00:45:14.000 I have the opposite approach.
00:45:15.000 I just go right over the head, satanic child traffickers, private islands, entities, demons.
00:45:21.000 Let's just go there.
00:45:22.000 I think different people have different ways of approaching this information, but I think, first of all, the conversation needs to be started.
00:45:29.000 We're doing our best here.
00:45:30.000 I totally agree.
00:45:30.000 The conversation starts with the people who are listening, the people who are taking notes, the people who are studying, because you guys need to be informed more than ever.
00:45:37.000 And the game's rigged on social media, the game's rigged here, the game's rigged everywhere online, but it's not in real life.
00:45:45.000 This is why, again, I'm not just trying to be a shill here, but I've been a big proponent of t-shirts, I've been a big proponent of people talking to their family members, their friends, and I think that is more imperative than ever of individuals doing their homework, doing their research, and whether it's having a light-hearted kind of approach of empathy, whether it's hitting them over the head with just the bigger truths going down the rabbit hole, who knows what will work, but I think Being brave and being able to have those conversations is the first step that we need to have.
00:46:11.000 But the other thing too is, outside of all of that, there's a real simple way to win.
00:46:16.000 It's called voting with your dollars.
00:46:17.000 So shout out to our friends at Public Square.
00:46:20.000 Download the Public Square app.
00:46:22.000 They've sponsored the show before, but this is a freebie because I really do believe in what they're doing.
00:46:26.000 It's an app that shows you businesses that have taken a pledge to support American values.
00:46:30.000 So you can make sure you're giving your money to companies that believe in the good.
00:46:35.000 And then, uh, don't buy anything from Anheuser-Busch.
00:46:38.000 There you go.
00:46:39.000 And don't donate to your alma mater, please.
00:46:41.000 Yeah, yeah, uh, Will Chamberlain liked to say, I don't know if he still does, I'm assuming he does, seize the endowments.
00:46:45.000 Yep, and that's one of the things that we're doing at my non-profit, National Progress Alliance, do not, it should be the easiest ask in the world, don't donate.
00:46:53.000 By the way, just, I personally don't like the hardcore approach, so I'll ask people, I'll say something like, which is true, I don't know if you guys were, I don't want to go down this rabbit hole, but I personally was duped by the Hunter Biden laptop story.
00:47:06.000 I thought it was a plant.
00:47:07.000 I'm betting you guys didn't.
00:47:10.000 Jury's out for me.
00:47:11.000 I wait for facts and proof, and I want proof three times before I start to accept something.
00:47:16.000 Luke probably went completely in the other direction.
00:47:18.000 Yeah.
00:47:18.000 And then I was probably in the middle, like, let's figure out where this came from.
00:47:21.000 Yeah, so the way that I try to reach people is I'll say, you know, I, like, I'll put it on myself, I fell for that.
00:47:26.000 What were you thinking about that?
00:47:28.000 Oh, or I, and then I'll say, well, which is also true, I didn't really know whether or not it was a lab leak hypothesis, but the lab leak hypothesis is true, but all these people said, people who were in authority, people who were supposed to be trusted,
00:47:42.000 I believe them and I was wrong. In other words, I was wrong, maybe they were wrong, I was wrong, but what about this?
00:47:48.000 What about this?
00:47:49.000 So if you can undermine their confidence in things in the past,
00:47:52.000 they can make their own conclusions about whether or not...
00:47:55.000 I think the best thing you do is be genuine and if you believe something,
00:47:59.000 say you actually believe in those things.
00:48:00.000 And I think living your life as an example, being like, hey, I don't do this.
00:48:04.000 I don't support this.
00:48:05.000 I don't want my money going towards this.
00:48:06.000 I don't want to spend my time investing in these really bad people that are connected to these really awful people.
00:48:11.000 With these chemicals and this thing and that thing.
00:48:13.000 And again, it's all personable.
00:48:15.000 So if you're seeing someone who's very kind of sensitive, who's very kind of introvert, you obviously don't do my method.
00:48:21.000 You obviously use Tim's method.
00:48:23.000 But if you see someone who is open minded, someone eager, obviously you just beat them over the head straight down the rabbit hole.
00:48:30.000 I like to steel man the opposition.
00:48:32.000 I kind of take that role on this show.
00:48:34.000 A lot of times people in the audience think I'm like an idiot, but like I will act as if we are wrong and I will give you Yeah, it's so important to do that, and it's also so unfortunate that you have to.
00:48:45.000 It's so unfortunate that woke people will not have conversations with you.
00:48:48.000 They will not engage in discussion, dialogue, debate.
00:48:50.000 The smart ones understand why they can't.
00:48:51.000 Well, a hundred percent.
00:48:54.000 It's such a bankrupt ideology, and the people forwarding it have profoundly mediocre minds.
00:49:00.000 Well, it's not that.
00:49:01.000 It's that there is no ideology.
00:49:03.000 So what do you mean?
00:49:04.000 So I mean this is my consistent position on what wokeness is.
00:49:07.000 Wokeness is the modern left liberal culture formulated by social media algorithms.
00:49:13.000 It is characterized by cult-like adherence to liberal social orthodoxy.
00:49:17.000 That's all it is.
00:49:18.000 And the example of this is Ukraine.
00:49:21.000 Why woke people support a war in Ukraine makes literally no sense.
00:49:23.000 It doesn't follow any academic theories.
00:49:25.000 Okay, so let's, I think we have to unpack it.
00:49:29.000 Have you, have you read Tim Urban, you know, Tim Urban, The Wait But Why?
00:49:33.000 What's Our Problem?
00:49:34.000 No, but you were mentioning it before.
00:49:35.000 This is like one of the best books I've ever read.
00:49:37.000 But in chapter five about social justice, he has a chart.
00:49:40.000 I think we sent you that chart.
00:49:43.000 If you could pull it up, great.
00:49:44.000 But he talks about, I do think it's a coherent ideology.
00:49:49.000 Yeah.
00:49:50.000 There it is, from Tim Urban's Wait But Why blogger, What's Our Problem?
00:49:57.000 It's fascinating how it originates in Marxism, goes down the tree to critical theory.
00:50:04.000 This is the best book since the 2020 Cynical Theories by Helen Pluckrose and James Lindsay on this topic, in my opinion.
00:50:11.000 And you have he talks about social justice fundamentalism at one end and liberal social justice I think you and I I think everybody in this room frankly falls on the side of liberal social justice Not equality of outcome equality of opportunity not using race to divide But being what what is now microaggression colorblind to people we make those judgments I do think that there's a fundamental core tenant or there are fundamental core tenants of the ideology and Racism is everywhere.
00:50:38.000 It's baked into the system.
00:50:40.000 There's whiteness.
00:50:42.000 Your privilege seeks to preserve itself.
00:50:44.000 I mean, et cetera, et cetera.
00:50:45.000 And that is a component of what wokeness is, but it's also just... When I try to envision wokeness as an object, it is like a Jackson Pollock painting.
00:50:56.000 But that's because wokeness isn't an ideology.
00:50:58.000 It is not this thing.
00:51:00.000 These things They frustrate me when I see, and I think this chart is probably correct, you know, going from postmodernism down and critical race theory and intersectionality, but it applies an academic worldview into the modern problem, which has already been shown by numerous data points, including, there was the Zach Roberts, LexisNexis Twitter threat from a few years ago, that all around the same time, everywhere in the world, the same thing happened, where there was a massive spike in the same concepts, ideas, words,
00:51:30.000 And what likely happened is, and I'll put it this way, the reason why I don't believe we're dealing with an ideology of the left, we're dealing with algorithmic corruption, is that Jack Dorsey was not woke until he started eating the own refuse of his own social media platform.
00:51:45.000 Okay, so can we spend some time on this?
00:51:48.000 Because I think this is really important.
00:51:50.000 Okay, so this is the frame that I suggest thinking about this.
00:51:54.000 At the top you have theory.
00:51:57.000 Below that, you have institutions, and downstream from that, you have individuals' beliefs.
00:52:02.000 So, 2017, we had this, almost this exact conversation.
00:52:05.000 Yeah.
00:52:05.000 A very heated debate.
00:52:06.000 I would call it lively.
00:52:08.000 I thought it was great.
00:52:09.000 No, I thought it was great, too.
00:52:10.000 I thought it was fantastic with James and Helen in the room.
00:52:14.000 So, at the level of theory, you have what we tried to do with Sokol Squared, at the Grievance Studies hoax.
00:52:20.000 We tried to undermine the theory.
00:52:21.000 We tried to show that it's bogus.
00:52:22.000 We tried to attack it.
00:52:24.000 What Reid and I are going around the world doing, street epistemology, is to show that it goes from the theory to the institutions, the institutions to the people.
00:52:32.000 The institutions, the academies, K through 12, and we can talk about that, more than happy to talk about that, are academic institutions.
00:52:41.000 You get promoted and tenure based upon the papers.
00:52:45.000 you write usually seven papers in seven years unless you write something that's morally fashionable.
00:52:50.000 It's very difficult to get published and we're going to do the University of San Francisco story,
00:52:55.000 I hope, because I think that this encapsulates it perfectly.
00:52:58.000 Educational administrators themselves participate in the ideology. So there's a
00:53:03.000 structure and so when you say algorithms, I want to say two things that are important about this.
00:53:10.000 Woke people occupy, so they get their degree five, six, seven years later, they go out, they become administrators, they become managers, etc.
00:53:19.000 They bring the madness that they took with them, that they think it's true, based upon what they perceive to be legitimate bodies of literature.
00:53:26.000 Microaggressions, trigger warnings, safe spaces, belonging, equity, all this nonsense.
00:53:30.000 They take that with them and they institutionalize that.
00:53:33.000 We can cover it if you want, but it's a reverse Trojan horse Martin Bailey.
00:53:38.000 We can cover that.
00:53:39.000 Now, to get to your point about the algorithms, yes, you are correct, but the algorithms themselves wouldn't even make sense unless you understood that model.
00:53:48.000 Like, this doesn't come ex nihilo, like manna from heaven, right?
00:53:51.000 This comes at a very specific place.
00:53:54.000 I'm starting to think that institutions include social media networks now.
00:53:57.000 Hold on.
00:53:59.000 One of the first big political trends was probably loose change 9-11, which was certainly not within the confines of this academia, and the Ron Paul revolution, which was absolutely not in the confines of this academia.
00:54:11.000 What happens is you see, what I believe it is, that humans have tendencies Uh, rage makes someone more likely to share something than any other emotion, anger.
00:54:22.000 And I always shout out CGP's Grey, uh, CGP Grey's video, this video will make you angry.
00:54:27.000 And he explains this and he was like, it's really great, he's like, this group is talking about the other group, the other group's talking about this group, they're not actually talking to each other, they're talking about each other.
00:54:35.000 And he's like, trust me, I am not talking about your group.
00:54:38.000 So what I see here is, yes, the ideologies of the left do exist, and they do have a degree of prominence among the modern left today.
00:54:47.000 But there's also, I don't know, how many ideologies exist?
00:54:50.000 A hundred thousand?
00:54:52.000 Countless.
00:54:52.000 Millions?
00:54:53.000 Countless?
00:54:54.000 So why is it that this is the one?
00:54:56.000 I think the issue is... That's a genealogical question that I could answer.
00:55:00.000 But I don't think... The chart actually answers that.
00:55:04.000 Right, but this chart then omits every other ideology and the path that's taken and its degree of prominence in modern culture.
00:55:09.000 Right, but every other ideology isn't the dominant moral orthodoxy now.
00:55:13.000 This wokeism is the dominant moral orthodoxy.
00:55:15.000 We see it everywhere.
00:55:16.000 And it's because of social media algorithms, not because of universities.
00:55:20.000 Okay, so that's the question.
00:55:22.000 Why are the social media algorithms catering to the moral orthodoxy?
00:55:26.000 So, very simply put, advertisers fund moral orthodoxy.
00:55:31.000 Before wokeness took over how did this begin first?
00:55:34.000 We had in what was it?
00:55:35.000 What year was it Luke 2008 2009 the Ron Paul revolution right around there?
00:55:38.000 Yep, you're starting to see the emergence of moral of Moralizing through social media with people listening to Ron Paul's message and it resonating the point where they were going around slapping stickers everywhere Anti-intervention sound currency a lot of really really awesome things that I completely agree with today but eventually These social media platforms started to adopt algorithms that would show you more of what you engaged with.
00:56:01.000 Two things happen.
00:56:03.000 People are more likely to share something that makes them angry, and advertisers are less likely to advertise on something that's morally objectionable.
00:56:10.000 In modern culture at the time, 2008, we find racism to be objectionable.
00:56:14.000 So a big advertiser says, I don't want my product on racist content.
00:56:19.000 Okay, so you just smuggled something in there.
00:56:22.000 We find it to be objectionable because This is morally fashionable.
00:56:29.000 Like, this is what's morally fashionable today.
00:56:31.000 But this was pre-wokeness.
00:56:34.000 Yeah, I guess the main wokeness really kicked in 2012, 2010, so 2008 I would give its pre-wokeness.
00:56:42.000 But you're talking, in that case, you're talking about a kind of tribalism, right?
00:56:46.000 You're not talking about the algorithms themselves.
00:56:50.000 Do they feed outrage?
00:56:52.000 For sure.
00:56:52.000 Yes.
00:56:53.000 Yeah, for sure.
00:56:54.000 Every single thing you've said is true.
00:56:57.000 And it's, I don't mean to use a big word but it's the only word to describe, it's covariant.
00:57:02.000 Like both of those variables act upon each other at the same time to enforce the dominant moral orthodoxy even more than it was in the first place.
00:57:10.000 So what's happening is, this is why I say that the problem is not the ideology.
00:57:14.000 The ideology has existed for a long time.
00:57:16.000 The ideology is a component of the problem.
00:57:18.000 The problem is we are in a feedback loop of algorithms quadrupling upon quadrupling our problem.
00:57:24.000 It's exponential growth.
00:57:25.000 So, the example of this I give is 2008, it starts with the viral Ron Paul revolution, which is a good thing.
00:57:31.000 Then you start to see, my favorite example is Mike.com, which started off as a website exploiting this, and they were producing libertarian content.
00:57:39.000 However, within that libertarian content was anti-police brutality content, because libertarians didn't like that either.
00:57:45.000 There was also Second Amendment audits that were going viral, where people were like, the cops stopped me, and you get these videos of people challenging police officers.
00:57:52.000 Then, people start adding in a racial component.
00:57:55.000 Why?
00:57:56.000 Because they start putting up videos of police brutality against black people.
00:57:59.000 These videos start getting plastered all over Facebook because it makes money.
00:58:05.000 At one point, a website that was dedicated to nothing but police brutality videos was the 400th most viewed website in the world.
00:58:15.000 I'm not going to say what website that was.
00:58:17.000 And they were paying their writers an exorbitant amount of money and all they did was post videos of police brutality.
00:58:22.000 This starts fracturing the minds of 10-year-olds in 2010 who are now 23-year-olds who are voting in this election whose entire worldview has been built upon a machine that started with Libertarian So here's what happens.
00:58:38.000 When libertarians are sharing anti-police content, and they're getting a lot of views, because people don't like injustice.
00:58:43.000 It is the epitome of injustice when a cop violates our rights.
00:58:46.000 That's supposed to be upholding the law.
00:58:49.000 Then you add in a racial component, and now you have an exponential growth.
00:58:52.000 The people who hate racism, and the people who hate police brutality, everyone's seeing it and everyone's sharing it.
00:58:58.000 Then, Mike.com shifts its business model and says, this gets more views, let's do more of this instead.
00:59:06.000 The company then slowly, rather exponentially, rapidly shifts into a woke social justice company.
00:59:12.000 But here's the thing.
00:59:13.000 When this happened, we didn't have the word woke.
00:59:14.000 I mean sort of, it was here and there.
00:59:16.000 What was the word?
00:59:18.000 Intersectional feminist.
00:59:19.000 Which then gave rise to social justice and social justice warriors.
00:59:23.000 I mean, at first it was feminism, then it was intersectional feminism, then the SJW, then we ended up with the complaints about critical race theory, then people pointed out actually critical gender theories in there too, and now it's woke.
00:59:36.000 What we're seeing is a feedback loop of social media algorithms funneling refuse back into the mouths of people in what I would only describe as a human centipede of ideology.
00:59:44.000 Just to clarify, would you say that woke is the way that the ideology is funneled?
00:59:51.000 Because what what so when people say what does woke even mean okay well when when the average person describes woke they may mean a bunch of different things but the one unifying factor is they're referring to the modern left liberal orthodoxy that was created by social media algorithms and and the example I give is Ukraine War is the easiest way to understand this.
01:00:14.000 Why is it that Hasan Piker will be like, here are these things that I believe, trans rights, and then also, I also am for war in Ukraine, and you're like, what do these things have to do with each other, and why is it that this individual has no principles?
01:00:28.000 It's just...
01:00:29.000 He just follows the orthodoxy.
01:00:31.000 You think there's a woke on the other side of the coin?
01:00:34.000 Like for people that are obsessed with Donald Trump?
01:00:36.000 Like a right-wing version?
01:00:37.000 But it's so small and there's no institutional power behind it, I find it to be negligible.
01:00:40.000 But it could become a problem if it were let, if it got control of the machine.
01:00:44.000 I think it was white nationalism.
01:00:46.000 Yeah, I think that the anti-woke people are starting to use some of the tools of the woke now.
01:00:51.000 And I want to come back to what you said because I think it's important.
01:00:54.000 This is both a very complicated problem and a non-complicated problem.
01:00:57.000 with whom they disagree. So I think that they're adopting the tools to buttress
01:01:01.000 the ideology. This is both a very complicated problem and a non-complicated
01:01:08.000 problem. So I don't mean to muddy the waters too much, but I do want to say
01:01:12.000 this because it's been really bumming me out. So I read a piece in the
01:01:15.000 Epistemology of Democracy. It was about by Keith Stanovich.
01:01:20.000 It was about my side And my side bias, this will horrify you when you hear it, but my side bias is exactly what you think it would be.
01:01:27.000 My side is right, your side is wrong.
01:01:29.000 But this, the article or the chapter in the book argued that there's literally nothing you can do to overcome my side bias.
01:01:38.000 No training, no education, nothing.
01:01:40.000 But when you combine that That research chapter with what you just said, not only can you not do it, there's no amount of education that can get over my side bias, you're now talking the algorithm being held hostage to the algorithm.
01:01:57.000 Yes.
01:01:57.000 And the humans being held hostage by the humans.
01:02:00.000 So even if, right, so even if there weren't the case that you're being held hostage by the algorithm, you're still suffering, you know, this my side bias thing, no training, no education, etc.
01:02:10.000 You just can't, so what are you going to do?
01:02:11.000 So that is, I think, what is often missing in these conversations, what I just like to throw in there.
01:02:18.000 We know there's a problem.
01:02:19.000 We've been talking, you and I have been talking for years, we know this is a problem and I think what people are hungry for now is what are we going to do about it?
01:02:26.000 Psychedelics?
01:02:28.000 You said one thing, right?
01:02:30.000 We're going to stop donating to Alma Mater, we're going to stop buying, you know, we're going to vote with our dollars, if you will.
01:02:35.000 But I think what would be helpful, I think, to a lot of your viewers is ending every one of these sessions, these conversations, with what are you going to do about it?
01:02:43.000 Like, empowering them to do something.
01:02:45.000 And so, obviously, that's why I wanted to lead with the Anheuser-Busch story, because it seems like this is particularly effective.
01:02:50.000 Let me tell you, I've got these tweets that I love to tweet, where I say something like, Sterilize your children to prevent overpopulation.
01:02:58.000 It's a joke, right?
01:02:59.000 Or I'll say something about leftists will abort their children or sterilize their children.
01:03:05.000 And then I had one tweet where I said, and this is a good thing because it stops overpopulation and humans are causing climate change.
01:03:12.000 Not one time has any publication ever criticized me or written about those tweets.
01:03:19.000 Yet when I say don't buy beer, I get 500 articles saying far-right fascist, outraged, whiny loser, because you can tell what's working.
01:03:30.000 They don't want to bring up the fact that they're aborting their kids and sterilizing them, because that's bad for them in the long run.
01:03:37.000 If they highlight that, they have to point out the things we're promoting are detrimental and deleterious to our ideology, and then when I say something like, don't buy beer, they're like, quick, we gotta put a stop to that.
01:03:48.000 Not only is it affecting our potential sales, because it's sponsored by this company, But there's truth, the fact that if Budweiser sales do drop and they are forced to apologize, that will be a major shift in the culture war.
01:04:02.000 The moment a massive multinational corporation says, we are sorry for being woke, woke people will lose their minds, because that means... We're sorry, we made a genuine mistake, we wish we hadn't done it, we'd like to get you back.
01:04:17.000 And real quick, Netflix never apologized for Cuties, and they still have Big Mouth.
01:04:24.000 But when they started getting rid of some of the woke shows and stuff they had, and they talked about it, they saw subscriber increase.
01:04:31.000 So they're probably learning an important lesson, but the apology is everything.
01:04:35.000 Real quick, just one last point.
01:04:38.000 Antifa uses violence, and the far left uses violence, and in the summer of love they got away with it, because they have law enforcement to a certain degree on their side, either because police are too feckless, or the DAs are bought and paid for.
01:04:50.000 So these big companies, as I've long stated, do not fear Dave Rubin, classical liberals showing up with pitchforks.
01:04:57.000 They do fear Antifa, who do it and get away with it.
01:05:00.000 If Budweiser is willing to make an apology, To all of us for sponsoring Dylan Mulvaney, that means they are saying the threat of physical force from Antifa is not as scary as us actually going out of business.
01:05:12.000 Yeah, and I want to add to that, and I want to just say to the people listening to this, one of the things that I've seen over and over again is people are very afraid to piss off the far left.
01:05:23.000 My comment to you is, they already hate you.
01:05:26.000 They already hate you because you're going to say something or not toe the line perfectly.
01:05:30.000 Do not be afraid of these people.
01:05:32.000 Do not be held hostage to them.
01:05:33.000 Be based.
01:05:35.000 Also, the cartels and Islam, don't be held hostage to violent extremists of any kind.
01:05:39.000 No offense people, not Islam at all, but I mean, you know, what's his name?
01:05:43.000 Richard Dawkins was afraid to even criticize or talk about it.
01:05:45.000 That's really sad.
01:05:47.000 Every religion should be criticized.
01:05:49.000 Shout out to Steven Crowder, potentially one of the most fearless guys, who's constantly making jokes and saying F you to all of these extremists for telling them to shut up.
01:05:56.000 You said earlier how these things could cause people to lose their minds.
01:06:00.000 You were saying this, my side bonus, I actually jokingly said psychedelics, but I think being able to clear your mind helps you see the danger of your side.
01:06:11.000 Recording yourself on video, uploading it, and watching yourself say it, and having no choice but to accept that's what you are, you'll see what you did wrong.
01:06:19.000 The psychedelics, you've got to do what's legal and what's safe for you, but being able to clear your mind, whether through fasting, I think helps you see the crap on your side.
01:06:27.000 Let's jump to this story from TimCast.com.
01:06:30.000 Female swimmer assaulted by trans activists while speaking on women's rights.
01:06:35.000 Quote, this is proof that women need sex protected spaces.
01:06:38.000 So you may have seen the story.
01:06:40.000 This is about Riley Gaines.
01:06:41.000 She is a woman's rights activist.
01:06:44.000 I guess you'd call that a feminist who has taken issue with with males competing in female sports.
01:06:48.000 Right.
01:06:49.000 She was going to be speaking at an event and a mob of far left extremists chased her into a room and then they pinned her down for about three hours.
01:06:57.000 Demanded ransom.
01:06:57.000 Yeah!
01:06:58.000 They said, pay us, pay us money.
01:07:00.000 That's extortion, I guess?
01:07:02.000 Look at what happened with Brett Weinstein and Evergreen when they were mocking and holding these administrators hostage.
01:07:07.000 It's not the first time.
01:07:08.000 Yeah, and if anybody doesn't know the Brett story, which you should, either watch the Jordan Peterson or the Mike Nena has wonderful coverage of that.
01:07:15.000 I actually, a long time ago, produced a miniature documentary at Evergreen with Brett, and so you can search for Brett Weinstein, Evergreen, Tim Cass or something.
01:07:25.000 I want you on the show, Brett, if you're listening.
01:07:28.000 So what was the latest development here with Riley Gaines?
01:07:31.000 So what I wanted to talk about is, is this a story?
01:07:33.000 I had no question about it, but this kind of stuff, this isn't extreme, but it happened to Charles Murray, I mean, it happens all the time.
01:07:40.000 Which really the story, to me, which really the story is what the university did.
01:07:46.000 They said that they were peaceful protests.
01:07:49.000 They offered them counseling.
01:07:51.000 They said transphobia has no place in the university.
01:07:54.000 The statement, if you could pull the statement up, it's an utterly remarkable, it's opposite land.
01:08:01.000 And this is what is happening in our universities.
01:08:03.000 And this is why It's so important to talk about this. It's again what I had
01:08:08.000 mentioned. This is the legitimacy crisis a crisis of legitimacy in institutions and this is
01:08:13.000 contributing to that.
01:08:14.000 It's the misuse of the word phobia. That's for sure.
01:08:16.000 Because saying that you don't want biological men to compete in physical sports with
01:08:20.000 biological females doesn't mean you're afraid of trans women's prisons. Yeah, it's not that you're
01:08:25.000 afraid of it.
01:08:26.000 I mean, I'm not afraid.
01:08:27.000 I just, I'm concerned, maybe.
01:08:29.000 And it's not like this person's saying anything radical or crazy.
01:08:32.000 She's just making logical, political points that are not out of the ordinary.
01:08:38.000 And to have a group of people chase her down to physically assault her.
01:08:41.000 I mean, I was chased by, you know, some of these mobs before.
01:08:43.000 I was assaulted by one of these mobs before.
01:08:46.000 It's not fun to deal with.
01:08:47.000 But the police officers here to cower and not do anything here, and the university to back the attackers here. It is a deranged statement and
01:08:56.000 anybody who reads that if you ever had any question whatsoever that the academies are ideologically
01:09:03.000 captured, read their statement. So the New York Post has some of it. They
01:09:06.000 say following the mayhem, Jamila Moore, vice president for student affairs and
01:09:11.000 enrollment management, emailed students thanking them for taking part in the event.
01:09:15.000 It took tremendous bravery to stand in the challenging space.
01:09:18.000 I am proud of the moments where we listened and asked insightful questions.
01:09:21.000 I am also proud of the moments when our students demonstrated the value of free speech and the right to protest peacefully.
01:09:28.000 Welcome to the modern era.
01:09:29.000 And you see what they did, free speech, right?
01:09:31.000 You see how they... They chased her into a room, demanded money, threatened her life... Alright.
01:09:37.000 And they framed it under free speech.
01:09:40.000 That's what they do.
01:09:42.000 Is someone playing something or something?
01:09:45.000 Yeah, I think someone's downstairs on the phone.
01:09:47.000 Can you tell them to get off their phone if that's okay?
01:09:49.000 Yeah, I was actually just communicating with Surge.
01:09:52.000 I missed the top part of the quote that they said, but they were happy that the protests were peaceful.
01:09:58.000 They weren't peaceful.
01:09:59.000 But look at what happened with the Nashville Six.
01:10:02.000 Almost immediately, the Democrats This is a huge question.
01:10:06.000 where they start saying the Nashville 3, they shift the entire narrative to these
01:10:08.000 Democrats who get ousted for insurrection at their state capitol.
01:10:11.000 Insurrection.
01:10:12.000 You think that all the institutions, all the colleges, I mean obviously not all of them,
01:10:16.000 but what percentage do you think?
01:10:17.000 This is a huge question.
01:10:19.000 So I don't know if you've been following the work by Chris Ruffo at New College in Florida
01:10:25.000 So he is working hard to dismantle diversity bureaucracies and speaking of which we are in Austin and One of the things that I'm doing is I'm a founding faculty at the University of Austin and there's be none of this bullshit At the University of Austin I can assure you there's none of this bullshit.
01:10:41.000 How's that going?
01:10:42.000 Is it a is it is it?
01:10:43.000 phenomenal phenomenal well over a hundred million dollars who got whoa and Yeah.
01:10:48.000 Is it a new university?
01:10:49.000 Yep.
01:10:49.000 It's a free speech university.
01:10:51.000 It's what a university should be.
01:10:53.000 It's what a traditional university kind of was.
01:10:56.000 How old is it now?
01:10:57.000 Well, it's the first class.
01:10:58.000 We have a summer program, the Forbidden Classes program, and it'll be online in 2024.
01:11:03.000 We've got to get a high-speed rail in this city now.
01:11:05.000 Is it online only?
01:11:06.000 No, it's on-ground only.
01:11:11.000 On ground only.
01:11:11.000 Yeah, and they have classes this summer where I'll be teaching the forbidden classes, the Kathleen Stock, with gender, Mark Andreessen, Neil Ferguson, Ayan Hirsi Ali.
01:11:22.000 How big is it?
01:11:22.000 How many square feet?
01:11:23.000 What do you got?
01:11:24.000 No, it's just, it's literally being constructed now.
01:11:26.000 Wow.
01:11:27.000 It's in Austin.
01:11:28.000 But so, you ask a great question, what percentage of universities So I think a question is what percentage of the people believe are caught in the orbit of the ideology?
01:11:38.000 We don't really know the answer to that question because they've been so successful at creating a culture of fear that people won't honestly admit if they believe these truly deranged propositions.
01:11:50.000 So we don't really know, but we know from whether or not there are offices of diversity, equity, inclusion, which are virtually every university, if not everyone, maybe not Hillsdale, Liberty, the conservative, We know that there are bias response teams at over 150 universities, excuse me, 250 universities where you can, and that's from Julian Melcher in the Wall Street Journal published that a few years ago, it's been updated, where people can, students or anyone can file a report that is lodged with the police if they even, basically for thought crimes.
01:12:23.000 So our universities are experiencing ideological capture.
01:12:28.000 This has been the base of my disagreement with Tim.
01:12:30.000 The base of my disagreement with Tim is that this is the nucleation point.
01:12:33.000 This is the point from which it all erupts.
01:12:37.000 And what we haven't talked about at all, but I probably should mention
01:12:39.000 because it's so important for context, is that I can't, all my publications, et cetera,
01:12:47.000 etc. Or someone...
01:12:49.000 Let me take it off myself.
01:12:50.000 Someone who has a lot of publications, research, go to Google Scholar, teaching, a quarter century of teaching, etc., etc.
01:12:59.000 Somebody like that can't just go into the public schools and start teaching.
01:13:02.000 You need a teaching certificate.
01:13:04.000 All of the colleges of education that grant teaching certificates, literally all of them are woke.
01:13:10.000 They're predicated on Paulo Freire's Pedagogy of the Press.
01:13:13.000 James Lindsay, guest of your show, your friend, also talks about this pretty extensively.
01:13:18.000 This is a woke indoctrination mill for teachers in K-12 systems, and we get this madness.
01:13:25.000 So, I disagree with you.
01:13:28.000 I think you're wrong.
01:13:30.000 What we saw with, like, um... What was his name?
01:13:33.000 Nick Christakis?
01:13:34.000 Was that his name?
01:13:34.000 Yeah.
01:13:35.000 Nicholas Christakis.
01:13:36.000 Nicholas Christakis.
01:13:37.000 Princeton.
01:13:37.000 The students... Oh, no.
01:13:39.000 Yale.
01:13:39.000 Yale.
01:13:39.000 Sorry.
01:13:40.000 Freshman students already had this ideology within them.
01:13:43.000 Right.
01:13:43.000 They brought it to the universities.
01:13:45.000 Correct.
01:13:45.000 And the universities said, the customer is always right.
01:13:49.000 And then recycled back to them what they were demanding.
01:13:52.000 Okay, but you're actually making my point, unless I'm not understanding.
01:13:56.000 You're not understanding.
01:13:58.000 These 16-year-olds are on Facebook.
01:14:00.000 They're 10 years old.
01:14:02.000 They get a Facebook account.
01:14:02.000 They're not supposed to, but they do.
01:14:04.000 They see nothing but this weird algorithmic garbage.
01:14:07.000 They enter the university, and they demand these changes, and the schools say, yes, because you pay our bills.
01:14:13.000 So you're saying people from the university gatekeep who gets to teach at the elementary school.
01:14:17.000 Yeah, we're not disagreeing.
01:14:19.000 We're not disagreeing.
01:14:21.000 What I'm saying is the universities did not originate this.
01:14:25.000 There's elements of the ideology that obviously you can trace back to academia.
01:14:30.000 What I'm saying is the universities have these things within them because young people were manipulated, brought it to the universities and demanded it.
01:14:37.000 Okay, so if the causal explanation for this is solely algorithmic manipulation, Then how is it, given that you yourself have said there are an awful lot of ideologies, then how is it that it's a, what an utterly remarkable coincidence that it's the ideology taught in universities and it's the same one that the ideology pushes.
01:14:56.000 That's way too much.
01:14:57.000 When did it become, when did we start seeing these protests, this uproar, the far left
01:15:03.000 extremists?
01:15:04.000 It's been over the past ten years.
01:15:05.000 It has quite literally been since, LexisNexisDatus points out, the emergence of social media
01:15:10.000 around the world created an expansion of this ideology.
01:15:13.000 It seems like you both have a touch of my side bias, because you're a student of the
01:15:16.000 internet, Tim, and you're a student of the academia, it seems, and you're both right.
01:15:21.000 It's a confluence of events.
01:15:22.000 I'm not disagreeing with you.
01:15:25.000 He's disagreeing with me, which is totally 100% fine.
01:15:30.000 The algorithms amplified our world view.
01:15:36.000 At the same time as we started seeing the emergence of the far left stuff, we also saw an expansion of white nationalism.
01:15:42.000 I got it.
01:15:43.000 Critical pedagogy started before social media.
01:15:45.000 So critical pedagogy was the precursor to that.
01:15:52.000 Social media, Elissa, all these people, they are in a sense, they actually are, you know, everybody wants to be a victim, this is, John Heights is a victim culture, it is a victim culture, but they are actually victims.
01:16:02.000 They're epistemic victims of a deranged ideology.
01:16:05.000 Those people were then placed in seats of power from K through 12 up, they've been getting that, so the kids who came to Yale, they've been indoctrinated with that stuff for years.
01:16:18.000 This is what I don't quite understand, how it's an ideology.
01:16:21.000 Well, it's a suite of beliefs, right?
01:16:24.000 There's systemic racism.
01:16:25.000 Racism is baked into the system.
01:16:27.000 Privilege seeks to preserve itself.
01:16:29.000 All of those things co-occur.
01:16:32.000 Okay, well you've mentioned that a few times.
01:16:34.000 I'm trying not to go down that rabbit hole.
01:16:35.000 But this is a huge hole in the idea.
01:16:38.000 No, no, but it's not a huge hole, but it's also a lot of conservatives believe that too.
01:16:42.000 You know, I just talked to Douglas Murray the other day.
01:16:45.000 Douglas Murray was in Ukraine.
01:16:47.000 He is for the United States.
01:16:48.000 I mean, again, I'm trying not to go down to the Ukraine thing, but you have... It's just like... They oppose cryptocurrency too.
01:16:58.000 Balaji has some great stuff on that for why that is.
01:17:02.000 They're also very pro-vaccine.
01:17:04.000 It's a toxin to woke.
01:17:05.000 But all those things are in the orbit of the ideology.
01:17:09.000 All of them.
01:17:10.000 In essence of mindless adherence?
01:17:13.000 What do you mean?
01:17:14.000 Like, how is being at war in Ukraine, opposing cryptocurrencies, and being pro big pharmaceutical companies in any way associated?
01:17:23.000 I mean, in fact, I would argue being in favor of massive multinational corporations is actually paradoxical.
01:17:29.000 It's kind of like joining the military, assuming this consciousness.
01:17:32.000 Like, if Joe Biden tells you, then you're going to do it.
01:17:34.000 They've accepted, some people have accepted whatever they say.
01:17:38.000 Okay, so two things.
01:17:40.000 One, what Ronald Reagan said in his famous debate, there you go again.
01:17:45.000 You're assuming, and I don't mean that as a slander on you, but you're assuming a kind of rationality that the proponents of the ideology have that they simply do not have.
01:17:55.000 Oh, I'm quite the opposite.
01:17:57.000 I'm saying they quite literally have no rationality, and there is no coherent ideology at all.
01:18:02.000 What we're looking at is a sewer of refuse that has been mashed together, and what's happened is, I think regular people who are looking for answers seek experts who are anti-woke, and they find academics who then give them a very academic explanation.
01:18:17.000 Okay, so let's talk about a sewer that's been crammed together and force-fed, right?
01:18:24.000 Who is doing the force-feeding?
01:18:25.000 The algorithms.
01:18:27.000 Okay, again, I'm not saying you're wrong.
01:18:29.000 I'm telling you, in no uncertain terms, the data is utterly overwhelming for this.
01:18:35.000 The people force-feeding them are people who have gone through teacher training programs, their colleges of education, their university administrators.
01:18:45.000 We released a show a while Bill Asher released a show on my YouTube channel, Why Colleges Are Becoming Cults.
01:18:53.000 We know why this is true and the likelihood that two ideologies, that the same ideology could be both promoted in K-12 and academic institutions and be promoted by the algorithm at the same time is virtually zero.
01:19:08.000 So, Jack Dorsey, who pushed the free speech, winning the free speech party until 2014.
01:19:12.000 Who also lied before Congress and hasn't been subject to perjury yet.
01:19:15.000 After the fact.
01:19:16.000 Go ahead.
01:19:17.000 The original Twitter, when it was created, was the free speech wing of the Free Speech Party.
01:19:21.000 When he created Twitter, it was in a similar vein to the Ron Paul revolution of freedom, libertarianism, etc.
01:19:27.000 And then there was a shift.
01:19:28.000 Now, it's entirely possible that the explosion of data points we see from LexisNexis in every country on the planet at the exact same time could be because a globalist cabal got together, sat around a table and said, Implement Order Wokeness!
01:19:44.000 Our plan has come to fruition!
01:19:46.000 Ha ha ha!
01:19:46.000 And then they activated the device, which turned every country into a woke, you know, disaster zone.
01:19:52.000 Or it could be that the advent of cellular technology put a phone in everyone's pocket, and then we gave everyone access to communication devices, which amplified our social biases.
01:20:04.000 Okay.
01:20:04.000 Creating two things.
01:20:05.000 An interesting thing.
01:20:06.000 In countries like Iran and China, they had an inverse conversation.
01:20:11.000 They went completely anti-woke.
01:20:13.000 But still saw the emergence of the same language.
01:20:16.000 Okay, let me ask you a question then.
01:20:19.000 Taking everything that you just said, what is the disconfirmation criteria for that?
01:20:24.000 Like, what piece of evidence or rationale or reason would you have to hear that would throw that speculation or hypothesis into question?
01:20:34.000 The LexisNexis data that's been repeated on multiple occasions was incorrect?
01:20:39.000 The data is falsified.
01:20:41.000 The tracking of this terminology and the rise of this language was misinterpreted and, in fact, has always been around.
01:20:48.000 Something like that.
01:20:48.000 You could see the algorithmic code over the last 15 years if you had, like, snapshots of the code on a weekly database to see, like, was the algorithm actually inviting this discourse?
01:20:59.000 I spoke at universities after Occupy Wall Street.
01:21:02.000 I did not see this.
01:21:04.000 What year was that?
01:21:05.000 This is end of 2011.
01:21:07.000 I was being invited to speak at universities.
01:21:09.000 Into 2012 and 2013 and then... Yeah, I don't think it came to prominence.
01:21:15.000 It didn't come to prominence in the universities?
01:21:18.000 Well, I'm literally just going to say, it's also a geographical phenomenon.
01:21:22.000 So, for example, it's worse... I was in New York.
01:21:25.000 Well, it's worse on the coast.
01:21:26.000 It's worse in traditionally blue places.
01:21:29.000 When I was at Occupy Wall Street, wokeness was just starting to permeate.
01:21:34.000 into the leftist ideology, and it was confusing to a lot of people.
01:21:38.000 When I went and spoke at universities following Occupy Wall Street, all of the professors were classically liberal.
01:21:45.000 They were very much in that classically liberal vein, where I am right now, and you probably are.
01:21:51.000 And then over the next few years, especially with what we saw with Gamergate, We saw the expansion of wokeness, which came from the internet and from young people into the universities.
01:22:02.000 And then what we ended up seeing was a lot of people, you can see the effect with celebrities, you can see the effect with corporations, slowly start adopting what they think is prominent, and I think they do it.
01:22:13.000 Okay, I'll give you what you're looking for.
01:22:14.000 I'll give you what you're looking for, read the algorithm.
01:22:18.000 Let me ask you a question.
01:22:19.000 Let's say we had a pie chart.
01:22:21.000 And I'm claiming that it's K-12, colleges of education, college administrators, infused throughout the curriculum, diversity statements, all of the stuff that seek to both indoctrinate students in an indoctrination mill and perpetuate the ideology.
01:22:40.000 Let's say that you had two of those and they're not competing and you're going to fit them in a pie chart.
01:22:45.000 What percentage of that pie chart is algorithm and what percentage of it is academic institutions?
01:22:53.000 So what do you mean?
01:22:54.000 You mean like right now?
01:22:55.000 What's responsible for the freaking catastrophe?
01:22:57.000 Responsible?
01:22:58.000 100% algorithm.
01:23:01.000 Ninety, ninety-seven percent perhaps.
01:23:02.000 One hundred percent, okay.
01:23:04.000 Maybe ninety-seven.
01:23:05.000 Then our disagreement is far more substantive than I thought.
01:23:08.000 Yeah, so the issue is when... I think what's happening is people are seeking answers as to what this ideology is.
01:23:16.000 They're seeing specific examples of it.
01:23:19.000 They're, uh, many people are looking at the news in a very short-term perspective, and then they're following experts who they believe are academics, uh, I'm sorry, academics who they believe are experts on the subject matter, and who likely are in the specifics of, say, you know, critical race theory and critical theory, but they're not actually looking at the hard data of the social phenomenon that's occurred.
01:23:40.000 And the strange thing to me is, as I often try to explain, there's hundreds of thousands of ideologies that existed in universities 20 years ago, and sort of still do, but they're being pushed out.
01:23:49.000 Pre-social media, this wokeness did not exist.
01:23:54.000 At all!
01:23:54.000 at all. And so simply put, when Anheuser-Busch says, I don't want my content appearing next
01:24:04.000 to these things.
01:24:06.000 These things get removed from social media.
01:24:08.000 It's an amplification machine that keeps spinning up and recycling these ideologies which keep changing name.
01:24:15.000 And I think the issue is...
01:24:18.000 You know, I'm on the ground at Occupy Wall Street.
01:24:20.000 I'm on the ground at various protests in LA, all across the country.
01:24:23.000 I'm in Spain.
01:24:24.000 And I'm seeing similar things and I'm seeing similar behaviors that don't have a core ideological function.
01:24:29.000 And then today I'm seeing pro-vaccine stuff.
01:24:31.000 I'm seeing pro-Fauci stuff.
01:24:32.000 I'm seeing pro-Biden stuff.
01:24:33.000 I'm seeing vote blue no matter who.
01:24:35.000 Always use your corporeal form.
01:24:37.000 Chatbot GPT too.
01:24:39.000 And I'm seeing Ukraine war stuff.
01:24:41.000 None of these things fit your argument.
01:24:44.000 Now, it is true, I think, that universities have become mills of wokeness, but I think within these mills, it makes no sense that there's a rainbow flag next to a Ukraine flag.
01:24:55.000 There is no ideology that brings these things together.
01:24:59.000 What does, however, is a social media algorithm that says, adhere to orthodoxy.
01:25:03.000 Okay.
01:25:05.000 Or straight up the president telling people to do it, and they're so afraid of Donald Trump that they're just falling in line.
01:25:10.000 But why are they even afraid of Donald Trump?
01:25:11.000 Because they're like psychotic from poisoning the food supply, staring at the TV for four hours a day, eight hours a day.
01:25:18.000 Where in these theories, in these ideologies, is Trump derangement syndrome?
01:25:22.000 So the Trump derangement syndrome, as far as I know, I'm sure someone's written a paper about
01:25:31.000 it, but that's just something that people have used, that often say conservatives have used.
01:25:38.000 Sam Harris has an interesting thing about Trump derangement syndrome.
01:25:41.000 Well, he's Trump deranged.
01:25:44.000 The dude, Sam Harris, is clearly unwilling to reflect upon himself.
01:25:48.000 He's a perfect example of what he's complained about with Islam.
01:25:51.000 He doesn't self-reflect upon the fact that he's wrong at all.
01:25:55.000 He doesn't look at any of the data.
01:25:57.000 He just hates Donald Trump to the point where he said he believed it was good to cheat to win elections.
01:26:03.000 So here's my suggestion to you.
01:26:05.000 Why don't you have him on and have a conversation with him?
01:26:07.000 Oh yeah, absolutely.
01:26:09.000 The point is Trump Derangement Syndrome exists.
01:26:12.000 The left refers to it as Trump Anxiety Disorder.
01:26:16.000 Just to be clear for people listening, by Trump Derangement Syndrome you mean?
01:26:21.000 When people become irrationally angry at the thought of Donald Trump, The truth-telling around Donald Trump, or anything he does.
01:26:30.000 For example, when they said that the terrorist was an austere scholar.
01:26:34.000 Anything he does is framed negatively, and any defense of him in any factual stance like, hey, you know Trump, when he threw the fish with Shinzo Abe into the Koi pond, Shinzo Abe actually did it first.
01:26:45.000 They say, you're a Trump supporter, they get irrationally angry.
01:26:48.000 Sam Harris being a good example.
01:26:50.000 He says that, Sam Harris believes subverting elections is a good thing because Trump is that bad.
01:26:55.000 He said, Sam Harris actually said that if there was legitimate evidence of Hunter Biden sexually abusing children, it would not be as bad as Donald Trump running a fraud university.
01:27:06.000 I heard that on Trigonometry.
01:27:08.000 True.
01:27:09.000 Now where in leftist ideology does that exist?
01:27:12.000 It doesn't.
01:27:13.000 What does is that Sam Harris is beholden to liberal social orthodoxy crafted by algorithms and he doesn't want to break from it.
01:27:20.000 I'll tack on part two of the definition of Trump derangement syndrome is that when people are obsessively in love with Donald Trump, it's another form of derangement.
01:27:26.000 The cult worship, that kind of thing, blind adherence.
01:27:30.000 I agree with you, but people don't use the word to mean that, but it does exist.
01:27:33.000 Okay, so there's so much to unpack in this.
01:27:36.000 So when people are accused of Trump derangement syndrome, they will give the response that that is a slur that people use who whitewash the fact that Donald Trump, and I'm not saying I'm... I
01:27:52.000 gotta hear this, yeah.
01:27:53.000 It's that Donald Trump is an enemy of democracy and he has subverted the
01:27:58.000 democracy and he would not leave office, allegedly, and he's a traitor and
01:28:03.000 whatever else happens. My response is it's a cult. Well, their response is not that this is an
01:28:10.000 irrational hatred of Donald Trump, it's that this is a rational hatred of Donald Trump.
01:28:15.000 But any objective, and I mean this literally, any objective person knows that it is irrational.
01:28:21.000 I mean, Donald Trump was impeached.
01:28:23.000 Why?
01:28:24.000 Because he made a phone call to Ukraine asking about Joe Biden withholding congressionally approved aid in exchange for a political favor.
01:28:34.000 I agree with you.
01:28:35.000 I agree that, again, perhaps it's a difference of degree.
01:28:41.000 Simply put, these people believe that because social media algorithms fed them false narratives and lies.
01:28:48.000 Now, they say that every cell in our body is completely changed within seven years.
01:28:53.000 That means someone alive today I'm just making a nonsensical point.
01:29:02.000 The point is, a person's entire being, if they've been following corporate news since 2015, is comprised of life that believes all of these psychotic lies about Donald Trump.
01:29:12.000 that we have debunked over and over and over again.
01:29:15.000 And I don't think Trump's a saint.
01:29:17.000 I don't think he's the greatest president who's ever lived or anything like that.
01:29:19.000 I think on foreign policy he's the best I've seen in my lifetime.
01:29:22.000 I can tell you all the things that I didn't like what he's done.
01:29:25.000 I can say that I was wrong when I took his side over Thomas Massey.
01:29:29.000 And then you go and talk to somebody and it's become a trope.
01:29:32.000 Stop making me defend Trump was a trope throughout 2016, 2017 when the media would lie.
01:29:38.000 They accused Donald Trump of being a traitor and having served the Russians as he was president.
01:29:45.000 And there are people who still believe this.
01:29:48.000 Now that's Trump derangement syndrome.
01:29:52.000 Well in that sense I think it's really important, your point is really important, and I think it's really important when we make mistakes that we've had about things in the past that we kind of fess up to them.
01:30:05.000 That's why I fessed up to you about the Hunter Biden laptop and me had fall fat.
01:30:10.000 So I did I did buy into a lot of that media narrative that maybe he seems unnaturally close to Russia, something's going on.
01:30:20.000 They lied.
01:30:21.000 Alphabank lies.
01:30:23.000 Well, I heard a lot of that.
01:30:24.000 There's a tape.
01:30:25.000 I'm sure you've heard that.
01:30:26.000 Russian ladies of the night.
01:30:27.000 Yeah, you heard that rumor.
01:30:28.000 Drinking a lot of beverages.
01:30:29.000 And this is my point.
01:30:30.000 Social media algorithms were feeding this content to people because it got clicks.
01:30:34.000 So what we saw was... Again, Tim, I'm not disagreeing with you.
01:30:38.000 This is a huge problem.
01:30:40.000 I'm not disagreeing with you.
01:30:42.000 I don't know how to get out of it.
01:30:43.000 Which is why I mentioned the my side bias, which is making it worse.
01:30:47.000 I think that we get out of it in a few ways.
01:30:51.000 I think people who are aware of the issue Should have as many kids as possible.
01:30:56.000 I know everyone points to me and asks me where my kids are.
01:30:58.000 You know, personal issues, but we'll talk, but you know, I'll keep that private.
01:31:01.000 So I think people should have as many kids as possible.
01:31:03.000 That doesn't mean your kids will think like you.
01:31:05.000 There's no guarantee, but there's also a combination of factors.
01:31:09.000 One, get out of cities.
01:31:11.000 We, you know, we see what's going on with Riley Gaines.
01:31:13.000 We see what's going on in Austin with Daniel Perry.
01:31:16.000 The lies the media is pushing on this guy, it's incredible.
01:31:19.000 So Daniel Perry's got a... Or get off social media altogether, like Lukyanov has been saying, or John Haidt, for kids, young girls, liberal girls, young women, white... Take your kids' phones away.
01:31:29.000 Oh no, it's actually, it's interesting, it's not YouTube, it's not, it's strictly social media, it's the problem that Haidt found.
01:31:37.000 Yeah, yeah, yeah.
01:31:38.000 Your kids should not be on social media.
01:31:40.000 So like texting?
01:31:40.000 They shouldn't have phones.
01:31:42.000 I don't think they should have phones.
01:31:42.000 Well, that's another thing.
01:31:44.000 There's no... Again, I'm not familiar with the data.
01:31:46.000 I haven't looked into it.
01:31:47.000 But people who are familiar with the data say it's not watching YouTube, it's not reading books, it's not texting.
01:31:52.000 It's strictly social media.
01:31:53.000 I will clarify.
01:31:54.000 They have these cell phones that can only make and receive phone calls and texts.
01:31:57.000 Those are totally fine in my opinion.
01:31:59.000 You know, I thought you guys were saying something interesting about Trump Derangement Syndrome.
01:32:01.000 You're saying that it's like an irrational fear of this, and you're saying that it's considered a rational fear.
01:32:06.000 Yeah, I'm saying that the people accused of Trump Derangement Syndrome will claim that it's not a syndrome, that it's not irrational, it's rational.
01:32:15.000 I think that's the response.
01:32:16.000 And I'm trying to look up what does that even mean?
01:32:17.000 What's rationality?
01:32:19.000 It's something that you can claim based on a reason.
01:32:21.000 What's a reason?
01:32:22.000 Well, something that you think is right.
01:32:23.000 What's anianism?
01:32:24.000 Like, it's so subjective.
01:32:27.000 This stuff.
01:32:28.000 I actually think it's a really good point when we're trying to break down why these people think they're rational.
01:32:33.000 They think they have a reason for it.
01:32:34.000 The reason I say that they are Trump deranged is the emotional vigor that comes along with it.
01:32:40.000 You come to me and you tell me that Donald Trump was an evil tyrant, I'll say, why do you think that?
01:32:44.000 I go to them and I say, I think Donald Trump was the greatest president on foreign policy.
01:32:47.000 They lose their minds.
01:32:49.000 There's something really profound to that, right?
01:32:52.000 There's something that people who can't engage, they just get anthropogenic global
01:32:59.000 warming, right?
01:33:00.000 Like anthropogenic global warming.
01:33:05.000 If you make an argument against it, or if you make an argument, no matter what your
01:33:08.000 argument is, there are some people who will say, well, what's your evidence?
01:33:11.000 And there are other people who will lose their minds.
01:33:13.000 And this is the point I'm making ultimately about the modern left and wokeness, is that climate change, Trump derangement syndrome, and Ukraine, Fauci, none of these things line up with academia.
01:33:25.000 But they are deranged on all of these counts, and equally defensible on all of these counts.
01:33:31.000 What's the best way to figure out if someone should fear a president?
01:33:36.000 I think looking at history and figuring out how you should fear a president, you look at history, you should probably fear all of them, including Donald Trump.
01:33:47.000 Amen, absolutely.
01:33:48.000 So you don't think that Trump presented, maybe unique is too strong of a word, you don't think that Trump presented an especially dangerous threat to our democracy?
01:34:00.000 Why do you say our democracy?
01:34:02.000 Well, I mean the United States as opposed to Canadian democracy.
01:34:05.000 Trump represented probably the best effort we had at a saving grace of the American Constitutional Republic.
01:34:11.000 You mean the whole clearing the swamp thing?
01:34:13.000 Not necessarily.
01:34:15.000 My view of Trump is that he was more like a bull in a china shop, and that the ivory tower has been constructed over the past hundred years, probably since just before 1913.
01:34:22.000 Okay, let me ask you this question.
01:34:25.000 On a scale from one to ten, We should fear Donald Trump how much, and we should fear Biden how much?
01:34:33.000 Biden?
01:34:34.000 Well, it's tough.
01:34:35.000 It's not so much Biden himself, but Biden, for all that's around him, I would put it at an 8 or a 9.
01:34:41.000 And Trump, I'd put it at...
01:34:44.000 Six five or six.
01:34:45.000 Okay, so so this is really I want to say something so every time I've asked you a question You've given me an answer and I know that that must sound like well, of course I've given you an answer because you've asked me a question that I consider it reasonable Maybe if I asked you about question about your sex life, you wouldn't answer the question, but maybe he would or maybe he would I know I wouldn't ask him so I wouldn't ask him so it doesn't make a difference, but there is something also I think it's an element of that Trump derangement syndrome that you mentioned.
01:35:13.000 There's a kind of excitability or a kind of anger in that or a kind of like when you ask someone a question, they get upset that you asked them the question.
01:35:22.000 Yeah, absolutely.
01:35:22.000 Or they don't want the questions asked at all.
01:35:25.000 And that, I think, is more of a concern of the byproduct of what's happening to people's consciousness and how they engage each other.
01:35:36.000 Because they go on social media and they are told, this is what you must be whether it's true or false, this.
01:35:43.000 And I wonder if it's a component of low testosterone or just outside of that argument agreeableness.
01:35:50.000 Individuals who fear being ostracized because of, you know, humans survive by being in social groups.
01:35:57.000 Yeah.
01:35:57.000 and those who are more willing to take risks to lead, if it's the right thing, so leaders versus followers,
01:36:02.000 the people who are more likely to fall into the follower category are like,
01:36:05.000 don't challenge the order, I will not have you get me ostracized.
01:36:11.000 Whereas my view is kind of like, ostracize me, I don't care, I'll do fine on my own.
01:36:16.000 So I don't have that fear that these people probably do.
01:36:20.000 They get angry because their brain understands that what is being told to them is likely true,
01:36:27.000 but if it's true, it means that.
01:36:30.000 The social order they're using to survive is wrong.
01:36:33.000 Now they're faced with the evolutionary fear of a risk to their lives.
01:36:37.000 Yeah, so that's the question.
01:36:39.000 How do you help people, in the literature it's called, build a golden bridge?
01:36:44.000 Like, how do you help people walk over who are awoke?
01:36:48.000 And I'm thinking, like, you know, now I'm thinking the FBI is going to come for me when I say red pill, white pill.
01:36:52.000 You're done.
01:36:52.000 They're coming in, they're busting in the door.
01:36:54.000 I think it's already, my social credit score is already terrible.
01:36:59.000 You know, and how do you help people and not create adversarial relationships to say, oh, wow, geez, you know, I thought this and...
01:37:08.000 Anheuser-Busch, we need that apology.
01:37:10.000 More than anything right now, we need that apology.
01:37:12.000 We need to say to every single person who is blindly following social orthodoxy, you're on the wrong side of history.
01:37:19.000 Because that's what they say, because that's what they fear.
01:37:21.000 Okay, that's, by the way, as I've told Rita a thousand times, that's the, I've lost five friends now because of my stances on things.
01:37:29.000 And four of those five people, all for woke reasons, four of those five people have said to me the same thing, you're on the wrong side of history.
01:37:38.000 Exactly.
01:37:38.000 I hear that constantly.
01:37:39.000 And that's the only thing they have that attaches them to any of these arguments they might make, or any of these flag icons they'll put in their profiles, or in their windows, is that they fear being on the wrong side of history for one reason.
01:37:54.000 Being shunned from the majority means death.
01:37:56.000 Figuratively, but in the past, for humans, literally.
01:38:00.000 If Anheuser-Busch apologizes, we get the high ground and we say, you're on the wrong side of history on this one, then they get scared.
01:38:07.000 And they'll start to ask themselves, uh-oh, are we going to lose this one?
01:38:10.000 Should I be with them?
01:38:11.000 You know, people often ask, well, what are some of the things you're sure you're right about in the culture war?
01:38:16.000 Well, to me, it's pretty obvious.
01:38:18.000 Don't defund the police.
01:38:19.000 That's freaking insane.
01:38:21.000 The other thing is child gender affirmation, gender affirming care.
01:38:26.000 That is one I hear a lot of.
01:38:28.000 I've mentioned to my friends years ago, why don't you write down, so like three years ago, why don't you write down the beliefs that you have now that are new, that you've acquired in the last year or two, and why don't we revisit those in five years and see where we are?
01:38:41.000 Hence the wrong side of history, people.
01:38:43.000 And that's my other prediction too.
01:38:44.000 You think you're smart?
01:38:45.000 Don't think you're smart.
01:38:46.000 Make predictions.
01:38:47.000 My prediction is there'll be massive gaslighting at the end of this.
01:38:49.000 Oh, I never believed this.
01:38:50.000 Absolutely.
01:38:51.000 Yeah, because it's so ghastly.
01:38:53.000 They'll have to gaslight.
01:38:54.000 Well, if they want to be a part of the moral majority again.
01:38:57.000 There you go again.
01:38:58.000 And that's not the tribalism, but the one thing people want more than being right is to belong, right?
01:39:07.000 It's to be loved.
01:39:08.000 It's to be a member of a community.
01:39:10.000 Wokeness offers that very, very easily.
01:39:12.000 The signaling, the pronouns in the bio, etc.
01:39:15.000 We're gonna go to super chats, it's been a little long, so if you haven't already, would you kindly smash that like button, subscribe to this channel, share the show with your friends, and head over to TimCast.com, click join us, become a member, because we're gonna have a members-only uncensored show at about 10, 10 p.m., where that's where we're more likely to swear and be not-so-family-friendly and maybe get a little bit darker with the conversation, so you don't wanna miss it.
01:39:37.000 And you can also join the Discord server after you're a member where you can hang out with like-minded individuals.
01:39:41.000 And then if you join at the $25 level or you're a member for at least six months, even call into the show and ask questions of us and our guests.
01:39:48.000 But let's read Super Chats.
01:39:49.000 All right.
01:39:50.000 Freedom Jeffrey, 1776 says, Hi Tim, can't wait to see you guys Friday in Austin.
01:39:55.000 Awesome.
01:39:56.000 Will we be able to take video in the venue?
01:39:59.000 I don't know.
01:39:59.000 Are we live?
01:40:01.000 It's live.
01:40:01.000 We should be fine, right?
01:40:03.000 Yeah, but there might be venue rules, so I want to say I don't... I would love for people to be able to.
01:40:09.000 Talk about building a golden bridge.
01:40:11.000 Playing music.
01:40:11.000 Yeah, I'm totally fine with it.
01:40:12.000 And I think after the show, too, we're going to give Ian a guitar and make him play music.
01:40:15.000 Hell yes.
01:40:16.000 Let's do it.
01:40:16.000 Frequent measure.
01:40:17.000 Because we've got, I think, a couple hours after the show of, like, what do you want to do?
01:40:20.000 That's how you bridge the golden bridge.
01:40:22.000 Like, you need people to come to your show that are woke and that love Donald Trump.
01:40:26.000 A hundred percent.
01:40:27.000 And they all don't even know.
01:40:28.000 They don't ask.
01:40:29.000 It doesn't matter.
01:40:29.000 They're just vibing.
01:40:30.000 I couldn't agree more.
01:40:31.000 They look at each other as people.
01:40:32.000 Yes.
01:40:33.000 All right.
01:40:35.000 Let me read some more Super Chats.
01:40:38.000 Raymond G. Stanley Jr.
01:40:38.000 says, Tim, I enjoyed the Jorge segment.
01:40:41.000 Always nice, shining a light.
01:40:43.000 Uh, in, it's, when it's the darkest before the dawn.
01:40:46.000 We are based chads on that red pill and, oh, oops, sorry, FBI kid.
01:40:50.000 Go F your shelves.
01:40:51.000 Let's go, Brandon.
01:40:52.000 Yeah, Jorge Masvidal.
01:40:54.000 This is the, the, one of the greatest moments.
01:40:57.000 I just, I can't believe this happened.
01:40:59.000 Donald Trump sitting with Mike Tyson, Kid Rock, and Dana White, watching a UFC fight.
01:41:05.000 I saw that.
01:41:05.000 Yeah, I saw that.
01:41:07.000 It was great.
01:41:07.000 Shout out to Trump as the greatest president and then says let's go Brandon and the whole
01:41:11.000 crowd is cheering let's go Brandon as Joe Rogan is holding the microphone.
01:41:15.000 How is this real life?
01:41:18.000 But talk about a white pill moment.
01:41:20.000 The whole audience yelling let's go Brandon.
01:41:23.000 I've been in Miami the last few months.
01:41:25.000 I love it there.
01:41:26.000 The people there are incredible.
01:41:27.000 Lots of anti-communists, Cubans out there, that I resonate my dislike for communists as a Polish-born person.
01:41:34.000 So, beautiful place, beautiful people.
01:41:36.000 Can't recommend it enough.
01:41:37.000 Grant says, the real victory of the Budweiser boycott is making Bud Light the drink of the left and forcing them to buy and drink lots of Bud Light.
01:41:45.000 I saw a funny post.
01:41:45.000 They said, Bud Light is the beer for people who don't want to drink beer but want to drink lots of it.
01:41:52.000 Oh yeah.
01:41:53.000 And I'm like, yeah.
01:41:56.000 Talk about fifth generational warfare.
01:41:58.000 If you want to wipe someone out, that's another way of doing it.
01:41:59.000 Like, hey, get a bunch of beer.
01:42:01.000 Hey, go take this other product from Big Pharma.
01:42:03.000 When low-tar cigarettes were popular, people smoked twice as much of them.
01:42:09.000 Yeah, there's a lot of water in Bud Light.
01:42:12.000 Bud Light's a terrible beer.
01:42:13.000 I know!
01:42:14.000 Like, I'm not saying that for any other reason, but it's just a terrible beer.
01:42:17.000 Independent of who markets it.
01:42:18.000 It's gotta be, like, one of the worst beers.
01:42:21.000 It's just so basic, bland, not very tasty.
01:42:24.000 I think that's why people tend to go for it, because it's not overly tasty.
01:42:27.000 It's rice beer, yeah.
01:42:29.000 I'm into, like, I like the thicker, stoutier.
01:42:32.000 Yeah, Guinness.
01:42:33.000 Dark, chocolatey.
01:42:34.000 I'm a huge fan of Guinness.
01:42:35.000 You said Newcastle.
01:42:36.000 I love Newcastle Brown Ale.
01:42:37.000 So good.
01:42:38.000 Let's read some more.
01:42:39.000 But I try not to drink beer.
01:42:40.000 Too many carbs.
01:42:41.000 Yeah, I've kind of given up.
01:42:42.000 I don't like beer.
01:42:43.000 All right, let's read some more.
01:42:44.000 Max Reddick says, Medhi Hassan is an evil person.
01:42:47.000 Oh my God, he's a lunatic.
01:42:48.000 He is defending the government on the Twitter files, arguing semantics, saying the government simply pointed out to Twitter that people were breaking their rules.
01:42:55.000 Insane.
01:42:55.000 He's just not an honest broker of conversation.
01:42:58.000 None of these people are.
01:42:59.000 He's just not honest.
01:43:01.000 Just none of these people are.
01:43:02.000 He's been in the news pretty heavily the last week.
01:43:05.000 Can I just talk about that just for a quick second?
01:43:07.000 It's taken me a lot of work to really get it into my mind that some people have absolutely nothing of substance to add to a conversation.
01:43:16.000 In fact, even listening to them is a kind of cognitive poison.
01:43:20.000 It's like a slow epistemic punch in the head.
01:43:24.000 Oh my gosh, I get this thing, sometimes people will be talking right, and they'll say right at the end of the sentence, and I have to turn the video off when it's happening, because I'm like, I'm not going to get brainwashed talking like that.
01:43:36.000 Right?
01:43:36.000 Alright.
01:43:37.000 Max also says, Tim, I know you don't like to say when guests are coming on, but would you mind breaking that rule once, and let us know when the serfs are coming on.
01:43:45.000 I need to see that guy get pressed.
01:43:47.000 So, uh, what's the guy's name?
01:43:49.000 Lance.
01:43:50.000 Lance?
01:43:50.000 Yeah.
01:43:51.000 I mean, he tweets at me and I say, come on the show, because I always do.
01:43:54.000 Like whenever a leftist tweets at me or something, I'll be like, bro, you're welcome to come on the show and have a conversation.
01:43:57.000 So that's, okay.
01:43:58.000 So that's vital.
01:43:59.000 And that's piggybacking off what you said.
01:44:02.000 When people with whom you have a disagreement, you know, like I've had a disagreement with Matt Walsh.
01:44:07.000 I invited him to a conversation.
01:44:08.000 Of course, he doesn't respond.
01:44:10.000 I think that's one of the things you have going for you.
01:44:12.000 If someone disagrees with you, I've always heard you say in the past, we'd love to have a conversation.
01:44:17.000 And then 97% of leftists ignore it or lie or turn it into a drama moment.
01:44:22.000 100% and they say, Tim doesn't talk to people he disagrees with, right?
01:44:25.000 I get that constantly.
01:44:27.000 Why don't you go on left-wing media?
01:44:28.000 Well, because I'm not invited on left-wing media.
01:44:30.000 You know, we invited one guy on the show, and he publicly says, I'll do it, and then privately goes, I'm not going on your show.
01:44:35.000 And then he takes it and turns it into a segment where he's like, Tim Pool is scared to debate me and all this stuff, and I'm like... So there you go.
01:44:41.000 I mean, but that's their tactic.
01:44:42.000 Right.
01:44:42.000 It's a woke tactic.
01:44:43.000 But this dude from the surfs tweeted at me, and I said, come on the show, and he was like, are you serious?
01:44:47.000 Hell yeah.
01:44:48.000 And I'm like, awesome, dude.
01:44:49.000 I look forward to it.
01:44:49.000 Like, you're allowed to have your opinions, man.
01:44:52.000 And we'll talk about them.
01:44:53.000 We've had Destiny the Omni-Liberal on a couple times.
01:44:56.000 You know, a lot of people don't like him, they disagree with him, he gets things wrong or whatever, but I'm like, I think Destiny's great.
01:45:00.000 And he's going to be here Saturday, too.
01:45:02.000 You want to hear something totally crazy?
01:45:04.000 I am in Romania, giving talks, doing all this stuff, and I'm like walking on a back road in the middle of freaking nowhere.
01:45:15.000 And this is a totally true story.
01:45:17.000 And this woman comes running up to me and says, Oh my God, oh my God, I saw you on Destiny.
01:45:25.000 In the middle, yeah, in the middle of nowhere, in some village in Romania.
01:45:31.000 Wow!
01:45:31.000 Yeah, I'm gonna tell him that story when I see him.
01:45:33.000 All right, oh snap, it's Dave says, Luke, liquor stores stayed open because people who drink on the daily, if COVID didn't kill them, the DTs would have.
01:45:40.000 Yep.
01:45:41.000 Quitting drinking cold is worse than the needle.
01:45:43.000 Your liver goes in complete shock.
01:45:45.000 Absolutely, yep, totally right.
01:45:47.000 I'm not disagreeing with you, but the optics were still bad when the government was shutting down mom-and-pop businesses.
01:45:51.000 And churches!
01:45:52.000 And churches, but liquor stores were okay, and I understand there's medical reasons.
01:45:57.000 That's how addictive some of this stuff is.
01:45:59.000 The church thing, to me, is the most egregious because it's the First Amendment.
01:46:03.000 It is like the most vital for us to experience life as we understand it, and they shut churches down.
01:46:10.000 I mean, that just, that blew my mind.
01:46:13.000 That's crazy.
01:46:15.000 All right, all right, we'll read some more.
01:46:17.000 Wrath of Paul says the Bud Light fiasco only distracts us from the bigger issues that actually kill our freedom such as the impending collapse of the petrodollar and the implementation of CBDCs around the world.
01:46:28.000 Well my view is not so simple on the petrodollar.
01:46:31.000 I think It's actually, I think the petrodollar is dying.
01:46:35.000 It's been dying for a decade plus.
01:46:37.000 Yeah, if you see the curve, it's going down.
01:46:39.000 And Donald Trump's view, this is my personal view, I could be totally wrong, I gotta ask the guy, but I think he sees this and he says, if we don't secure our borders, if we don't bring back manufacturing, if we don't get the American people back to work when the petrodollar falls, we're done.
01:46:53.000 And the Democrats were like, no, no, no, Trump, you're wrong.
01:46:56.000 We should destroy the planet before the petrodollar falls.
01:47:00.000 And so, even though it's still gonna fall, they're like, warring Ukraine at any cost, and now we got those Pentagon leaks, which, assuming that are real, shows that Russia actually ain't losing.
01:47:10.000 That Ukrainian air defense forces are gonna be depleted by the end of May, that the money we just sent to Ukraine is not enough, and that only around 8% of Russian troops that have been reported have actually died.
01:47:21.000 So when they say 200,000, the numbers from these leaks, 17.
01:47:26.000 That I haven't heard.
01:47:26.000 Now, the argument is that these are, these are, it's propaganda and the numbers are manipulated.
01:47:33.000 And the language and the context is all manipulated, perhaps.
01:47:36.000 Has there been any statement from the Pentagon if it's true or false?
01:47:38.000 Well, according to the military analysts, the people pretty much speaking for the Pentagon,
01:47:42.000 they're like, yeah, everything here released is true, but they just faked the numbers here
01:47:45.000 for this specific reason, which is very convenient to our official story.
01:47:48.000 That's what they're going with here.
01:47:49.000 But there's a lot of other things going around.
01:47:52.000 I was talking about this in my YouTube video today, specifically with China brokering a peace deal between Iran and Saudi Arabia.
01:47:58.000 Saudi Arabia has just called for the end of the Yemen war, which is a proxy war.
01:48:02.000 They're bringing the Sunnis and Chias together.
01:48:04.000 There's a big move here against the dollar here, and this is going directly against Israel and Americans' foreign policy in the Middle East, which is going to be some very significant ramifications for.
01:48:13.000 You know, I'm reading this here super chat about Ian, and I think Ian's right.
01:48:17.000 Duncan Chestnut says, Ian landed a 20 here.
01:48:20.000 Apology is not enough.
01:48:21.000 Make them an example.
01:48:22.000 Going woke will collapse your company, not just inconvenience it.
01:48:25.000 And you know what?
01:48:27.000 I think that's actually a decent point.
01:48:30.000 What I was saying earlier is that if they apologize, then we go out and we buy as much of their product as possible.
01:48:35.000 Because what we want people to see is, we want all of the analysts to be like, well, when they did this ad campaign, their stock collapsed and their sales collapsed.
01:48:43.000 When they apologized for it, they made a billion dollars.
01:48:46.000 So I don't understand.
01:48:47.000 Why would we want to punish them for apologizing?
01:48:50.000 Well, it's not about apologizing.
01:48:51.000 It's that if they don't... Well, so that's what he's saying.
01:48:53.000 Apology's not enough.
01:48:54.000 Make him an example.
01:48:55.000 My view is, if they don't apologize, then I hope they completely collapse.
01:49:00.000 Because then the message is... Oh, that I'm 100% fine with.
01:49:03.000 But if they apologize, then apologizing for doing something wrong should be rewarded.
01:49:07.000 But you shouldn't do... Aristotle actually says it.
01:49:10.000 Sorry.
01:49:10.000 But you shouldn't do a bad thing to fear remorse.
01:49:14.000 You should fear remorse only once you've done a bad thing.
01:49:16.000 Sorry.
01:49:17.000 If someone walks up to somebody and just punches them and knocks them on the ground and then stands there and goes, sorry.
01:49:22.000 I'm not buying it, bro.
01:49:23.000 I'm not gonna buy your beer.
01:49:25.000 You just decked that guy.
01:49:26.000 Like, no.
01:49:26.000 No, no, no, no.
01:49:27.000 That apology's not enough in that situation.
01:49:28.000 This is different.
01:49:29.000 They just smeared Dylan Mulvaney, they made a fool of him, and they wrecked their brand.
01:49:34.000 We deserve to be dragged for this.
01:49:36.000 But let's make it a little different.
01:49:38.000 A group of people are standing in a big circle, and there's a guy named Bush, and then another guy standing next to him.
01:49:47.000 And everyone's yelling, so the Bush guy goes, alright, I'll please these people, and just punches the dude as hard as he can, like you described.
01:49:54.000 However, everyone starts booing, and then the guy's like, what's happening?
01:49:59.000 Why are you booing me?
01:49:59.000 I thought this is what you wanted.
01:50:00.000 And they were like, no, we don't want you to do this, and now we're going to take you down.
01:50:04.000 And he goes, no, no, no, no, I'm sorry, I'm sorry.
01:50:06.000 Yeah, I don't think he's actually sorry.
01:50:08.000 I think he's scared.
01:50:09.000 And I think when he then yells out to everyone, I am scared of you, please stop, it sends a message to every other corporation.
01:50:15.000 If you want to see your net worth plummet by $4 billion, play that game.
01:50:21.000 So for you, correct me if I'm wrong, this Bush thing is kind of like a line in the sand.
01:50:26.000 And this is where we can actually do something to make a substantive change.
01:50:30.000 I'll be honest, you know, Dylan Mulvaney is a product, as Vivek Ramaswamy said in a rather brilliant fashion, we thought the AI monster we would have to deal with is going to be a robot with laser eyes, but it's Dylan Mulvaney.
01:50:42.000 It's the manipulation young people are experiencing at the hands of social media.
01:50:46.000 And so, it's not so much a line in the sand.
01:50:48.000 When the story dropped, I didn't care.
01:50:50.000 When I first saw the story that Dylan Mulvaney was being sponsored by Bud Light, I went, well, how about that?
01:50:56.000 But when Kid Rock then fired at a bunch of Bud Light, I said, OK, I will do a short segment about this.
01:51:03.000 And so it wasn't even a main segment for me.
01:51:05.000 And then I got 1.5 million views, and I said, OK.
01:51:11.000 People have rallied around this idea and begun a boycott, and then Travis Tritt came out later, an event got cancelled, and I said, this might be the moment people actually decide I can do some good here.
01:51:24.000 And so with that I said, okay, let's follow the story and see what's happening.
01:51:28.000 Now that we're seeing real economic consequences, For me, it's opportunity.
01:51:33.000 It is, finally, we have a $4 billion get-woke-go-broke, and everyone is piling on.
01:51:39.000 And then the media started attacking me, and I said, over target.
01:51:43.000 Can I just, if I may, put in a plug for my two passions.
01:51:47.000 So, Star Trek has been destroyed.
01:51:50.000 Star Trek Discovery has been destroyed by woke madness.
01:51:53.000 It's turned the show into a living cesspool.
01:51:56.000 Doctor Who, the last two seasons, have been utterly destroyed.
01:52:00.000 And I would like to add, if I may, I would like apologies for destroying the franchise from both the BBC, you know, destroying Terry Nation's work, destroying Star Trek.
01:52:13.000 So please add those to your list.
01:52:15.000 I canceled my Paramount.
01:52:17.000 Yeah, Star Trek was It's one of my favorites.
01:52:21.000 Next Generation, of course.
01:52:23.000 Deep Space Nine.
01:52:24.000 And the first problem I had is they only do prequels.
01:52:28.000 So boring.
01:52:29.000 And then they, you know, get woke, go broke.
01:52:33.000 So it's unfortunate.
01:52:34.000 The prequel thing didn't bother me, but the woke arc, the woke story, everything about it was interesting.
01:52:40.000 I couldn't stand it.
01:52:40.000 It destroyed the franchise.
01:52:42.000 All right, let's read some more.
01:52:43.000 We got Insane Redneck.
01:52:44.000 He says, regarding encryption, my brother just finished his time in the army and just started work as a fort guard.
01:52:50.000 He told me the DoD has classes for staff purely for the sake of bypassing VPNs.
01:52:56.000 Unsurprising.
01:52:58.000 Brandon Allen says, look at Luke rocking that cowboy hat based LARP.
01:53:03.000 I'm not laughing, this is what I always wear.
01:53:04.000 He's making a joke.
01:53:05.000 I know, the FBI.
01:53:08.000 Alright, let's see.
01:53:11.000 What do we got?
01:53:12.000 D.L.
01:53:12.000 says, chickens with pew-pews.
01:53:14.000 You're not taking the eggs this time.
01:53:16.000 Quick, someone meme that.
01:53:17.000 Side note, ever think of quick access app?
01:53:20.000 With sub, that includes all your sub-based media for one low monthly price using access code.
01:53:25.000 The app's done.
01:53:27.000 The issue is it's really hard to get an app on the App Store for iTunes.
01:53:31.000 But I think the Android app is going up really soon.
01:53:34.000 And this means that you can log in and get access to all of the content and it'll be really easy to navigate and I'm really excited for it.
01:53:40.000 Super cool.
01:53:41.000 I have to say, I'm just watching this now.
01:53:45.000 Boy, your viewers are really thoughtful.
01:53:47.000 Some of these comments are just fantastic.
01:53:48.000 Comments are incredible.
01:53:49.000 I'm always in the comments section, always commenting as well.
01:53:52.000 The comments are really thoughtful.
01:53:53.000 But we get corrections in real time, too.
01:53:55.000 Yeah.
01:53:55.000 Like, we say something wrong, I can see it, and I'm like, oh, hey guys, someone just pointed out, like, we've gotten guns wrong.
01:54:01.000 And do you say, hey, I made a mistake?
01:54:02.000 Oh, yeah, yeah, yeah.
01:54:03.000 So that's what differentiates you from I mean, just think about how silly it would be if someone superchats, you said the name of the gun wrong, it was actually this, and then I went, well I don't care, I'm not correcting, like, they're literally listening to each other, and they know I'm wrong.
01:54:16.000 The only thing I can do is be like, I got that one wrong.
01:54:18.000 Yeah, so you have a good audience, that's good, that's really good.
01:54:21.000 Smart people posting smart comments.
01:54:23.000 I read every comment.
01:54:24.000 There's a big difference in a crowd that just watches for kind of in the background and people that really listen to the stuff, the content.
01:54:30.000 These people listen.
01:54:31.000 Here's a good one.
01:54:32.000 Miguel D says, I'm a staunch conservative and I'm sounding more like Bernie Sanders every day.
01:54:36.000 Billionaires are the problem.
01:54:37.000 They own the corrupt media, big tech and pay off the politicians.
01:54:41.000 Is it billionaires or is it money that's the problem?
01:54:43.000 No, it's neither.
01:54:46.000 It's not even billionaires.
01:54:47.000 Elon Musk's a billionaire.
01:54:48.000 The problem is corruption.
01:54:51.000 Bad people exist.
01:54:52.000 You know, look, a bad guy who's poor can go murder a bunch of people.
01:54:57.000 A bad guy who's very wealthy can cause massive destruction culturally and politically.
01:55:02.000 It's still just bad people doing bad things.
01:55:05.000 You know, Elon Musk is, you know, I got my criticisms of him.
01:55:08.000 You can talk about his deference to China or to whatever degree that is.
01:55:12.000 He denies it.
01:55:13.000 But then you can talk about buying Twitter and taking the W off the sign and then painting it white because the landlord complained.
01:55:19.000 Yeah, that was great.
01:55:19.000 He's a funny guy and he's doing good things.
01:55:21.000 I do think the value of wealth hoarding is a problem.
01:55:23.000 You really shouldn't be valuable to not spend your money.
01:55:26.000 That doesn't make any sense.
01:55:27.000 Currency is supposed to, like a current move, This guy playing with words over here.
01:55:33.000 Inflation fixes that problem.
01:55:36.000 Yeah, or deflating currency, something like that.
01:55:38.000 Well, this is why they do negative interest rates.
01:55:41.000 They want to try and get you to remove your money from bank accounts and go spend it or something.
01:55:45.000 The whole modern monetary theory, I think, is a big corrupt nightmare.
01:55:49.000 It's so crazy that people that have billions just sit with it in a bank account at 2% interest and then they're just like...
01:55:54.000 They're richer.
01:55:54.000 They make more money than I do.
01:55:56.000 I'm bullish on Bitcoin.
01:55:58.000 All right.
01:55:59.000 What is this?
01:56:00.000 Barker?
01:56:01.000 Lewis can't choose.
01:56:02.000 Okay.
01:56:04.000 Vivek on his podcast with Bill Burr talked about using and promoting the FBI not reform it.
01:56:09.000 He is a wolf in sheep's clothing.
01:56:11.000 Yeah, for sure he is.
01:56:12.000 That's why you should be following his work.
01:56:16.000 I think there's a degree of hidden agenda to a certain extent with Vivek, but I mostly think that he means it.
01:56:24.000 He knows what he's talking about.
01:56:25.000 And when I talked to him about his story, about how he got involved, basically he ran a multi-billion dollar company.
01:56:30.000 It was a great interview.
01:56:31.000 He got attacked by woke people.
01:56:33.000 It pissed him off because Woke Inc.
01:56:35.000 is a wonderful book for people who haven't read it.
01:56:37.000 I think he's emotionally slighted by the fact that he said, sure, I'll write a statement, did, and then went, it wasn't good enough, and then they resigned, and he's like, you eff with me, you get the horns, you know what I mean?
01:56:48.000 And I think he's going to play ball with the military-industrial complex.
01:56:51.000 He's been pretty straightforward that he's going to be a different guy after eight years in there, and that's why, because he will decide where the bombs fall.
01:56:57.000 I don't know about that.
01:56:58.000 He's the real deal.
01:56:59.000 He's like Obama, man.
01:57:02.000 I think he's more like a right-wing Andrew Yang, but I think he'd run circles around Andrew Yang.
01:57:06.000 Well, he's talking about a war with the cartels.
01:57:08.000 The CIA is not going to be happy with that.
01:57:10.000 Those are two of my favorite people, by the way.
01:57:11.000 I've been thinking about the cartels.
01:57:13.000 What do we do about that?
01:57:14.000 Like, do we mobilize the full force of the American military and go to war with them?
01:57:18.000 That's what he wants to do.
01:57:19.000 That's his proposal.
01:57:19.000 That's what he wants to do.
01:57:20.000 Like declare war on Mexico?
01:57:22.000 use at work with their government but invade the country?
01:57:24.000 Mexico is fast becoming a narco state.
01:57:25.000 Like if they're letting cartels run their country, we have no choice but to take control
01:57:29.000 and seize the land.
01:57:30.000 I mean, I would even consider taking it.
01:57:32.000 If they're not going to govern and they're just going to let them attack us?
01:57:36.000 That is another conversation.
01:57:38.000 That's a huge conversation.
01:57:39.000 Wait until you learn about government, Ian.
01:57:41.000 Let's read some more.
01:57:42.000 We got KCB who says, Tim, I do share your videos occasionally if I find them profound on Facebook, and none of them get any reactions.
01:57:49.000 It's impossible that no one would give it a like, even if they don't watch.
01:57:53.000 Your content is being hidden by Facebook.
01:57:55.000 Can confirm.
01:57:56.000 No, what are you going to do about it?
01:57:58.000 We started posting more and more onto Facebook because it is a platform we should be using to get access to more people.
01:58:04.000 And we put them on Rumble as well.
01:58:06.000 I'm actually beginning to think about uploading to Twitter.
01:58:10.000 I think Twitter is an avenue that's not being explored properly right now.
01:58:14.000 With Twitter Blue, you can actually upload, what is it, a couple hours or something?
01:58:16.000 Is that what they're doing?
01:58:18.000 I don't know.
01:58:18.000 It's been a long time, yeah.
01:58:19.000 So I'm thinking maybe we need to just run the TimCastIRL channel like we do for Instagram and just start producing more and more content for it.
01:58:30.000 All right, let's see what we got here.
01:58:32.000 Jomas says, I install steam boilers for a living.
01:58:35.000 Beers after a really dirty one brings the guys together.
01:58:38.000 Just saying, it's not all evil gents.
01:58:40.000 Morale matters.
01:58:42.000 No, that's why I said I like beer.
01:58:43.000 I think beer is totally fine.
01:58:44.000 I'm not going to drink it, but I understand why people do.
01:58:46.000 Totally.
01:58:47.000 I mean, and I actually like Blue Moon, but I haven't had one in a really long time.
01:58:51.000 I really try not to drink.
01:58:52.000 I don't know, you don't drink, right?
01:58:53.000 No.
01:58:53.000 I try not to drink at all.
01:58:54.000 I just love drinking.
01:58:56.000 I just find it so enjoyable, but I'm trying to cut down on my carbs and my wheat.
01:59:01.000 So, no sugars and no wheat, or minimal.
01:59:05.000 Smart man.
01:59:06.000 Yeah, and I feel better when I'm at a better weight, you know?
01:59:10.000 I think this healthy at every size is BS.
01:59:12.000 But there is something about, you know, some minimal diet, cutting sugar and minimizing alcohol.
01:59:19.000 Although I will say, you know, so I went from gin and tonic to half gin, half tonic double
01:59:27.000 gin to quadruple gin, half tonic, and that's been working out for me and then I don't have
01:59:33.000 anything else.
01:59:34.000 Yeah, just replace the beer with beef liver.
01:59:36.000 Just have a bunch of beef liver with the boys.
01:59:38.000 Well, hard alcohol I think is better for it.
01:59:41.000 You know, it's like that movie from, I think it was like the 80s, where they're shooting into the house, the vampires are in the house, they're shooting in the house, and the bullets aren't hurting the vampires, even though they're hitting them, but it's making holes in the barn.
01:59:56.000 Oh, and the sun's coming through.
01:59:57.000 Yeah, and the sun's coming through, and it's burning the vampires.
02:00:00.000 That's what alcohol, to me, is.
02:00:01.000 It's the sugar.
02:00:02.000 Oh, there you go.
02:00:03.000 It's the sugar that gets you.
02:00:04.000 Same with the meat, it breaks down into sugar.
02:00:06.000 Let's grab one more idea.
02:00:08.000 What does it say?
02:00:09.000 Chris Mendilla.
02:00:10.000 Is that what it says?
02:00:11.000 Mendilla.
02:00:12.000 Billionaires don't keep their money in their bank accounts.
02:00:14.000 They invest in companies and stocks.
02:00:16.000 Most of these billionaires, like Elon, have very little liquid cash.
02:00:19.000 Most of their worth is from stocks.
02:00:20.000 Particularly Elon.
02:00:21.000 Everyone says he's worth $200 billion, but that's just stock in Tesla.
02:00:25.000 He can't do anything with that.
02:00:26.000 What they can do is take loans against their stock, and then they have a ton of money, and then they can cash out their stock when they're allowed to, and then pay back loans and things like that.
02:00:34.000 And pay a hefty tax bill.
02:00:36.000 That's right.
02:00:37.000 You paid a lot of taxes.
02:00:39.000 So, if you haven't already, would you kindly smash that like button, subscribe to this channel, share the show with your friends if you really like it, because that's the most powerful way for podcasts to grow, is word of mouth.
02:00:48.000 Head over to TimCast.com, become a member, because we're going to have an uncensored members-only show coming up in about ten minutes.
02:00:55.000 And if you join the Discord, you can get access to actually calling into the show after six months of membership or at least 25 bucks.
02:01:03.000 Then we screen callers and we choose a few every night.
02:01:05.000 So we just started doing this recently and it's the most fun thing, I think, all day.
02:01:10.000 So do that.
02:01:11.000 TimCast.com.
02:01:13.000 You can follow the show at TimCast IRL on Instagram and I think everywhere.
02:01:17.000 And you can follow me personally at TimCast.
02:01:19.000 Peter, do you want to shout anything out?
02:01:20.000 Peter Boghossian, I'm on Twitter at Peter Boghossian, B-O-G-H-O-S-S-I-N.
02:01:25.000 I have a YouTube channel.
02:01:27.000 I think it's, if you just put in Peter Boghossian, it will come up.
02:01:30.000 And I have a non-profit, National Progress Alliance, and we fight wokeness and promote free speech and open discourse full-time, so.
02:01:37.000 Appreciate the same thing, liking our channel, subscribing, and sharing.
02:01:42.000 Thank you.
02:01:43.000 How do you spell that again?
02:01:44.000 B-O-G-H-O-S-S-I-A-N.
02:01:49.000 Peter, thank you so much for coming on.
02:01:50.000 My YouTube channel is youtube.com forward slash we are change.
02:01:53.000 I started going live sporadically, randomly, doing little talks.
02:01:58.000 If you want to check that out, definitely subscribe.
02:02:00.000 I don't know when.
02:02:00.000 I don't have a schedule, so just make sure you subscribe there.
02:02:03.000 I'm also going to be doing a members-only meetup this Thursday in Austin, and if you want to find out about that, check out LukeUnfiltered.com.
02:02:10.000 LukeUnfiltered.com.
02:02:12.000 And I'm really excited to meet some of you guys here in Austin.
02:02:14.000 I think Luke and I may be doing a video this week.
02:02:17.000 Yeah, a live one!
02:02:18.000 It's going to be live, you guys, on Luke's channel, We Are Change.
02:02:21.000 It was really fun listening to you guys tonight.
02:02:22.000 Peter, thanks for coming, man.
02:02:23.000 It was really great.
02:02:24.000 I appreciate it.
02:02:25.000 Thank you.
02:02:26.000 I want to say I really enjoy the spirited discourse, and I think one of the reasons that your podcast has been successful, you and Rogan, is because you do have these kind of—and I like it when people tell me I'm wrong.
02:02:39.000 I find that to be in that Tim Urban's book, What's Our Problem, that's the higher level, that's what you want to do.
02:02:46.000 That's a reason that cements and glues friendships, is people say, you know, they call you out and you don't take it personally, I don't take it personally, I ask you a question, you answer a question, I ask you a question.
02:02:56.000 So I think that there's something really lovely about that and that's what cements relationships.
02:03:04.000 That's what we should all be doing.
02:03:06.000 And we also have Mr. Duprea on the microphone.
02:03:09.000 No camera though.
02:03:10.000 Yeah, no camera guys.
02:03:11.000 High energy surge in the house.
02:03:13.000 Yeah, but yeah, I'm just hanging out.