Valuetainment - February 19, 2026


“What If We Make SUPER Humans?” - AI Gene Editing Future SPARKS Human Enhancement CONTROVERSY


Episode Stats

Length

23 minutes

Words per Minute

196.88959

Word Count

4,680

Sentence Count

313

Misogynist Sentences

4

Hate Speech Sentences

2


Summary

In this episode, I sit down with neuroscientist, neurophysiologist, and author of the book "Aging is Dead" to talk about the growing field of aging research and how technology can help reverse aging. We cover everything from stem cell therapy, gene engineering, and artificial intelligence (AI) in general.


Transcript

00:00:00.000 The day ChadGBT came out was revolutionary to everything.
00:00:05.580 Everybody looked at, we used to Google.
00:00:07.900 Then, you know, I think it was first week of ChadGBT,
00:00:11.000 hey, do me a favor and create a song in Tupac's voice,
00:00:16.300 but the way that the president speaks.
00:00:19.060 And then you saw that one clip came out.
00:00:20.920 Wait a minute, that's amazing.
00:00:22.520 ChadGBT, I want you to take this research
00:00:24.380 and then all this data started coming out.
00:00:26.020 And like this, things have changed.
00:00:27.520 Two, three years, however long it's been.
00:00:28.760 How quickly do you think we're going to have new advancement
00:00:33.920 the next 5, 10, 20 years that will be just as shocking as OpenAI?
00:00:39.420 I'm super excited about how, you know, these language models,
00:00:46.400 these basically like LLMs and how AI is going to help really advance science
00:00:52.920 and particularly science in the field of aging.
00:00:55.940 You know, because at the end of the day,
00:00:57.680 you can do everything in your capacity with your diet and your lifestyle
00:01:02.020 to really give yourself that, you know, edge in terms of like aging better
00:01:07.440 and living longer.
00:01:08.380 But at the end of the day, you do have a certain genetic potential.
00:01:11.780 Like you could eat, you know, healthy and exercise
00:01:14.320 and still not live to be 120, right?
00:01:17.920 There is a genetic component to living to be 100.
00:01:20.440 Healthy and exercise and still not live to 100.
00:01:23.600 Living to 100, believe it or not, is largely genetic.
00:01:27.060 But that doesn't mean that, you know, diet and lifestyle don't matter.
00:01:30.640 They matter a great deal.
00:01:32.000 It's a matter of, you know, basically your quality of life
00:01:34.220 and not having Alzheimer's disease and cardiovascular disease
00:01:37.120 and type 2 diabetes and things like that later in life,
00:01:39.740 really degrading the quality of your life
00:01:41.360 and perhaps giving you an extra 5 years, right?
00:01:43.840 But you're not going to live 25, 30 years more by just being healthy, right?
00:01:48.820 So this is where gene engineering comes in.
00:01:51.660 And that's where I think the AI also comes in
00:01:54.540 because it's really going to, I think,
00:01:56.400 exponentially move that field forward.
00:01:59.240 Whereas it's been kind of like, you know,
00:02:01.760 it's been progressing but at a pretty slow rate.
00:02:04.160 What do you think we'll see?
00:02:05.100 Like give me three things you'll say.
00:02:07.720 These are three crazy wild things we may see.
00:02:09.880 Yeah.
00:02:10.340 So there's this – you've heard of stem cells, right?
00:02:12.700 Of course.
00:02:13.060 And have you heard of – do you know who Shinya Yamanaka is,
00:02:16.460 the Japanese scientist?
00:02:17.500 He won the Nobel Prize in 2006 for basically discovering
00:02:21.280 that you could take any cell from a person,
00:02:23.540 an 80-year-old person, an old cell.
00:02:25.600 You know, we're constantly getting –
00:02:27.160 skin cells are falling off our body constantly.
00:02:29.320 You can take some of those old skin cells
00:02:30.980 from an 80-year-old woman, for example,
00:02:33.920 and you could put four different what are called transcription factors.
00:02:37.900 It's basically just, you know, a gene that basically
00:02:40.440 is a master regulator of many, many different genes.
00:02:42.600 So it activates genes or it, you know, silences them so they're not active.
00:02:46.820 And you could put four of them on this skin cell, okay?
00:02:50.200 And you could revert that skin cell into what's called a pluripotent stem cell.
00:02:54.880 So that would be a cell that could become pretty much any type of cell in the body.
00:02:59.840 And so this is induced pluripotent stem cells.
00:03:02.560 And this is what Dr. Yamanaka won the Nobel Prize for
00:03:05.320 because it's really the ultimate reversal of aging.
00:03:08.340 If you take an old cell and turn it into a stem cell,
00:03:11.560 that can become any type of cell, a neuron, an eye cell, you know, a heart cell, right?
00:03:17.600 But so this is where it gets really exciting.
00:03:20.600 You don't want to basically – you don't want – if you want to reverse aging,
00:03:23.100 you don't want the cell to lose its identity, right?
00:03:26.560 You want to take a cell and reverse its aging but keep it that same cell, right?
00:03:30.760 So I just told you the skin cell becomes a stem cell,
00:03:32.780 and then it can become any type of cell.
00:03:34.240 So now we have some new data – this is all animal data –
00:03:37.600 where you can basically take this old skin cell
00:03:40.560 and just kind of pulse those four different, you know, those big genes,
00:03:44.980 those transcription factors on the skin cell, and you can epigenetically wipe out.
00:03:49.800 So epigenetics is like those factors that are sitting on top of our DNA,
00:03:53.840 like, you know, methyl groups, and they're activating genes or deactivating them.
00:03:57.220 You can wipe them out and make the cell young but still keep its identity.
00:04:02.920 And this has been done in animals, so essentially you're reversing the aging.
00:04:07.820 No one's been able to – I mean, there's a lot of hurdles to overcome in terms of, like,
00:04:11.240 how do we get this into humans?
00:04:13.120 You know, the delivery system is kind of questionable
00:04:16.560 because what they're doing in animals is giving it this virus,
00:04:19.180 and it's like, well, the virus could potentially integrate into the genome
00:04:22.780 and maybe be oncogenic.
00:04:24.200 So there's a lot of things to overcome before it goes to humans,
00:04:26.760 and I think that's where AI is going to come in
00:04:28.620 and help us figure out how we can translate this data into humans.
00:04:33.360 And I don't know if that made sense to you, but it's kind of this futurism type of stuff,
00:04:36.560 but it's very exciting.
00:04:37.980 It is.
00:04:38.700 But I guess what I want to go more to is –
00:04:40.620 I watched a movie yesterday, Saturday, called Mercy.
00:04:44.820 It just came out with Chris Pratt.
00:04:46.080 And it's like – it's one of these movies.
00:04:48.920 What was that movie back in the days where it could predict somebody committing a future crime minority report?
00:04:55.920 And Chris Pratt, who's a cop, comes in, and this software, this judge is AI.
00:05:02.720 Her name is Mercy.
00:05:04.000 And he comes in, and all of a sudden he wakes up.
00:05:07.780 He doesn't know what's going on.
00:05:08.900 And Mercy says, you just killed your wife, and you stabbed your wife XYZ times.
00:05:13.540 I was like, wait, what?
00:05:14.560 I would never do this.
00:05:15.480 And the AI convinces him that he did.
00:05:19.440 And then eventually you have 90 minutes to make the argument that you didn't do this.
00:05:24.500 So pull up this file, and put up that file, and pull up – you become the lawyer for yourself.
00:05:28.300 Very interesting, futuristic.
00:05:30.300 And the way it ended, the story – I'm not going to tell you what it ended, but the line that he says.
00:05:35.660 He says, here's what we learned, that both humans and AI can make mistakes.
00:05:40.640 But when you watch it, it was kind of like, man, this is very realistic.
00:05:45.660 We can get here very soon.
00:05:49.000 For the average individual, if you say, okay, what are some things people want to do?
00:05:52.840 We want to live a healthy life, enjoy it, live long, pain-free, right?
00:05:59.340 Enjoy ourselves to the point where, you know, I can still move my body in my 90s, in my 80s, maybe make it to 100s.
00:06:08.280 What massive, resounding, you know, innovation could happen that will shock everybody?
00:06:13.760 Is there anything?
00:06:14.320 Because we've heard about stem cells, right?
00:06:16.140 We've heard about stem cells.
00:06:17.440 I went to Beverly Hills.
00:06:18.460 One of our clients invited me.
00:06:19.580 I've got 80 million stem cells on my back.
00:06:21.480 They said, this is going to fix your lower back problem.
00:06:24.040 Nothing happened.
00:06:25.140 Is there anything that you think could happen that's revolutionary?
00:06:28.700 Well, I think that stem cells could be part of that, I would say, equation.
00:06:32.920 So we can have, you know, stem cells then that are tuning up every organ in our body, right?
00:06:38.540 And not just for regeneration or recovery.
00:06:40.680 But also we can grow organs.
00:06:43.140 Like, that's a new thing where people, scientists are now growing, like, even 3D printing human organs.
00:06:50.040 And so it's like, oh, your heart's giving out.
00:06:51.760 We're going to grow a new organ.
00:06:52.860 It's from your own cells.
00:06:54.680 So you're not going to have that rejection, you know, that rejection type of phenotype that happens.
00:06:59.400 So this organ regrowing thing is really cool where you're basically going to get a tune-up.
00:07:04.500 Oh, I need a new liver.
00:07:05.980 You know, it's not working.
00:07:07.140 At the end of the day, I need to inject some neural cells.
00:07:10.680 Stem cells in my brain because I'm losing, you know, my brain's atrophying.
00:07:15.080 I'm losing cells in my brain.
00:07:17.180 So I do think that's all on the horizon.
00:07:19.240 Can that be measurable?
00:07:20.320 Like, can it be done in a way where I instantly can see results?
00:07:24.220 Like, let's just say, you know, systems attached to my brain and it says, okay, what we just did,
00:07:30.340 it took you from this number to this number.
00:07:32.000 Look where you're at now just by doing this test.
00:07:34.080 Instead of, it could work 52% of the time it works.
00:07:37.840 Like, will there be things that I know for a fact will work?
00:07:40.680 Well, if you're looking at, for example, brain mass, yes.
00:07:43.920 I mean, if you see that you're injecting stem cells and now you're growing new neurons in the hippocampus,
00:07:48.860 part of your brain, which is involved with learning and memory, then that would be something that you can measure.
00:07:53.800 By the way, exercise can regrow brain cells in your hippocampus as well.
00:07:59.060 So that's one of the things that we do have control over right now.
00:08:02.460 We can talk about that.
00:08:03.800 And then gene therapy.
00:08:05.020 So, like, we do know there are a variety of what are called longevity genes.
00:08:09.900 Like, these are genes that are found in people that live to be 100.
00:08:13.200 So these are centenarians.
00:08:14.700 Or semi-supercentenarians.
00:08:17.200 So these are people that live to be 105.
00:08:19.560 Or the supercentenarians.
00:08:20.860 So these are people that live to be 110 and older.
00:08:22.800 There's a common denominator of really highly active genes that are found in these individuals.
00:08:28.760 And these genes happen to be genes that are involved in basically the cellular stress response,
00:08:35.800 which is essentially when you stress your body, your body adapts and responds and goes,
00:08:41.760 uh-oh, there's stress going on.
00:08:43.700 I need to respond to this so that I can stay alive.
00:08:46.560 And it does things like increases antioxidant genes or anti-inflammatory genes or genes involved
00:08:53.000 in making stem cells.
00:08:54.380 All kinds of these, like, stress response genes are really highly active in people that actually
00:08:59.700 do live to be 100 plus.
00:09:02.480 And so gene therapy can then deliver those genes to the right organs in humans.
00:09:08.420 And then, again, you'll be able to, you know, now deal with the stresses of aging better.
00:09:13.180 And I think that's another thing on the horizon.
00:09:15.480 So, for example, Dr. George Church, he's really the godfather of gene engineering,
00:09:19.300 played a role in discovering the human genome.
00:09:22.340 He's done studies in rodents and mice where they took, you know, three or four genes and
00:09:28.860 made them, you know, soluble, injected them in mice, and it extended their life expectancy.
00:09:33.100 It reversed aging of their organs.
00:09:35.000 He's now doing clinical studies in dogs.
00:09:37.980 It's much easier to get FDA approval to do this sort of things in dogs and humans.
00:09:41.020 But it's like, you know, once you get to the dogs, you have humans that are like,
00:09:44.560 of course, I want my dog to live longer.
00:09:46.720 And so they're going to, you know, start to approve those sort of therapies for dogs.
00:09:51.220 And then it's easier to then start them in humans.
00:09:52.980 So I think, again, AI is going to play a role in getting some of these gene therapies,
00:09:57.200 you know, from bench side, from, you know, this preclinical data into humans.
00:10:01.420 And all this is super exciting.
00:10:03.400 Oh, I'm excited.
00:10:04.600 And my optimism is the fact that you guys are going to figure things out.
00:10:08.140 So that's what I'm relying on.
00:10:10.400 I think the speed of using AI and gene, what was that one company video that we saw?
00:10:17.320 Was it CRISPR?
00:10:18.460 Was it called CRISPR?
00:10:19.840 Was it called CRISPR?
00:10:20.680 Yeah, CRISPR.
00:10:21.020 So is that kind of what you're talking about, that that's the direction we could be going?
00:10:23.920 CRISPR is one of the directions.
00:10:27.020 So gene therapy versus gene editing.
00:10:29.960 So gene editing would be CRISPR.
00:10:31.600 That's a technology that's used to change just like a single nucleotide of DNA, which can change the function of a gene.
00:10:38.500 So, for example, some people have, you know, these variations in our genes that make us not have them work as good.
00:10:45.200 Some people have diseases because of it, right?
00:10:47.180 Cystic fibrosis being one, right?
00:10:48.680 Muscular dystrophy, right?
00:10:49.740 Things like that where you have just a little one nucleotide or two nucleotide change in a gene, and it completely alters the function.
00:10:56.240 So CRISPR can come in and actually change that gene.
00:10:58.720 So you can basically take someone who has a disease and completely wipe that disease out and reverse it so they don't have it anymore.
00:11:05.760 Or perhaps you can alter the DNA of a gene and make it work better.
00:11:11.680 Gene therapy is where you're basically just putting more of a gene there.
00:11:15.720 So let's say we all, we have these longevity genes.
00:11:18.020 For example, FOXO3 being one.
00:11:20.380 This is a major transcription factor.
00:11:22.680 Again, it's regulating multiple genes involved in everything I just said, antioxidant production, anti-inflammatory production, you know, stem cell production.
00:11:30.880 All these things, the response to all kinds of different stressors that happens throughout our life as we age.
00:11:37.120 And you're basically able to handle that stress better.
00:11:39.380 So what you do with gene therapy is just give you a boost.
00:11:42.320 You get a boost more of that gene that's going to make everything else work better, kind of like a genetic tune-up.
00:11:47.300 So there's a little bit of difference between gene therapy and gene editing, and I think they can both work.
00:11:52.940 And it really depends.
00:11:53.760 It'll be more of an individual sort of personalized type of approach because everyone, we have different genetic makeups, and we have different benefits, and there's things that we aren't as good at depending on our genetic makeup, right?
00:12:03.700 And so it's really kind of cool.
00:12:06.220 It is.
00:12:06.720 It's the tuning up you can get.
00:12:07.440 It is.
00:12:07.900 And I think I saw an article a few months ago where it said, you know, there's going to be a time we can go in and say, you know, I'd want my daughter to be 5'11".
00:12:15.460 I want her to have blonde hair.
00:12:17.420 I want her to have blue eyes.
00:12:19.000 Will we get to a point that people would choose what their kids are going to look like?
00:12:21.900 Because that's pretty scary if you can get to that point.
00:12:23.900 Well, I mean, that's the ethical question because, yes, you could potentially do that with the gene editing technology and CRISPR and some of the, you know, some of the other types of technology that come out and a little bit more precise that are part of CRISPR.
00:12:37.760 I think that, you know, right now, I mean, think about IVF, right?
00:12:40.800 You go and people are getting screening.
00:12:43.520 You're screening for genetic diseases if an embryo has a disease, you know, you don't do it.
00:12:47.800 So at a certain point, like, where do you draw the line?
00:12:49.780 Okay, well, I don't want my child to have cystic fibrosis.
00:12:52.160 Okay, that's acceptable, right?
00:12:53.960 Because that's, like, obviously.
00:12:55.300 But that's the risk versus this is exactly what I want them to look like, right?
00:12:59.580 Right.
00:12:59.820 I mean, that's, and that's where it's like, well, what if you want to make, what if you want to make an embryo that is smarter, that is more resilient to disease?
00:13:08.700 You know, these things are probably going to be more acceptable in the community than phenotypes, than looks, right?
00:13:16.540 So it's like, it's hard.
00:13:17.520 It's like this blurry line and where, what's ethical, what's not, and how do you define that?
00:13:21.900 And it's going to be hard.
00:13:23.120 There's going to be all sorts of questions and, I would say, problems that are going to arise as this technology advances.
00:13:29.800 And it is very exciting, right?
00:13:31.180 I mean, you're talking about what if we could tune up humans where we're smarter, we're more disease resistant, we live longer.
00:13:38.260 I mean, who doesn't want that?
00:13:39.260 What would you take?
00:13:40.100 What would you do?
00:13:41.040 You're in this world.
00:13:41.820 Like, what would you want to see?
00:13:43.460 Because for me, when I think about, you know, kids, what you would want it to look like, what you would, you know, how smart you would want them to be.
00:13:52.940 Then it gets me to go think about, you know, the whole, I don't know if you saw the article came out last week about Chinese billionaires doing birth tourism and dropping off surrogates with hundreds of girls.
00:14:05.400 And, hey, I want you to go to U.S., become a citizen, then come back in.
00:14:08.620 Okay, so now let's take it even further.
00:14:10.240 And then now we're making superhumans, 150 IQ, strong, they can jump 48 inches.
00:14:16.600 And, I mean, what does the world look like in 50, 100 years if we're able to do that?
00:14:20.780 Right.
00:14:21.280 I mean, it's a really good question.
00:14:23.140 Not to mention the fact that, you know, if you start to screen for intelligence and, like, how do you define intelligence, right?
00:14:29.580 Because we have this, like, neurodivergence.
00:14:32.760 You know, people are so different.
00:14:33.960 And some people are very analytical.
00:14:35.560 You have that engineering type of brain.
00:14:38.300 And then you have the more conspiracy type and maybe a little more paranoid.
00:14:42.140 But those people also play a role in society.
00:14:44.620 They can connect the dots real easy.
00:14:46.120 And maybe sometimes they're making conclusions that might be not correct.
00:14:49.360 But sometimes they find things that people that are really super, super analytical don't.
00:14:53.460 And so it's a good question because, like, then you have to ask yourself, like, well, am I screening for a certain personality type?
00:15:01.700 And how is that going to change society, right?
00:15:03.960 I mean, because it's hard to draw that line.
00:15:08.220 And I know we're getting into these crazy ethical questions.
00:15:10.480 At the end of the day, I'm excited about healthier living.
00:15:13.760 I'm excited about, you know, yeah, like intelligence.
00:15:16.860 Again, how do you define intelligence?
00:15:18.420 Like, that's a whole other question.
00:15:20.540 We do know some genes that are involved in intelligence.
00:15:22.800 One actually is called Clotho.
00:15:24.840 And that gene is also involved in living longer and protecting against Alzheimer's disease.
00:15:29.840 It's very interesting.
00:15:30.620 Some people have different variations of that gene that makes it more active.
00:15:35.180 And those individuals don't get Alzheimer's disease.
00:15:38.360 They do have a higher IQ.
00:15:39.740 They are smarter and they live longer.
00:15:41.540 And so, you know, there is an initiative now for some researchers that are really trying to make a type of Clotho
00:15:49.320 that people can sort of inject and take and get to the brain and help protect them from Alzheimer's disease,
00:15:54.780 help, you know, make them live longer, perhaps be a little smarter.
00:15:57.640 And that's actually research that's going on right now.
00:16:00.200 So, you know, it's not like this stuff is like totally futurism stuff.
00:16:05.420 This is research that's happening now.
00:16:07.300 And I do think that with the help of AI, it will advance it.
00:16:11.060 With that said, you said something that I think is key, and that is AI makes mistakes.
00:16:15.920 And I can't emphasize that enough.
00:16:18.140 And certainly with each model that has come out, you know, over the past couple of years,
00:16:22.340 it's gotten better, especially if we're talking about the science, the field of science and research.
00:16:27.120 But I will say this, it depends a lot on your prompt, what input you're putting in,
00:16:33.160 because what I've noticed about AI is it also likes to kind of appease you.
00:16:38.580 And so if you're biased it a certain way with the way you're asking a question,
00:16:44.460 you might get a very different answer.
00:16:45.820 And I've tested this.
00:16:46.980 It also depends on the model that you're using.
00:16:49.140 I think that the free models don't ask science or research questions.
00:16:53.240 It's not going to be accurate.
00:16:54.380 If you're wanting to actually get a science, really evidence-based, you know, answer,
00:17:00.100 you have to have the paid version for one and really a deep research.
00:17:05.100 And that's something that's able to go out and look at the scientific literature.
00:17:08.260 And that is something that I would trust much more than just using like a GPT 5.2 free,
00:17:14.460 you know, version of it.
00:17:15.980 So I'll say that.
00:17:17.620 I'll leave it at that, because you can definitely get wrong answers
00:17:22.160 when it comes to asking a science-based or medicine-based question.
00:17:25.220 Yesterday, my CTO came to me and said,
00:17:26.660 hey, look, they measure every month which language learning model is the most accurate.
00:17:33.000 And right now, for the last month, it was showing me that Gemini is at the top at 38.5% accuracy
00:17:39.380 versus a lot of other things.
00:17:41.000 So even until today, it's still making mistakes.
00:17:42.920 And they're openly saying, we don't know it all.
00:17:45.120 We are making mistakes.
00:17:46.000 But, you know, in regards to brain, right, the way we think,
00:17:51.880 back in the days, or maybe even still today, we would say, oh, my God,
00:17:55.160 he's got a photographic memory.
00:17:57.300 He just remembers.
00:17:57.900 How do you remember that?
00:18:00.180 You know, how do you read a book and you retain so much of it, right?
00:18:05.520 How impressive will that be in 10 years, 20 years,
00:18:09.460 where we can all, you know, people choose to put a Wikipedia in here,
00:18:13.480 all the facts that are in the world, the facts, like the day you were born,
00:18:18.520 the day I was born, the day, you know, World War II or such and such person died
00:18:22.940 or this happened.
00:18:24.140 Once that's in there, what do you think will be impressive in 20, 30 years
00:18:27.440 with human beings?
00:18:28.480 What would we look at and say, man, she's pretty impressive.
00:18:31.600 He's impressive.
00:18:32.620 What will impress us?
00:18:34.040 I know.
00:18:34.480 That's a, it's an interesting thought experiment because it might be something
00:18:39.260 that is unexpected and surprising, like more just empathy or human qualities
00:18:44.380 or arts or being able to make something really creative that you're not really
00:18:48.640 going to be able to do with AI, right?
00:18:52.420 I don't know.
00:18:53.340 Or like being able to sing a beautiful song.
00:18:55.300 Like AI can make that music.
00:18:56.380 You're talking about Tupac.
00:18:57.880 Right.
00:18:58.100 I mean, my husband and I make AI music for my son to help him.
00:19:02.380 He's been learning Mandarin.
00:19:03.920 And it's like amazing, like the music that AI can make.
00:19:07.680 It's almost frightening because you might go, wow, like, you know,
00:19:10.920 what's going to happen to the music industry?
00:19:13.800 Yes.
00:19:14.160 But I do think that at the end of the day, what's going to happen is as we get
00:19:17.540 more, you know, advanced with AI and as you mentioned, perhaps we do become
00:19:22.960 smarter maybe through gene engineering and all sorts of other techniques
00:19:27.360 and advancements.
00:19:28.780 I think it's going to go back to the human condition and that's going to be
00:19:32.680 more impressive.
00:19:33.580 I think that's, I don't know.
00:19:34.860 That's just a guess.
00:19:35.960 What do you mean by human condition?
00:19:37.780 I mean, I mean, you know, the arts and empathy and just, you know, being
00:19:42.800 creative and, and creating something that's, that you're not going to be able
00:19:49.060 to get with just facts.
00:19:50.520 Right.
00:19:52.880 Yeah.
00:19:53.280 I guess in bodybuilding, they'll say, you know, this is the natural Mr.
00:19:58.780 Olympia competition, right?
00:20:00.320 They put nothing in their body and this is open.
00:20:03.100 You know, this guy is, you know, you can use anything and everything and he's
00:20:06.880 the Mr.
00:20:07.200 Olympia champion.
00:20:08.020 I wonder if we're going to get to the, even today, I think we have an Olympic
00:20:11.480 games today that they're talking about called the enhanced games or something.
00:20:16.020 I don't know.
00:20:16.340 Have you heard about the enhanced games?
00:20:17.460 It's literally called enhanced games.
00:20:20.180 And in this enhanced games, they encourage you to do PD.
00:20:24.680 Everybody's on PD.
00:20:25.600 And just let's find out what does it look like if we're freely able to use
00:20:29.420 everything.
00:20:29.840 Right.
00:20:30.460 So I wonder, I wonder if it's going to be like, you know, he naturally looks like
00:20:34.500 that.
00:20:34.860 That's natural beauty.
00:20:36.520 You know, she's done no surgeries.
00:20:39.000 He's done no surgeries.
00:20:40.600 He hasn't put the stuff in his brain to be interested in what he has to say or
00:20:45.300 all his podcasts are going to be what?
00:20:46.560 Then are we going to watch AI podcasts?
00:20:49.140 Is there going to be two personalities like Gemini's Mary is doing a podcast with
00:20:55.220 open AI's Bobby and they're debating politics and we can program one to be a
00:20:59.960 leftist, one to be a conservative and go ahead and debate this issue and state
00:21:03.460 the facts.
00:21:04.600 Will we be going there?
00:21:05.900 Are we going to be like, man, I would much rather talk to, listen to these guys
00:21:09.180 because they're regular human beings making mistakes.
00:21:11.100 And is it going to be attractive to be dumb?
00:21:13.640 Is it going to be attractive to make, you know, dumb comments?
00:21:16.020 I really wonder what's going to happen in 20, 30 years because it's going to be a
00:21:19.600 commodity if you're smart.
00:21:20.700 So what?
00:21:21.660 I can have the same thing you have.
00:21:22.880 I just put a chip in my head.
00:21:24.300 You're no longer competitive to me.
00:21:25.880 So why go study?
00:21:27.160 Like, you know, we would hear people say, I knew how to study for two days,
00:21:31.680 memorize everything, go ace the test and forgot all of it.
00:21:34.660 Right.
00:21:35.100 And that's what college is really all about.
00:21:36.760 But I really wonder what's going to happen and what will be turned, what will we be turned
00:21:41.940 on to?
00:21:42.920 Mistakes.
00:21:43.340 I love it.
00:21:44.060 We all make them.
00:21:45.080 I think so as well.
00:21:46.400 I think so as well.
00:21:47.520 Yesterday, I'm talking to a guy who's a very successful business guy.
00:21:51.020 He's telling me he's got a hundred million dollar EBIT on all this other stuff.
00:21:53.800 And he said, hey, how come my content is not, you know, doing better?
00:21:58.520 How come my podcast is not doing better?
00:22:00.140 And I watch him.
00:22:00.800 I'm like, okay, he's a billionaire.
00:22:02.000 He's done this.
00:22:02.540 He's done that.
00:22:02.980 He's done this.
00:22:03.540 So on everything, it looks good, bodybuilding, physique.
00:22:07.560 I said, you're too perfect.
00:22:10.100 I think the audience likes imperfection.
00:22:13.000 The audience likes to see flaws in people.
00:22:15.880 Passion.
00:22:16.680 Yeah.
00:22:17.180 Realness.
00:22:18.000 Vulnerability.
00:22:18.740 Sincerity.
00:22:20.060 I think, again, we don't know.
00:22:22.540 I think in 20 years, as this advances, we're going to want to see human error.
00:22:29.560 We're going to want to see that again.
00:22:31.620 Just like today, you see a lot of companies rely on AI on customer service.
00:22:35.460 Some companies are like, clients are like, I don't want to talk to a machine.
00:22:38.920 I want to talk to a human being.
00:22:40.520 I want to talk to somebody with service.
00:22:42.020 So anyways, we'll see what will happen with this.
00:22:43.760 I'm just wondering sometimes where you're at with it.
00:22:46.060 Yeah.
00:22:46.660 I feel the same way when I get an email in, and it's so obvious when it's AI generated.
00:22:50.720 I crave the real, authentic email, not AI generated, right?
00:22:56.200 Yeah.
00:22:56.460 When we set out to create a shoe that blends comfort, function, and luxury, we had the
00:23:02.760 choice to make it fast.
00:23:04.240 We had the choice to make it cheap.
00:23:06.400 We chose neither.
00:23:08.000 Instead, we chose Tuscany.
00:23:10.060 We chose true Italian craftsmanship.
00:23:12.960 Each pair touched by 50 skilled hands.
00:23:15.300 We chose patience, spending two years perfecting every detail, and we chose the finest quality
00:23:21.320 at every step.
00:23:23.460 Introducing the Future Looks Bright collection.
00:23:26.920 Not rushed.
00:23:28.340 Not disposable.
00:23:29.820 Not ordinary.
00:23:31.180 Rather intentional.
00:23:33.880 Luxurious.
00:23:35.180 Timeless.
00:23:35.620 If you enjoyed this video, you want to watch more videos like this, click here.
00:23:43.360 And if you want to watch the entire podcast, click here.
00:23:46.120 Thank you.