The Peter Attia Drive - July 15, 2024


#309 ‒ AI in medicine: its potential to revolutionize disease prediction, diagnosis, and outcomes, causes for concern in medicine and beyond, and more | Isaac Kohane, M.D., Ph.D.


Episode Stats

Length

1 hour and 55 minutes

Words per Minute

163.1663

Word Count

18,814

Sentence Count

1,084

Misogynist Sentences

2

Hate Speech Sentences

4


Summary

In this episode, Dr. Zach Cohen joins me to talk about the evolution of artificial intelligence (AI) from science fiction to real-world applications. We talk about where AI is today, where it is in the past, and where it could be in the future.


Transcript

00:00:00.000 Hey, everyone. Welcome to the Drive podcast. I'm your host, Peter Atiyah. This podcast,
00:00:16.540 my website, and my weekly newsletter all focus on the goal of translating the science of longevity
00:00:21.520 into something accessible for everyone. Our goal is to provide the best content in health and
00:00:26.720 wellness, and we've established a great team of analysts to make this happen. It is extremely
00:00:31.660 important to me to provide all of this content without relying on paid ads. To do this, our work
00:00:36.960 is made entirely possible by our members, and in return, we offer exclusive member-only content
00:00:42.700 and benefits above and beyond what is available for free. If you want to take your knowledge of
00:00:47.940 this space to the next level, it's our goal to ensure members get back much more than the price
00:00:53.200 of the subscription. If you want to learn more about the benefits of our premium membership,
00:00:57.980 head over to peteratiyahmd.com forward slash subscribe.
00:01:04.180 My guest this week is Isaac Cohen, who goes by Zach. Zach is a physician scientist and chair
00:01:10.180 of the Department of Biomedical Informatics at Harvard Medical School, and he's an associate
00:01:14.460 professor of medicine at the Brigham and Women's Hospital. Zach has published several hundred papers
00:01:18.940 in the medical literature and authored the widely used books, Microarrays for Integrative Genomics
00:01:24.040 and the AI Revolution in Medicine, GPT-4 and beyond. He is also the editor-in-chief of the newly launched
00:01:30.660 New England Journal of Medicine AI. In this episode, we talk about the evolution of AI. It wasn't really
00:01:37.280 clear to me until we did this interview that we're really in the third generation of AI, and Zach has been
00:01:41.780 a part of both the second and obviously the current generation. We talk about AI's abilities to impact
00:01:47.560 medicine today. In other words, where is it having an impact? And where will it have an impact in the
00:01:52.820 near term? What seems very likely? And of course, we talk about what the future can hold. And obviously
00:01:58.280 here, you're starting to think a little bit about the difference between science fiction and potentially
00:02:02.300 where we hope it could go. Very interesting podcast for me, really a topic I know so little about,
00:02:08.120 which tend to be some of my favorite episodes. So without further delay, please enjoy my conversation
00:02:13.100 with Zach Cohen. Well, Zach, thank you so much for joining me today. This is a topic that's highly
00:02:23.480 relevant and one that I've wanted to talk about for some time, but wasn't sure who to speak with. And
00:02:29.280 we eventually kind of found our way to you. So again, thanks for making the time and sharing your
00:02:33.580 expertise. Give folks a little bit of a sense of your background. What was your path through medical
00:02:38.780 school and training? It was not a very typical path. No. So what happened was I grew up in
00:02:44.760 Switzerland. Nobody in my family was a doctor, come to United States, decide to major in biology.
00:02:52.340 And then I get nerd sniped by computing back in the seventies and the late seventies. And so I minor in
00:02:59.900 computer science, but I still complete my degree in biology and I go to medical school. And then in the
00:03:06.820 middle of medical school's first year, I realized, holy smokes, this is not what I expected. It's a
00:03:12.380 noble profession, but it's not a science. It's an art. It's not a science. And I thought I was going
00:03:17.580 into science. And so I bail out for a while to do a PhD in computer science. And this is during the
00:03:24.960 1980s now, early 1980s. And it's a heyday of AI. It's actually a second heyday. We're going through the
00:03:32.660 third heyday. And it was a time of great promise. And with a retrospective scope, very clear that it
00:03:41.600 was not going to be successful. There was a lot of over-promising. There is today, but unlike today,
00:03:47.780 we had not released it to the public. It was not actually working in the way that we thought it
00:03:53.940 would go on at work. And it certainly didn't scale. It was a very interesting period. And
00:04:00.420 my thesis advisor, Peter Solovich, a professor at MIT, said, Zach, you should finish your clinical
00:04:06.620 training because I'm not getting a lot of respect from clinicians. And so to bring rational decision
00:04:13.280 making to the clinic, you really want to finish your clinical training. And so I finished medical
00:04:19.040 school, did a residency in pediatrics and pediatric endocrinology, which was actually extremely
00:04:25.180 enjoyable. But when I was done, I restarted my research in computing, started a lab at Children's
00:04:32.580 Hospital in Boston, and then a center of biomedical informatics at the medical school. Like in almost
00:04:39.840 every other endeavor, getting money gets attention from the powers that be. And so I was getting a lot
00:04:46.560 of grants. And so they asked me to start the center and then eventually a new department of
00:04:52.840 biomedical informatics that I'm the chair of. We have now 16 professors or assistant professors
00:04:58.900 of biomedical informatics. Then I had been involved in a lot of machine learning projects,
00:05:06.120 but like everybody else, I was taken by surprise, except perhaps a little bit earlier than most,
00:05:10.520 by large language models. I got a call from Peter Lee in October of 22. And actually I didn't get a
00:05:17.140 call. It was an email right out of a Michael Crichton novel. It said, Zach, if you'll answer
00:05:22.780 the phone, I can't tell you what it's about, but it'd be well worth your while. And so I get a call
00:05:29.440 from Peter Lee and I knew him from before. He was a professor of computer science at CMU and also
00:05:36.000 department chair there. And then he went to ARPA and then he went to Microsoft and he tells me about
00:05:40.800 GPT-4. And this was before any of us had heard about chat GPT, which is initially GPT-3.5. He tells
00:05:48.040 me about GPT-4 and he gets me early access to it when no one else knows that exists. Only a few
00:05:55.400 people do. And I start trying it against hard cases. I get called down. I just remember from my
00:06:03.200 training. I get called down to the nursery. It's a child with a small phallus and a hole at the base
00:06:10.540 of the phallus and they can't palpate testicles and they want to know what to do because I'm a
00:06:16.560 pediatric endocrinologist. And so I asked GPT-4, what would you do? What are you thinking about?
00:06:22.940 And it runs me through the whole workup of these very rare cases of ambiguous genitalia. In this case,
00:06:30.860 it was congenital adrenal hyperplasia where the making of excess androgens during pregnancy and
00:06:38.420 then subsequently in birth causes the clitoris to swell from the glands of the penis of the phallus
00:06:46.360 and the labia minora to fuse to form the shaft of what looks like a penis. But there's no testicles,
00:06:55.920 there's ovaries. And so there's a whole endocrine workup with genetic tests, hormonal tests,
00:07:04.620 ultrasound, and it does it all. And it blows my mind. It really blows my mind because very few of us in
00:07:12.720 computer science really thought that these large language models would scale up the way they do.
00:07:18.420 It was just not expected. And talking to Bill Gates about this after Peter Leon introduced me to
00:07:26.500 problem, and he told me that his line engineers in Microsoft research, a lot of his fanciest computer
00:07:32.160 scientists did not expect this. But the line engineers at Microsoft were just watching the
00:07:37.660 scale up, you know, GPT 0, 1, 2, and they just saw it was going to keep on scaling up with the size of
00:07:44.600 the data and with the size of the model. And they said, yeah, of course, it's going to achieve this
00:07:49.520 kind of expertise. But the rest of us, I think because we value our own intellects so much, we
00:07:56.160 couldn't imagine how we'd get that kind of conversational expertise just by scaling up the
00:08:03.340 model and the data set. Well, Zach, that's actually kind of a perfect introduction to how I want to
00:08:09.460 think about this today, which is to say, look, there's nobody listening to us who hasn't heard
00:08:13.700 the term AI, and yet virtually no one really understands what is going on. So if we want to
00:08:19.280 talk about how AI can change medicine, I think we have to first invest some serious bandwidth in
00:08:26.200 understanding AI. Now, you alluded to the fact that when you were doing your PhD in the early 80s,
00:08:30.240 you were in the second generation of AI, which leads me to assume that the first generation was
00:08:36.460 shortly following World War II. And that's probably why someone by the name of Alan Turing
00:08:41.820 has his name on something called the Turing test. So maybe you can talk us through what Alan Turing
00:08:48.300 posited, what the Turing test was and proposed to be, and really what Gen 1 AI was. We don't have to
00:08:54.740 spend too much time on it, but clearly it didn't work. But let's maybe talk a little bit about
00:08:58.740 the postulates around it and what it was. After World War II, we had computing machines.
00:09:07.000 And anybody who was a serious computer scientist could see that you could have these processes
00:09:13.920 that could generate other processes. And you could see how these processes could take inputs
00:09:19.800 and become more sophisticated. And as a result, shortly after World War II, we actually had artificial
00:09:30.960 neural networks, the perceptron, which was modeled, roughly speaking, on the ideas of a neuron that could
00:09:40.740 take inputs from the environment and then have certain expectations.
00:09:47.120 And if you updated the neurons as to what was going on, it would update the weights going into
00:09:55.640 that artificial neuron. And so going back to Turing, he just came up with a test that said,
00:10:03.420 essentially, if a computational entity could maintain, essentially, its side of the conversation
00:10:10.820 without revealing that it was a computer and that others would mistake it for a human, then for all
00:10:19.780 intents and purposes, that would be intelligent behavior. And there's been all sorts of additional
00:10:26.900 constraints put on it. And one of the hallmarks of AI, frankly, is that it keeps on moving the goalposts
00:10:34.840 of what we consider to be intelligent behavior. If you had told someone in the 60s that the world chess
00:10:43.360 masters were going to be beaten by a computer program, they'd say, well, that's AI, really, that's AI.
00:10:49.820 And then when Kasparov was beaten by Deep Blue, by the IBM machine, people said, well, it's just doing
00:10:57.260 search very well. It's searching through all the possible moves in the future. It also knows all the
00:11:02.540 grandmaster moves. It has a huge encyclopedia store of all the different grandmaster moves.
00:11:07.900 And this is not really intelligent behavior. If you told people it could recognize human faces
00:11:14.640 and find your grandmother in a picture on any picture in the internet, they'd say, well, that's
00:11:20.920 intelligence. And of course, when we did it, no, that was not intelligent. And then when we said it
00:11:28.280 could write a rap poem about Peter Atia based on your webpage, and it did that, well, that would be
00:11:37.700 intelligent, that would be creative. But then if you said it's doing it based on having created a
00:11:44.220 computational model based on all the text ever generated by human beings, as much as we can gather,
00:11:50.940 which is one to six terabytes of data. And this computational model basically is predicting what is
00:11:57.860 the next word that's going to say, not just the next word, but of the millions of words that could be,
00:12:02.880 what are the probabilities of that next word? That is what's generating that rap. There's
00:12:08.720 people who are arguing that's not intelligence. So the goalposts around the Turing test keep getting
00:12:14.980 moved. So I just have to say that I no longer find that an interesting topic because it's what it's
00:12:22.000 actually doing. And whether you want to call it intelligent or not, that's up to you. It's like
00:12:26.480 discussing whether is a dog intelligent, is a baby intelligent before it can recognize
00:12:33.620 constancy of objects. Initially, babies, if you hide something from it, it's gone and it comes back.
00:12:39.880 It's a surprise. But at some point early on, they learn there's constancy of objects, even when they
00:12:44.720 don't see them. There's this spectrum of intelligent behavior. And I'd just like to remind myself that
00:12:53.780 there's a very simple computational model predicting the next word called a Markov model.
00:13:00.260 And several years ago, people were studying songbirds, and they were able to predict the full
00:13:07.000 song, the next note, and the next note of the songbird, just using a very simple Markov model.
00:13:14.240 So from that perspective, I know we think that we're all very smart, but the fact that you and I,
00:13:20.680 without thinking too hard about it, can come up with fluid speech. Okay, so the model's now
00:13:25.000 a trillion parameters. It's not a simple Markov model, but it's still a model. And perhaps later,
00:13:31.140 we'll talk about how this plays into, unfortunately, the late Kahneman's notions of thinking fast and
00:13:37.820 thinking slow. And his notion of system one, which is this sort of pattern recognition,
00:13:42.100 which is very much similar to what I think we're seeing here. And system two, which is the more
00:13:46.920 deliberate and much more conscious kind of thinking that we pride ourselves on. But a lot of what we
00:13:53.200 do is this sort of reflexive, very fast pattern recognition.
00:13:58.840 So if we go back to World War II, that's to your point where we saw basically rule-based computing
00:14:05.380 come of age. And anybody who's gone back and watched movies about the Manhattan Project,
00:14:11.140 or the decoding of all the sorts of things took place, Enigma, for example. Again, that's straight
00:14:18.780 rules-based computational power. And we're obviously at the limits of, I can only go so far. But it seems
00:14:25.420 that there was a long hiatus before we went from there to kind of like maybe what some have called
00:14:32.620 context-based computation, what your Siri does or Alexa, which is a step quite beyond that. And then
00:14:42.300 of course, you would go from there to what you've already talked about, Blue or Watson, where you have
00:14:48.640 computers that are probably going even one step further. And then of course, where we are now,
00:14:54.880 which is GPT-4. I want to talk a little bit about the computational side of that. But more what I want
00:15:00.860 to get at is this idea that there seems to be a very non-linear pace at which this is happening.
00:15:07.600 And I hear your point. I'd never thought of it that way. I hear your point about the goalpost moving,
00:15:12.840 but I think your instinct around majoring in the right thing is also relevant, which is let's focus
00:15:20.780 less on the fact that we're never quite hitting the asymptote definitionally. Let's look at the actual
00:15:26.380 output and it is staggeringly different. So what was it that was taking place during the period of your PhD,
00:15:33.200 what you're calling wave two of AI? What was the objective and where was the failure?
00:15:38.800 So the objective was in the first era, you wrote computer programs in assembler language or in languages
00:15:45.960 like Fortran. And there was a limit of what you could do. You had to be a real computational
00:15:52.580 programmer to do something in that mode. In wave two in the 1970s, we came up with these rule-based
00:16:01.700 systems where we said rules in what looked like English. If there is a patient who has a fever
00:16:08.400 and you get an isolate from the lab and that bacteria in the isolate is gram-positive,
00:16:14.700 then you might have a streptococcal infection with a probability of so-and-so. And these rule-based
00:16:21.580 systems, which you're now programming in the level of human knowledge, not in computer code,
00:16:28.600 the problem with that was several fold. A, you're going to generate tens of thousands of these rules
00:16:33.720 rules. And these rules would interact in ways that you could not anticipate. And we did not know
00:16:39.460 enough. And we could not pull out of human beings the right probabilities. And what is the right
00:16:46.460 probability of you have a fever and you don't see anything on the blood test? What else is going on?
00:16:54.920 And there's a large set of possibilities. And getting all those rules out of human beings ended up
00:16:59.580 being extremely expensive and the results were not stable. And for that reason, because we didn't
00:17:06.280 have much data online, we could not go to the next step, which is have data to actually drive these
00:17:13.340 models. What were the data sources then? Books, textbooks, and journals as interpreted by human
00:17:21.460 experts. That's why some of these were called expert systems, because they were derived from
00:17:26.760 introspection by experts who would then come up with the rules, with the probabilities. And some of the
00:17:35.100 early work, like for example, there was a program called Mycin run by Ted Shortliff out of Stanford, who
00:17:41.540 developed a antibiotic advisor that was a set of rules based on what he and his colleagues sussed out from
00:17:52.280 the different infectious disease textbooks and infectious disease experts. And it stayed only up
00:17:58.820 to date as long as they kept on looking at the literature, adding rules, fine-tuning it. There's
00:18:05.060 an interaction between two rules that was not desirable. Then you had to adjust that. Very labor-intensive.
00:18:12.120 And then if there's a new thing, you'd have to add some new rules. If AIDS happened, you'd have to say,
00:18:19.180 oh, there's this new pathogen. I have to make a bunch of rules. The probability is going to be
00:18:25.200 different if you're an IV drug abuser or if you're a male, a homosexual. And so it was very, very hard
00:18:32.040 to keep up. And in fact, people didn't. What was the language that it was programmed in? Was this
00:18:37.000 Fortran? No, no. These were so-called rule-based systems. And so the languages, for example,
00:18:42.720 system mycin was called e-mycin, essential mycin. So these looked like English.
00:18:48.740 Super labor-intensive.
00:18:50.300 Super labor-intensive. And there's no way you could keep it up to date. And at that time,
00:18:56.060 there was no electronic medical records. They were all paper records. So not informed by what was
00:18:59.940 going on in the clinic. Three revolutions had to happen in order for us to have what we have today.
00:19:08.120 And that's why I think we had such a quantum jump recently.
00:19:13.300 Before we get to that, that's the exciting question, but I just want to go back to the
00:19:16.820 Gen 2. Were there other industries that were having more success than medicine? Were there
00:19:21.660 applications in the military? Were there applications elsewhere in government where
00:19:25.620 they got a little closer to utility?
00:19:28.400 Yes. So there was a company which was a remnant of... Back in the 1970s, there were a whole bunch of
00:19:35.900 computer companies around what we called 128 in Boston. And these were companies that were famous
00:19:42.280 back then, like Wang Computer, like Digital Equipment Corporation. And it's a very sad story for Boston
00:19:49.260 because that was before Silicon Valley got its pearl of computer companies around it. And one of the
00:19:57.560 companies, Digital Equipment Corporation, built a program called R1. And R1 was an expert in configuring
00:20:05.380 the mini computers that you ordered. So you wanted some capabilities and it would actually
00:20:09.760 configure all the industrial components, the processors, the disk, and it would know about
00:20:16.520 all the exceptions and what you needed to know, what cabling, what memory configuration, all that was
00:20:22.520 done. And it basically replaced several individuals who had that very, very rare knowledge to configure
00:20:29.720 their systems. It was also used in several government logistics efforts. But even those efforts, although
00:20:37.740 they were successful and used commercially, were limited because it turns out human beings, once you
00:20:43.160 got to about three, four, five, six thousand rules, no single human being could keep track of all the ways
00:20:50.480 these rules could work. We used to call this the complexity barrier, that these rules would interact in
00:20:57.520 unexpected ways and you'd get incorrect answers, things that were not commonsensical because you'd
00:21:07.220 actually not captured everything about the real world. And so it was very narrowly focused. And if the
00:21:15.380 expertise was a little bit outside the area of focus, if let's say it was an infectious disease program
00:21:20.960 and there was a little bit of influence from the cardiac status of the patient and you had not
00:21:28.020 accurately modeled that, its performance would degrade rapidly. Similarly, if there was in digital
00:21:35.220 equipment a new model that had a complete different part that had not included and that there were some
00:21:43.200 dependencies that were not modeled, it would degrade in performance. So these systems were very brittle,
00:21:48.460 did not show common sense. They had expert behavior, but it was very narrowly done. There were applications
00:21:55.600 of medicine back then that survived till today. For example, already back then we had these systems
00:22:02.040 doing interpretation of ECGs pretty competently, at least a first pass until they would be reviewed by
00:22:10.320 an expert cardiologist. There's also a program that interpreted what's called serum protein
00:22:15.540 electrophoresis, where you look at proteins separated out by an electric gradient to make a diagnosis,
00:22:23.260 let's say of myeloma or other protein disorders. And those also were deployed clinically, but they only
00:22:30.920 worked very much in narrow areas. They were by no stretch of an imagination, general purpose reasoning
00:22:37.340 machines.
00:22:37.900 So let's get back to the three things. There are three things that have taken the relative failures of first and second
00:22:46.000 attempts at AI and got us to where we are today. I can guess what they are, but let's just have you walk us through
00:22:52.880 them.
00:22:53.540 The first one was just lots of data. And we needed to have a lot of online data to be able to develop models of
00:23:03.920 interesting performance and quality. So ImageNet was one of the first such data sets, collections of millions of
00:23:13.080 images with annotations, importantly. This has a cat in it. This has a dog in it. This is a blueberry muffin. This has a
00:23:20.320 human in it. And having that was absolutely essential to allow us to train the first very successful neural network
00:23:30.320 models. And so having those large data sets was extremely important. The other, and there's equivalent in
00:23:39.760 medicine, which is we did not have a lot of textual information about medicine until PubMed went online. So all the
00:23:50.520 literature, medical literature, at least we have an abstract of it in PubMed. Plus we have for a subset of it that's open
00:23:58.380 access because government has paid for it through grants. There's something called PubMed Central, which has the full
00:24:04.100 text. So all of a sudden that has opened up over the last 10 years. And then electronic health records, after Obama
00:24:12.660 signed the HITECH Act, electronic health records, which also ruined the lives of many doctors, also happened to also
00:24:19.540 generate a lot of text for the use in these systems. So that's large amounts of data being generated online. The second was
00:24:28.160 the neural network models themselves. So the perceptron that I mentioned that was developed
00:24:33.240 not too long after World War II was shown by one of the pioneers of AI, Marvin Minsky, to have fundamental
00:24:41.700 limitations in that it could not do certain mathematical functions like what's called an exclusive
00:24:47.760 ore gate. Because of that, people said these neural networks are not going to scale. But there were a few true
00:24:53.360 believers who kept on pushing and making more and more advanced architectures and those multi-level deep
00:25:03.060 neural networks. So instead of having one neural network, we layer on top of one neural network, another
00:25:09.860 one, and another one, and another one, so that the output of the first layer gets propagated up to the
00:25:16.540 second layer of neurons to the third layer and fourth layer and so on.
00:25:20.780 And I'm sorry, was this a theoretical mathematical breakthrough or a technological breakthrough?
00:25:26.420 Both. It was both because having those insight that these, we could actually come up with all the
00:25:33.540 mathematical functions that we needed to, we could simulate them with these multi-level networks,
00:25:38.440 whereas it was a theoretical insight, but we would have never made anything out of it if not for the
00:25:43.880 fact of sweaty teenagers, mostly teenage boys, playing video games. In order to have first-person shooters
00:25:53.120 capable of running high-resolution pictures of aliens or monsters in high-resolution, 24-bit color,
00:26:04.900 60 frames per second, we needed to have processors, very parallel processors,
00:26:10.700 that would allow you to do the linear algebra that allow you to calculate what was going to be the
00:26:17.580 intensity of color on every dot of the screen at 60 frames per second.
00:26:22.320 And that's literally just because of the matrix multiplication
00:26:25.080 math that's required to do this. You have N by M matrices that are so big, and you're crossing and
00:26:33.040 dotting huge matrices.
00:26:34.940 Huge matrices. And it turns out that's something that can be run in parallel. So you want to have
00:26:40.940 multiple parallel processors capable of rendering those images, again, at 60 frames per second. So
00:26:47.840 basically, millions of bits on your screen being rendered at 24 or 32-bit color. And in order to do
00:26:55.180 that, you need to have that linear algebra that you just referred to being run in parallel.
00:26:59.460 And so these parallel processors called graphical processing units, GPUs, were developed. And the GPUs
00:27:09.780 were developed by several companies. And some of them stayed in business, some didn't, but they were
00:27:14.960 aptly essential to the success of video games. Now, it then occurred to many smart mathematicians and
00:27:21.900 computer scientists that the same linear algebra that was used to drive that computation for images
00:27:29.320 could also be used to calculate the weights of the edges between the neurons in a neural network.
00:27:37.860 So the mathematics of updating the weights in response to stimuli, let's say, of a neural network,
00:27:45.520 updating of those weights can be done all in linear algebra. And if you have this processor, so a typical
00:27:55.140 computer has a central processing unit. So that's one processing unit. A GPU has tens of thousands of
00:28:04.840 processors that do this one very simple thing, linear algebra. And so by having this parallelism that
00:28:12.420 only supercomputers would have typically on your simple PC, because you needed to show the graphics
00:28:20.260 at 60 frames per second, gave us all of a sudden these commodity chips that allowed us to calculate
00:28:26.460 the performance of these multi-level neural networks. So that theoretical breakthrough was the second
00:28:31.740 part, but would not have happened without the actual implementation capability that we had with the GPUs.
00:28:40.600 And so NVIDIA would be the most successful example of this, presumably?
00:28:45.680 It was not the first, but it's definitely the most successful example. And there's a variety of
00:28:49.820 reasons why it was successful and created an ecosystem of implementers who built their neural
00:28:56.380 network deep learning systems on top of the NVIDIA architecture.
00:29:02.000 Would you go back and look at the calendar and say this was the year or quarter when there was
00:29:06.500 escape velocity achieved there? Yeah. So it was probably around 2012 when there was an ongoing
00:29:13.340 contest every year saying who has the best image recognition software. And these deep neural networks
00:29:22.060 running off GPUs were able to outperform significantly all their other competitors
00:29:29.400 in image recognition in 2012. That's very clearly when everybody just woke up and said, whoa, we knew about
00:29:36.560 neural networks. We didn't realize that these convolutional neural networks were going to be
00:29:41.980 this effective. And it seems that the only thing that's going to stop us is computational speed and the size
00:29:50.440 of our data sets. That moved things very fast along in the imaging space with very soon consequences in
00:29:59.620 medicine. It was only six years later that we saw journal articles about recognition of retinopathy,
00:30:08.380 diseases affecting the retina, the back of your eye and diabetes. And a paper coming out of all places
00:30:15.360 from Google saying we can recognize different stages of retinopathy based on the images of the back of
00:30:23.020 the eye. And that also was a wake up call because yes, part of the goalpost moving is great that we
00:30:28.660 could recognize cats and dogs in web pages. But now all of a sudden, this thing that we thought was
00:30:35.880 specialized human expertise could be done by that same stack of software. Just if you gave it enough
00:30:43.380 cases of these retinopathies, it would actually work well. And furthermore, what was wild was that
00:30:50.120 there's something called transfer learning, where you tune up these networks, get them to recognize
00:30:55.800 cats and dogs. And in the process of recognizing cats and dogs, it learns how to recognize little
00:31:01.040 circles and lines and fuzziness and so on. You did a lot better in training up the neural network
00:31:08.560 first on the entire set of images and then on the retinas. And if you just went straight to,
00:31:15.180 I'm just going to train on the retinas. And so that transfer learning was impressive.
00:31:20.980 And then the other thing as a doctor was impressive to many of us. I was actually asked to write an
00:31:26.200 editorial for the Journal of the American Medical Association in 2018 when a Google article was
00:31:32.560 written. What was impressive to us was that what was the main role of doctors in that publication?
00:31:39.120 It was just twofold. One was to just label the images that were used for training. This is
00:31:45.700 a retinopathy. It's not retinopathy. And then to serve as judges of its performance. And that was it.
00:31:53.580 The rest of it was computer scientists working with GPUs and images, tuning it. And that was it.
00:32:00.480 Didn't look anything like medical school. And you were having expert level recognition of
00:32:05.840 retinopathy. That was a wake-up call. You've alluded to the 2017 paper by Google,
00:32:14.080 Attention is All That is Needed, I think is the title of the paper. Attention is All You Need.
00:32:19.060 That's not what I'm referring to. I'm also referring to a 2018 paper in JAMA.
00:32:23.360 Ah, I'm sorry.
00:32:24.700 You're talking about the great paper, Attention is All You Need. That was about the invention of the
00:32:29.320 transformer, which is a specific type of neural network architecture. I was talking about these
00:32:35.540 were vanilla, fairly vanilla convolutional neural networks, the same one that can detect dogs and
00:32:41.560 cats. It was a big medical application, retinopathy 2018. Except for computer scientists, no one noticed
00:32:47.940 the attention is all you need paper. And Google had this wonderful paper that said, you know,
00:32:56.520 if we recognize not just text that co-locates together, because previously, so we're going to
00:33:03.740 get away from images for a second. There was this notion that I can recognize a lot of similarities
00:33:11.520 in text. If I see which words occur together, I can implicate the meaning of a word by the company
00:33:18.780 it keeps. And so if I see this word and it has around it, kingdom, crown, throne, it's about a king.
00:33:30.080 And similarly for queen and so on. That kind of association in which we created what was called
00:33:37.420 embedding vectors, which just in plain English, it's a string of numbers that says for any given word,
00:33:45.920 what's the probability? How often do these other words co-occur with it? And just using those
00:33:51.960 embeddings, those vectors, those lists of numbers that describe the co-occurrence of other words,
00:33:59.700 we were able to do a lot of what's called natural language processing, which you're looking at text
00:34:04.600 and saying, this is what it means. This is what's going on. But then in the 2017 paper,
00:34:11.220 they actually took a next step, which was the insight that where exactly the thing that we
00:34:19.380 were focusing on was in the sentence, what was before and after the actual ordering of it
00:34:24.980 mattered, not just the simple co-occurrence, that knowing what position that word was in the sentence
00:34:32.480 actually made a difference. That paper showed the performance went way up in terms of recognition.
00:34:40.340 And that transformer architecture that came from that paper made it clear for a number of
00:34:47.780 researchers, not me, that if you scaled that transformer architecture up to a larger model
00:34:54.680 so that the position dependence and this vector was learned across many, many more texts,
00:35:03.120 the whole internet, you could train it to do various tasks. This transformer model, which is called
00:35:08.200 the pre-trained model. So I apologize, I find it very boring to talk about because unless I'm working
00:35:14.080 with fellow nerds, this transformer, this pre-trained model, can think of it as the equivalent of an
00:35:20.180 equation with multiple variables. In the case of GPT-4, we think it's about a trillion variables.
00:35:27.540 It's like an equation where you have a number in front of each variable, a coefficient,
00:35:31.840 that's about a trillion long. And this model can be used for various purposes. One is the chatbot
00:35:41.900 purpose, which is given this sequence of words, what is the next word that's going to be said?
00:35:47.860 Now, that's not the only thing you could use this model for, but that's, turns out to have been
00:35:53.040 the breakthrough application of the transformer model for text.
00:35:57.460 Just to round out what you said earlier, Zach, would you say that is the third thing that enabled
00:36:02.300 this third wave of AI, the transformer? It was not what I was thinking about. For me,
00:36:07.400 I was thinking of the real breakthrough in data-driven AI. I put around the 2012 era. This is
00:36:13.520 yet another, if you talk to me in 2018, I would have already told you we're in a new heyday and
00:36:20.340 everybody would agree with you. There was a lot of excitement about AI just because of the image
00:36:25.300 recognition capabilities. This was an additional capability that's beyond what many of us were
00:36:32.740 expecting just from the scale-up of the neural network. The three, just to make sure I'm consistent,
00:36:39.280 was large data sets, multi-level neural networks, aka deep neural networks, and the GPU infrastructure.
00:36:46.440 That brought us well through the 2012 to 2018. The 2017 blip that became what we now know to be
00:36:58.760 this whole large language model transformer architecture, that development, unanticipated
00:37:04.920 for many of us, but that was already on the heels of a ascendant AI era. There was already billions of
00:37:10.760 dollars of frothy investment in frothy companies, some of which did well and many of which did not
00:37:17.800 do so well. The transformer architecture has revolutionized many parts of the human condition,
00:37:24.520 I think, but it was already part of it. I think the third wave. There's something about GPT where I feel
00:37:34.040 like most people by the time GPT-3 came out or certainly by 3.5, this was now outside of the
00:37:41.880 purview of computer scientists, people in the industry who were investing in it. This was now
00:37:48.920 becoming as much a verb as Google was in probably the early 2000s. There were clearly people who knew
00:37:57.580 what Google was in 96 and 97, but by 2000, everybody knew what Google was, right? Something about GPT 3.5
00:38:05.900 or 4 was kind of the tipping point where I don't think you can not know what it is at this point.
00:38:11.900 I don't know if that's relevant to the story, meaning does that sort of speak to what trajectory
00:38:19.180 we're on now? The other thing that I think, Zach, has become so audible in the past year
00:38:26.140 is the elevation in the discussion of how to regulate this thing, which seems like something
00:38:34.540 you would only argue about if you felt that there were a chance for this thing to be harmful to us
00:38:41.980 in some way that we do not yet perceive. So what can you say about that? Because that's obviously a nod
00:38:48.380 to the technical evolution of AI, that very serious people are having discussions about
00:38:56.620 pausing, moratoriums, regulations. There was no public discussion of that in the 80s,
00:39:02.060 which may have spoke to the fact that in the 80s, it just wasn't powerful enough to pose a threat.
00:39:06.620 So can you maybe give us a sense of what people are debating now? What is the smart, sensible,
00:39:13.420 reasonable argument on both sides of this? And let's just have you decide what the two sides are.
00:39:19.340 I'm assuming one side says, pedal to the metal. Let's go forth on development. Don't regulate this.
00:39:25.260 Let's just go nuts. The other side is, no, we need to have some breaks and barriers.
00:39:30.300 Not quite that. So you're absolutely right that chatbots have now become a commonly used noun. And that
00:39:36.940 probably happened with the emergence of GPT 3.5. And that appeared around, I think, December of 2022.
00:39:44.700 But now, yes, because out of the box, that pre-trained model I told you about could tell
00:39:51.020 you things like, how do I kill myself? How do I manufacture a toxin? It could allow you to do a lot
00:39:58.300 of harmful things. So there was that level of concern. We can talk about what's been done about
00:40:04.860 those first order efforts. Then there's been a group of scientists who interestingly went from
00:40:14.300 saying, we'll never actually get general intelligence from this particular architecture to saying, oh my
00:40:21.420 gosh, this technology is able to inference in a way that I had not anticipated. And now I'm so worried
00:40:30.300 that either because it's malevolent or just because it's trying to do something that has bad side effects
00:40:37.020 for humanity, it presents an existential threat. Now, on the other side, I don't believe is anybody saying,
00:40:44.220 let's just go heads down and let's see how fast we can get to artificial general intelligence.
00:40:51.020 Or if they do think that, they're not saying it openly.
00:40:54.060 Can you just define AGI, Zach? I think we've all heard the term, but is there a quasi-accepted
00:41:00.060 definition? First of all, there's not. And I hate myself for even bringing it up because it starts-
00:41:05.100 I was going to bring it up before you, anyway, it was inevitable.
00:41:08.380 That was an unfortunate slip because artificial general intelligence means a lot of things to a
00:41:13.580 lot of people. And I slipped because I think it's, again, a moving target and it's very much
00:41:19.820 eye on the beholder. There's a guy called Eliezer Yudkowsky, one of the so-called doomers.
00:41:24.620 And he comes up with great scenarios of how a sufficiently intelligent system could figure out
00:41:33.340 how to persuade human beings to do bad things or control of our infrastructure to bring down our
00:41:40.780 communications infrastructure or airplanes out of the sky. And we can talk about whether that's
00:41:45.580 relevant or not. And on the other side, we have, let's say, OpenAI and Google.
00:41:50.780 But what was fascinating to me is that OpenAI, which working with Microsoft generated GPT-4,
00:41:58.060 we're not saying publicly at all, let's not regulate it. In fact, they were saying,
00:42:03.100 please regulate me. Sam Altman went on a world tour where he said, we should be very concerned about
00:42:08.540 this. We should regulate AI. And he was before Congress saying, we should regulate AI. And so,
00:42:15.980 I feel a bit churlish about saying this because Sam was kind enough to write forward to the book I
00:42:21.260 wrote with Peter Lee and Kerry Goldberg on GPT-4 and the revolution in medicine. But I was wondering,
00:42:29.580 why were they insisting so much on regulation? And there's two interpretations. One is just a sincere,
00:42:35.820 and it could very well be that. Sincere wish that it be regulated so we check these machines,
00:42:41.500 these programs to make sure they don't actually do anything harmful. The other possibility,
00:42:46.060 unfortunately, is something called regulatory locking, which means I'm a very well-funded company,
00:42:51.820 and we're going to create regulations with Congress about what is required, which boxes do you have to
00:42:57.660 check in order to be allowed to run. If you're a small company, you're not going to have a
00:43:03.660 bevy of lawyers with big checks to comply with all the regulatory requirements. And so, I think Sam is,
00:43:13.180 I don't know him personally, I imagine he's a very well-motivated individual. But whether it's for
00:43:19.100 the reason of regulatory lock-in or for genuine concern, there has not been any statements of,
00:43:26.460 let's go heads down. They do say, let's be regulated. Now, having said that, before you even
00:43:32.780 go with a doomer scenario, I think there is someone just as potentially evil that we have to worry about,
00:43:38.380 another intelligence, and that's human beings. And how do human beings use these great tools?
00:43:44.540 So, just as we know for a fact that one of the earliest users of GPT-4 were high schoolers trying
00:43:52.940 to do their homework and solve hard puzzles given to them, we also know that various parties have
00:43:59.740 been using the amazing text generation and interactive capabilities of these programs to
00:44:06.380 spread misinformation, to chatbots, and there's a variety of malign things that could be done by
00:44:13.580 third parties using these engines. And I think that's, for me, the clear and present danger today,
00:44:20.140 which is how do individuals decide to use these general purpose programs?
00:44:27.180 If you look at what's going on in the Ukraine-Russian war, I see more and more autonomous
00:44:34.380 vehicles flying and carrying weaponry and dropping bombs. And we see in our own military a lot more
00:44:44.620 autonomous drones with greater and greater autonomous capabilities. Those are purpose-built
00:44:51.980 to actually do dangerous things. And a lot of science fiction fans will refer to Skynet from the
00:45:01.820 Terminator series, but we're literally building it right now.
00:45:05.260 In the Terminator, Zach, they kind of refer to a moment, I don't remember the year, like 1997 or
00:45:11.900 something. And I think they talk about how Skynet became, quote, self-aware. And somehow when it became
00:45:17.340 self-aware, it just decided to destroy humans. Is self-aware movie speak for AGI? Like, what do you
00:45:25.100 think self-aware means in more technical terms? Or is it super intelligence? There's so many terms here,
00:45:32.300 and I don't know what they mean. Okay. So self-awareness means a process by which the
00:45:38.620 intelligent entity can look back, look inwardly at its own processes and recognize itself. Now, that's
00:45:46.060 very hand-wavy, but Douglas Hofstra has probably done the most thoughtful and clear writing about what
00:45:56.140 self-awareness means. I will not do it justice, but if you really want to read a wonderful book that
00:46:02.460 spends a whole book trying to explain it, it's called I Am A Strange Loop. And in I Am A Strange
00:46:08.540 Loop, he explains how if you have enough processing power and you can represent the processes that you
00:46:17.660 have essentially models of the processes that constitute you. In other words, you're able to look at what
00:46:23.340 you're thinking. You may have some sense of self-awareness. There's a bit of an act of faith
00:46:28.300 on that. Many AI researchers don't buy that definition. There's a difference between self-awareness
00:46:35.500 and actual raw intelligence. You can imagine a super powerful computer that would predict everything
00:46:43.740 that was going to happen around you and was not aware of itself as an entity. The fact remains,
00:46:49.180 you do need to have a minimal level of intelligence to be able to be self-aware. So a fly may not be
00:46:56.540 self-aware. It just goes and finds good-smelling poop and does whatever it's programmed to do on that.
00:47:04.220 But dogs have some self-awareness and awareness of their surroundings. They don't have perfect
00:47:11.900 self-awareness, like they don't recognize themselves in the mirror and they'll bark at that. Birds will
00:47:17.180 recognize themselves in mirrors. We recognize ourselves in many, many ways. So there is some
00:47:23.580 correlation between intelligence and self-awareness, but these are not necessarily dependent functions.
00:47:28.700 So what I'm hearing you say is, look, there are clear and present dangers associated with current
00:47:34.780 best AI tools in that humans can use them for nefarious purposes. It seems to me that the most scalable
00:47:43.100 example of that is still relatively small in that it's not existential threat to our species large,
00:47:50.780 correct? Well, yes and no. If I was trying to do gain of function research with a virus,
00:47:58.140 good point, I could use these tools very effectively. Yeah. That's a great example.
00:48:03.980 There's this disconnect and perhaps you understand the disconnect better than I do.
00:48:07.580 There's those real existential threats. And then there's this more fuzzy thing that we're worried
00:48:14.780 about correctly about bias, incorrect decisions, hallucinations. We can get into what that might be
00:48:22.140 and our use in the everyday of human condition. And there's concerns about mistakes that might be
00:48:28.220 made. There's concerns about displacement of workers that just as automation displaced a whole other series
00:48:37.420 of workers. Now that we have something that works in the knowledge industry automatically, just as
00:48:43.660 we're placing a lot of copy editors and illustrators with AI, where's that going to stop? It's now much
00:48:50.700 more in the white collar space. And so there is concern around the harm that could be generated there.
00:48:57.260 In the medical domain, are we getting good advice? Are we getting bad advice? Whose interests are being
00:49:02.940 optimized in these various decision procedures? That's another level that doesn't quite rise at
00:49:08.540 all to the level of extinction events. But a lot of policymakers and the public seem to be concerned
00:49:14.860 about it. Those are fair points. Let's now talk about that state of play within medicine. So I liked
00:49:20.700 your first example, almost one we take for granted, but you go and get an EKG at the doctor's office. This was
00:49:25.820 true 30 years ago, just as it is today. You get a pretty darn good readout. It's going to tell you if
00:49:31.500 you have an AV block. It's going to tell you if you have a bundle branch block. Put it this way,
00:49:36.540 they read EKGs better than I do. That's not saying much anymore, but they do. What was the next area
00:49:42.300 where we could see this? It seems to me that radiology is a field of medicine, which is of course,
00:49:48.940 image pixel based medicine that would be the most logical next place to see AI do good work. What
00:50:00.140 is the current state of AI in radiology? In all the visual based medical specialties,
00:50:07.740 it looks like AI can do as well as many experts. So what are the image appreciation subspecialties?
00:50:19.020 Pathology, when you're looking at slices of tissue under the microscope. Radiology,
00:50:23.020 where you're looking at x-rays or MRIs. Dermatology, where you're looking at pictures of the skin.
00:50:31.580 So in all those visual based specialties, the computer programs are doing by themselves as well
00:50:42.300 as many experts, but they're not replacing the doctors because that image recognition process
00:50:50.460 is only part of their job. Now, to be fair to your point, in radiology, we already today,
00:50:58.220 before AI in many hospitals would send x-rays by satellite to Australia or India where they would
00:51:06.220 be read overnight by a doctor or a specially trained person who had never seen the patient.
00:51:12.060 And then the reports filed back to us because they're 12 hours away from us overnight, we'd have
00:51:17.660 the results of those reads. And that same kind of function can be done automatically by AI. So that's
00:51:23.260 replacing a certain kind of doctor.
00:51:24.140 Let me dig into that a little bit more. So let's start with a relatively simple
00:51:29.740 type of image, such as a mammogram or a chest x-ray. So it's a single image. I mean, I guess
00:51:35.980 with a chest x-ray, you'll get an AP and a lateral, but let's just say you're looking at an AP
00:51:40.780 or a single mammogram. A radiologist will look at that. A radiologist will have clinical information
00:51:47.500 as well. So they will know why this patient presented in the case of the chest x-ray,
00:51:52.940 for example, in the ER in the middle of the night. Were they short of breath?
00:51:56.300 Do they have a fever? Do they have a previous x-ray? I can compare it to all sorts of information.
00:52:02.620 Are we not at the point now where all of that information could be given to the AI to enhance
00:52:09.420 the pre-test probability of whatever diagnosis it comes to?
00:52:12.860 I am delighted when you say pre-test probability. Don't talk dirty around me.
00:52:16.940 Love my Bayes theorem over here.
00:52:18.780 Yep. So you just said a lot, because what you just said actually went beyond what the straight
00:52:26.060 convolutional neural networks would do, because they actually could not replace radiologists,
00:52:30.380 because they could not do a good job of taking into account the previous history of the patient.
00:52:36.540 And it's required the emergence of transformers, where you can have multimodality. You have both
00:52:43.980 the image and the text. Now, they're going to do better than many, many radiologists today.
00:52:52.460 There is, I don't think, any threat yet to radiologists as a job. One of the most irritating
00:52:58.220 to doctors predictions was by Jeffrey Hinton, one of the intellectuals leaders of neural network
00:53:03.900 architecture. He said, I think it was in 2016, I have this approximately wrong, but in six years,
00:53:09.740 we wouldn't have no need for radiologists. And that was just clearly wrong. And the reason it was
00:53:16.220 wrong is A, they did not have these capabilities that we just talked about, about understanding
00:53:21.180 about the clinical context. But it's also the fact that we just don't have enough radiologists.
00:53:25.260 Meaning to do the training?
00:53:27.180 To actually do the work. So if you look at American medicine, I'll let you shut me down.
00:53:33.980 But if you look at the residency programs, we're not getting enough radiologists out. We have an
00:53:42.060 overabundance of applicants for interventional radiology. They're making a lot of money. It's
00:53:47.100 high prestige. But straight up radiology readers, not enough of them. Primary care doctors, I go around
00:53:54.620 medical schools and ask who's becoming a primary care doctor. Almost nobody. So primary care is
00:53:59.660 disappearing in the United States. In fact, Mass General and Brigham announced officially they're
00:54:05.020 not seeing primary care patients. People are still going to dermatology and they're still going to
00:54:10.060 plastic surgery. What I did, pediatric endocrinology, half of the slots nationally are not being filled.
00:54:17.500 Pediatric developmental disorders like autism, those slots, half of them filled.
00:54:23.740 PDID. There's a huge gap emerging in the available expertise. So it's not what we thought it was
00:54:33.180 going to be that we had a surplus of doctors that had to be replaced. It's just we have a surplus in
00:54:39.980 a few focused areas which are very popular. And then for all the work of primary care and primary
00:54:46.140 prevention, kind of stuff that you're interested in, we have almost no doctors available.
00:54:50.540 Yeah, let's go back to the radiologist for a second because, again, I'm fixated on this one
00:54:55.660 because it seems like the most, well, the closest one to address. And again, if you're saying, look,
00:55:01.500 we have a dearth of imaging radiologists who are able to work the emergency rooms, urgent care clinics,
00:55:08.140 and hospitals, wouldn't that be the first place we would want to apply our best of imaging recognition
00:55:15.820 with our super powerful GPUs and now plug them into our transformers with our language models
00:55:23.260 so that I can get clinical history, medical past history, previous images, current images,
00:55:30.380 and you don't have to send it to a radiologist in Australia to read it, who then has to send it back
00:55:35.900 to a radiologist here to check. Like, if we're just trying to fill a gap, that gap should be fillable,
00:55:40.780 shouldn't it? And that's exactly where it is being filled. And what keeps distracting me in this
00:55:46.380 conversation is that there's a whole other group of users of these AIs that we're not talking about,
00:55:52.620 which is the patients. And previously, none of these tools were available to patients. With the release
00:55:58.700 of GPT 3.5 and 4, and now Gemini and Claude III, they're being used by patients all the time in ways
00:56:06.540 that we had not anticipated. Let me give you an example. So there's a child who was having trouble
00:56:14.780 walking, having trouble chewing, and then started having intractable headaches. Mom brought him to
00:56:21.980 multiple doctors, they did multiple imaging studies, no diagnosis, kept on being in intractable
00:56:28.540 pain. She just typed into GPT-4 all the reports and asked GPT-4, what's the diagnosis? And GPT-4
00:56:36.220 said, tethered cord syndrome. She then went with all the imaging studies to a neurosurgeon and said,
00:56:42.220 what is this? He looked at it and said, tethered cord syndrome. And we have such an epidemic of
00:56:48.460 misdiagnosis and undiagnosed patients. Part of my background that I'll just mention briefly,
00:56:55.580 I'm the principal investigator of the coordinating center of something called the Undiagnosed Network.
00:56:59.740 It's a network with 12 academic hospitals down the West Coast from University of Washington,
00:57:05.580 Stanford, UCLA, to Baylor, up the East Coast, Harvard hospitals, NIH. And we see a few thousand
00:57:12.300 patients every year. And these are patients who have been undiagnosed and they're in pain. That's just a
00:57:17.980 small fraction of those who are undiagnosed. And yes, we bring to bear a whole bunch of computational
00:57:23.020 techniques and genomic sequencing to actually be able to help these individuals. But it's very clear
00:57:29.100 that there's a much larger burden out there of misdiagnosed individuals.
00:57:33.340 But the question for you, Zach, which is, does it surprise you that in that example, the mother
00:57:37.980 was the one that went to GPT-4 and inputted that? I mean, she had presumably been to many physicians
00:57:45.100 along the way. Were you surprised that one of the physicians along the way hadn't been the one to say,
00:57:51.260 gee, I don't know, but let's see what this GPT-4 thing can do?
00:57:54.460 Most clinicians I know do not have what I used to call the Google reflex. I remember when I was
00:58:02.620 on the wards and we had a child with dysmorphology, they look different. And I said to the fellows,
00:58:10.700 this is after residency, what is the diagnosis? And they said, I don't know, I don't know. I said,
00:58:16.140 he has this and this and this finding. What's the diagnosis? And I said, how would you find out?
00:58:20.780 They had no idea. I just said, let's take what I just said and type it into Google.
00:58:24.620 In the top three responses, there was the diagnosis. And that reflex, which they do
00:58:31.100 use in a civilian life, they did not have in the clinic. And doctors are in a very unhappy position
00:58:38.220 these days. They're really being driven very, very hard. And they're being told to use certain
00:58:44.140 technological tools. They're being turned into data entry clerks. They don't have the Google reflex.
00:58:49.740 They don't have the reflex, who has the time to look up a journal article? They don't do the Google
00:58:55.500 reflex. Even less, do they have the, let's look at the patient's history and see what GPT-4 would
00:59:02.380 come up with. I was gratified to see early on doctors saying, wow, look, I just took the patient
00:59:09.260 history, plugged into GPT-4 and said, write me a letter of prior authorization. And they were
00:59:14.380 actually tweeting about doing this, which on the one hand, I was very, very pleased for them
00:59:19.180 because it was saving them five minutes to write that letter to the insurance company saying,
00:59:24.300 please authorize my patient for this procedure. I was not pleased for them because if you use chat GPT,
00:59:30.460 you're using a program that is covered by open AI, as opposed to a version of GPT-4 that is being run
00:59:38.860 on protected Azure cloud by Microsoft, which is HIPAA covered. For those of you, the audience
00:59:45.260 doesn't know, HIPAA is the legal framework under which we protect patient privacy. And if you violate
00:59:50.940 it, you can be fined and even go to prison. So in other words, if a physician wants to put any
00:59:57.260 information into GPT-4, they better not identify it. That's right. So they just plunked in a patient
01:00:04.300 note into chat GPT. That's a HIPAA violation. If there's a Microsoft version of it, which is HIPAA
01:00:10.620 compliant, it's not. So they were using it to improve their lives. The doctors were using it
01:00:15.580 for improving the business, the administrative part of healthcare, which is incredibly important.
01:00:20.140 But by and large, only a few doctors use it for diagnostic acumen.
01:00:26.620 And then what about more involved radiology? So obviously a plain film is one of the more
01:00:32.460 straightforward things to do, although it's far from straightforward, as anybody knows who's
01:00:36.460 stared at a chest x-ray. But once we start to look at three-dimensional images, such as
01:00:41.420 cross-sectional images, CT scans, MRIs, or even more complicated images like ultrasound and things of
01:00:48.220 that nature, what is the current state of the art with respect to AI in the assistance of reading
01:00:55.900 these types of images?
01:00:57.580 So that's the very exciting news, which is, remember how I said it was important to have
01:01:03.340 a lot of data, one of the three ingredients in the breakthrough. So all of a sudden having
01:01:07.740 a lot of data around, for example, echocardiograms, the ultrasounds of your heart. Normally it takes
01:01:14.300 a lot of training to interpret those images correctly. So there is a recent study from the
01:01:20.700 Echo Clip Group led, I think, out of UCLA. And they took a million echocardiograms and a million
01:01:30.380 textual reports and essentially trained the model, both to create those embeddings I talked about
01:01:36.780 of the images and of the text.
01:01:39.020 Just to make sure people understand what we're talking about, this is not, here's a picture of
01:01:44.700 a cat, here's a description, cat. When you put the image in, you're putting a video in. Now you're
01:01:51.420 putting a multi-dimensional video because you have time scale, you have Doppler effects. This is a very
01:01:59.180 complicated video that is going in.
01:02:01.180 It's a very complicated video and it's three-dimensional and it's weird views from different angles.
01:02:08.300 And it's dependent on the user. In other words, the tech, the radiology tech can be good or bad.
01:02:16.860 If I was the one doing it, it would be awful.
01:02:20.780 The echo tech does not have medical school debt. They don't have to go to medical school. They don't
01:02:25.660 have to learn calculus. They don't have to learn physical chemistry, all the hoops that you have to
01:02:29.740 go through in medical school. You don't have the attitudinal debt of doctors. So in two years,
01:02:34.460 they get all those skills and they actually do a pretty good job.
01:02:37.020 They do a fantastic job. But my point is, their skill is very much an important determinant of
01:02:42.700 the quality of the image.
01:02:44.300 Yes. But what we still require these days is a cardiologist to then read it and interpret it.
01:02:50.460 Right. That's sort of where I'm going, by the way, is we're going to get rid of the
01:02:52.940 cardiologist before we get rid of the technician.
01:02:55.500 We're on the same page. My target in this conversation is nurse practitioners
01:03:00.220 and physician assistants with these tools can replace a lot of expert clinicians.
01:03:06.140 And there is a big open question. What is the real job for doctors in 10 years from now?
01:03:13.100 And I don't think we know the answer to that because you fast forward to the conversation just now.
01:03:18.780 Excellent. Well, let's think about it. We still haven't come to proceduralists.
01:03:22.780 So we still have to talk about the interventional radiologist, the interventional cardiologist,
01:03:26.460 and the surgeon. We can talk about the role of the surgeon and the da Vinci robot in a moment.
01:03:31.500 But I think what we're doing is we're kind of identifying the pecking order of physicians.
01:03:36.700 And let's not even think about it through the lens of replacement. Let's start with the lens of
01:03:40.780 augmentation, which is the radiologist can be the most easily augmented, the pathologist,
01:03:47.900 the dermatologist, the cardiologist who's looking at echoes and EKGs and stress tests. People who are
01:03:56.460 interpreting visual data and using visual data will be the most easily augmented. The second tranche of
01:04:03.500 that will be people who are interpreting language data plus visual data. So now we're talking about
01:04:09.020 your internist, your pediatrician, where you have to interpret symptoms and combine them with laboratory
01:04:15.260 values and combine it with a story and an image. Is that a fair assessment in terms of tier?
01:04:21.260 Absolutely a fair assessment. My only quibble, it's not a quibble, I'm going to keep on going back to
01:04:25.820 this, is in a place where we don't have primary care. The American Association of Medical Colleges
01:04:31.580 estimates that by 35, that's only 11 years from now, we'll be missing on the order of 50,000 primary
01:04:36.940 care doctors. As I told you, I can't get primary care at the Brigham or at MGH today. And in the absence of
01:04:43.420 that, you have to ask yourself, how can we replace these absent primary care practitioners with
01:04:50.220 nurse practitioners with physician assistants augmented by these AIs? Because there's literally
01:04:57.980 no doctor to replace. So tell me, Zach, where are we technologically on that augmentation? If NVIDIA
01:05:06.060 never came out with another chip, if they literally said, you know what, we are only interested in building
01:05:12.860 golf simulators, and we're done with the progress of this, and this is as good as it's going to get. Do we have
01:05:19.980 good enough GPUs, good enough multi-layer neural networks, that all you need is more data and training sets, that we
01:05:28.780 could now do the augmentation that has been described by us in the last five minutes?
01:05:33.740 The short answer is yes. Let me make it very concrete. Most concierge services cost in Boston
01:05:39.900 somewhere between $5,000 and $20,000 a year. You can get this very low cost concierge service that I'm
01:05:46.060 just amazed that have not done the following, called One Medical. One Medical was acquired by Amazon.
01:05:51.500 And they have a lot of nurse practitioners in there. And you can make an appointment,
01:05:55.100 and you can text with them. I believe that those individuals could be helped in ordering the right
01:06:02.380 imaging studies, the right EKGs, the right medications, and assess your continuing heart failure,
01:06:11.980 and only decide in the very few cases that you need to see a specialist cardiologist or a specialist
01:06:19.660 endocrinologist today. Just be a matter of just making the current models better, evaluating them,
01:06:26.540 because not all models are equal. A big question for us, this is the regulatory question, which is,
01:06:32.140 which ones do a better job? And they're not all equal. I don't think we need technological breakthroughs
01:06:38.860 to just make the current set of paraprofessionals work at the level of entry-level doctors. Let me quickly
01:06:47.180 say the old very bad joke. What do you call the medical student who graduates at the bottom of
01:06:52.300 his class? Doctor. And so if you could just merely get the bottom 50% of doctors to be as good as the
01:07:01.580 top 50%, that would be transformative for healthcare. Now, there are other superhuman capabilities that we
01:07:10.220 can go towards, and we can talk about if we want, that do require the next generation of algorithms,
01:07:18.060 NVIDIA architectures, and data sets. Everything stopped now, we could already transform medicine.
01:07:24.300 It's just a matter of the sweat equity to create the models, figure out how to include them in the
01:07:30.780 workflow, how to pay for them, how to create a reimbursement system, and a business model that
01:07:37.420 works for our society. But there's no technological barrier.
01:07:42.940 In my mind, everything we've talked about is take the best case example of medicine today
01:07:49.900 and augment it with AI such that you can raise everyone's level of care to that of the best,
01:07:56.300 best, no gaps, and it's scaled out. Okay, now let's talk about another problem, which is where do you
01:08:04.780 see the potential for AI in solving problems that we can't even solve on the best day at the best
01:08:13.740 hospitals with the best doctors? So let me give you an example. We can't really diagnose Alzheimer's
01:08:21.180 disease until it appears to be at a point that for all intents and purposes is irreversible.
01:08:29.580 Maybe on a good day, we can halt progression really, really early in a patient with just a whiff of MCI,
01:08:36.300 mild cognitive impairment, maybe with an early amyloid detection and an anti-amyloid drug.
01:08:43.100 But is it science fiction to imagine that there will be a day when an AI could listen to a person's
01:08:48.620 voice, watch the movements of their eyes, study the movements of their gait, and predict 20 years
01:08:57.020 in advance when a person is staring down the barrel of a neurodegenerative disease and act at a time
01:09:03.260 when maybe we could actually reverse it? How science fiction-y is that?
01:09:07.580 I don't believe it's science fiction at all. Do you know that looking at retinas today, images of retina,
01:09:13.660 straightforward convolutional neural network, not even ones that involve transformers,
01:09:17.820 can already tell you by looking at your retina, not just whether you have retinal disease,
01:09:23.020 but if you have hypertension, if you're a male, if you're female, how old you are,
01:09:28.700 and some estimate of your longevity. And that's just looking at the back of your eye
01:09:33.580 and seeing enough data. I was a small player in a study that appeared in Nature in 2005 with Bruce
01:09:40.540 Yankner. We were looking at the frontal lobes of individuals who had died for a variety of reasons,
01:09:47.500 often in accidents of various ages. And we saw, bad news for people like me, that after age 40,
01:09:54.380 your transcriptome, the genes that are switched on, fell off a cliff. Thirty percent of your transcriptome
01:10:00.860 went down. And so there seemed to be a big difference in the expression of genes around age 40,
01:10:07.900 but there was one 90-year-old who looked like the young guy. So maybe there's hope for some of us.
01:10:12.140 But then I thought about it afterwards, and there were other things that actually have much smoother
01:10:16.380 functions, which don't have quite a fall off, like our skin. So our skin ages. In fact, all our organs
01:10:23.980 age and they age at different rates. You're saying that the transcriptome of the skin,
01:10:28.940 you did not see this cliff-like effect at a given age, the way you saw it in the frontal cortex.
01:10:34.540 So different organs age at different rates, but having the right data sets and the ability to see
01:10:42.540 nuances that we don't notice makes it very clear to me that the early detection part, no problem.
01:10:49.580 It can be very straightforward. The treatment part, we can talk about it as well. But again,
01:10:54.940 we had early on from the very famous Framium Heart Study, a predictor of when you had going to have
01:11:01.500 heart disease based on just a handful of variables. Now we have these artificial intelligence models
01:11:07.500 that, based on hundreds of variables, can predict various other diseases. And it will do Alzheimer's,
01:11:16.220 I believe, very soon. I think you'll be able to see a combination of
01:11:21.500 gait, speech patterns, picture of your body, picture of your skin, and eye movements. Like you said,
01:11:29.740 will be a very accurate predictor. We just published, by the way recently, speaking about eyes,
01:11:34.300 a very nice study where in a car, just by looking at the driver, it can figure out what your blood sugar is.
01:11:42.460 Because diabetics previously have not been able to get driver licenses sometimes because of the worry about
01:11:49.260 them passing out because of hypogycemia. So there was a very nice study that showed that you could
01:11:52.940 just, by looking, have cameras pointed at the eyes, could actually figure out exactly what the blood
01:11:57.500 sugar is. So that kind of detection is, I think, fairly straightforward. It's a different question
01:12:03.740 about what you can do about it. Before we go to the what you can do about it,
01:12:07.100 I just want to go a little deeper on the predictive side. You brought up the Framingham model or the
01:12:12.220 multi-ethnic study on atherosclerosis, the MESA model. These are the two most popular models by far
01:12:17.100 looking at a major adverse cardiac event risk prediction. But you needed something else to
01:12:21.820 build those models, which was enough time to see the outcome. In the Framingham cohort,
01:12:27.100 which was the late 70s and early 80s, you then had the Framingham offspring cohort.
01:12:31.740 And then you had to be able to follow these people with their LDL-C and HDL-C and triglycerides.
01:12:36.700 And later, eventually, they incorporated calcium scores. So if today we said, look,
01:12:43.100 we want to be able to predict 30-year mortality, which is something no model can do today,
01:12:50.700 this is a big pet peeve of mine, is we generally talk about cardiovascular disease through the lens
01:12:55.580 of 10-year risk, which I think is ridiculous. We should talk about lifetime risk. But I would
01:13:00.220 settle for 30-year risk, frankly. And if we had a 30-year risk model where we could take
01:13:07.100 many more inputs, I would absolutely love to be looking at the retina. I believe, by the way,
01:13:12.540 Zach, that retinal examination should be a part of medicine today for everybody.
01:13:17.180 I would take a retinal exam over a hemoglobin A1C all day, every day. I'd never look at another A1C
01:13:24.620 again if I could see the retina of every one of my patients. But my point is, even if effective today,
01:13:30.540 we could define the data set, and let's overdo it, and we can prune things later. But we want to see
01:13:35.820 these 50 things in everybody to predict every disease. Is there any way to get around the fact
01:13:41.340 that we're going to need 30 years to see this come to fruition in terms of watching how the story plays
01:13:46.300 out? Or are we basically going to say, no, we're going to do this over five years? It won't be that
01:13:51.340 useful because a five-year predictor basically means you're already catching people in the throes of
01:13:55.100 the disease. I'll say three words, electronic health records. So that turns out not to be the
01:14:02.060 answer in the United States. Why? Because in the United States, we move around. We don't stay in any
01:14:09.020 given healthcare system that long. So very rarely will I have all the measurements made on you,
01:14:15.820 Peter, all your glycohemoglobins, all your blood pressures, all your clinic visits, all the imaging
01:14:21.180 studies that you've had. However, that's not the case in Israel, for example. In Israel, they have
01:14:28.460 these HMOs, health maintenance organizations. And one of them, Clarit, I have a good relationship with
01:14:35.020 because they published all the big COVID studies looking at the efficacy of the vaccine. And why
01:14:42.540 could they do that? Because they had the whole population available. And they have about 20,
01:14:48.220 25 years worth of data on all their patients in detail and family relationships. So if you have
01:14:55.980 that kind of data, and Kaiser Permanente also has that kind of data, I think you can actually come
01:15:02.060 close. But you're not going to be able to get retina, gait, voice, because we still have to get those
01:15:07.900 prospectively. I'm going to claim that there are proxies, rough proxies, but for gait, false.
01:15:15.900 And for hearing problems, visits to the audiologist. Now, these are noisier measurements. And so,
01:15:23.820 those of us who are data junkies, like I am, always keep mumbling to ourselves, perfect is the
01:15:30.060 enemy of good. Waiting 30 years to have the perfect data set is not the right answer to help patients
01:15:36.300 now. And there are things that we could know now that are knowable today that we just don't know
01:15:42.860 because we haven't bothered to look. I'll give you a quick example. I did a study of autism using
01:15:49.180 electronic health records, maybe 15 years ago. And I saw there was a lot of GI problems. And I
01:15:55.660 talked to a pediatric expert, and they said, it was a little bit dismissive. They said, brain bad,
01:16:00.940 tummy hurt. I've seen a lot of inflammatory bowel disease. It just doesn't make sense to me that this
01:16:06.220 is somehow effect of brain function. To make a long story short, we did a massive study. We're
01:16:11.660 looking forward to tens of thousands of individuals. And sure enough, we found subgroups of patients who
01:16:15.820 had immunological problems associated with their autism, and they had type 1 diabetes,
01:16:21.500 inflammatory bowel disease, lots of infections. Those were knowable, but they were not known. And I had,
01:16:27.020 frankly, parents coming to me more thankful than for anything else I had ever done for them
01:16:30.860 clinically, because I was telling these parents, they weren't hallucinating that these kids have
01:16:35.420 these problems. They just weren't being recognized by medicine because no one had the big wide angle
01:16:41.340 to see these trends. So, without knowing the field of Alzheimer's the way I do other fields,
01:16:48.140 I bet you there are trends in Alzheimer's that you can pick up today by looking at enough patients
01:16:53.660 that you'll find some that have more frontotemporal components, some that have more effective
01:16:58.780 components, some that have more of an infectious and immunological component. Those are knowable
01:17:04.220 today. Zach, you've already alluded to the fact that we're dealing with a customer, if the physician
01:17:11.580 is the customer, who is not necessarily the most tech-forward customer. And truthfully, like many customers
01:17:20.060 of AI, runs the risk of being marginalized by the technology if the technology gets good enough.
01:17:25.900 And yet, you need the customer to access the patient to make the data system better,
01:17:33.900 to make the training set better. So, how do you see the interplay over the next decade of that dynamic?
01:17:42.620 That's the right question. Because in order for these AI models to work, you need a lot of data,
01:17:47.900 a lot of patients. Where is that data going to come from? So, there are some healthcare systems,
01:17:52.860 like the Mayo Institute, who think they can get enough data in that fashion. There are some data
01:18:00.940 companies that are trying to get relationships with healthcare systems where they can get de-identified
01:18:06.540 data. I'm betting on something else. There is a trend where consumers are going to have increased
01:18:13.580 access to their own data. The 21st Century Cures Act was passed by Congress, and it said that patients
01:18:20.220 should be given access to their own data programmatically. Now, they're not expecting
01:18:24.860 your grandmother to write a program to access the data programmatically, but by having a right to it,
01:18:30.540 it enables others to do so. So, for example, Apple has something called Apple Health. It has this big
01:18:36.060 heart icon on it. If you're one of the 800 hospitals that they've already hooked up with,
01:18:40.780 Pass General or Brigham Women's, and you're a patient there, if you authenticate yourself to it,
01:18:45.260 if you give it your username and password, it will download into your iPhone, your labs, your meds,
01:18:51.900 your diagnoses, your procedures, as well as all the wearable stuff, your blood pressure that you get
01:18:58.220 as an outpatient, and various other forms of data. That's already happening now. There's not a lot of
01:19:04.140 companies that are taking advantage of that, but right now that data is available on tens of millions
01:19:08.780 of Americans. Isn't it interesting, Zach, how unfriendly that data is in its current form? I'll
01:19:15.420 give you just a silly example in our practice. So, if we send a patient to LabCorp or Boston Heart or
01:19:21.340 Pick Your Favorite Lab, and we want to generate our own internal reports based on those where we want
01:19:28.940 to do some analysis on that, lay out trend sheets, we have to use our own internal software. It's almost
01:19:37.420 impossible to scrape those data out of the labs because they're sending you PDF reports. Their
01:19:44.860 APIs are garbage. Nothing about this is user-friendly. So, even if you have the My Heart thing or whatever,
01:19:53.020 the My Health thing come on your phone, it's not navigable. It's not searchable. It doesn't show you
01:19:58.540 trends over time. Like, is there a more user-hostile industry from a data perspective than the health
01:20:04.140 industry right now? No, no. And there's a good reason why, because they're keeping you captive.
01:20:10.860 But Peter, the good news is you're speaking to a real nerd. Let me tell you two ways where we could
01:20:17.020 solve your problem. One, if it's in that Apple Health thing, someone can actually write a program,
01:20:22.140 an app on the iPhone, which will take those data as numbers and not have to scrape it. And it could run
01:20:28.620 it through your own trending programs. You could actually use it directly. Also, Gemini and GPT-4,
01:20:34.540 you can actually give it those PDFs. And actually, with the right prompting, it will actually take
01:20:41.100 those data and turn them into tabular spreadsheets. We can't do that because of HIPAA, correct?
01:20:47.340 If the patient gets it from the patient portal, absolutely, you can do that.
01:20:50.700 The patient can do that, but I can't use a patient's data that way.
01:20:54.380 If the patient gives it to you, absolutely. Really? Oh, yes. But it's not de-identified.
01:21:00.940 It doesn't matter. If a patient says, Peter, you can take my 50 LabCorp reports for the last 10 years,
01:21:09.260 and you can run them through ChatGPT to scrape it out and give me an Excel spreadsheet that will
01:21:15.900 perfectly tabularize everything that we can then run into our model to build trends and look for
01:21:21.100 things. I didn't think that was doable, actually. So it's not doable through ChatGPT because your
01:21:26.220 lawyers would say, Peter, you're going to get a million dollars in fines from HIPAA. I'm not a
01:21:31.100 shill for Microsoft. I don't own any stock. But if you do GPT on the Azure cloud that's HIPAA protected,
01:21:37.740 you absolutely can use it with patient consent. 100% you can do it. GPT is being used
01:21:43.660 with patient data out of Stanford right now. EPIC's using GPT-4, and it's absolutely legitimately
01:21:51.980 usable by you. People don't understand that. We've now just totally bypassed OCRs.
01:21:57.420 We do not need to waste our time with, for people not in the acronyms, optical character recognition,
01:22:02.460 which is 15 years ago what we were trying to do to scrape this data. Peter, let me tell you,
01:22:08.940 there's New England Journal of Medicine. I'm on the editorial board there. And we just published
01:22:12.700 three months ago, a picture of the week back of this 72-year-old. And it looks like a bunch of red
01:22:18.620 marks. To me, it looks like something to scratch themselves. And it says, blah, blah, blah. They
01:22:23.100 had trouble sleeping. This is the image of the week. Image of the week. And I took that whole thing,
01:22:27.820 and I took out one important fact, and then gave it to GPT-4, the image and the text. And it came up with
01:22:35.500 the two things I thought it would be, either bleomycin toxicity, which I don't know what
01:22:39.660 that looks like, and shiitake mushroom toxicity. What I'd removed is the fact that the guy had eaten
01:22:47.100 mushrooms the day before. So this thing just like looking at the picture.
01:22:52.860 GPT-4 spit this out? Yes.
01:22:55.900 I don't think most doctors know this, Zach. I don't think most doctors understand. First of all,
01:23:01.820 I can't tell you how many times I get a rash. Well, I try to send a picture to my doctor,
01:23:06.700 or my kid gets a rash, and I'm trying to send a picture to their pediatrician,
01:23:10.700 and they don't know what it is. And it's like, we're rubbing two sticks together,
01:23:14.380 and you're telling me about the Zippo lighter. Yes. And that's what I'm saying is patients without
01:23:19.260 primary care doctors. I know I keep repeating myself. They understand that they have a Zippo
01:23:22.860 lighter waiting three months because of a rash or the symptoms. They say, I'll use this Zippo lighter.
01:23:28.540 It's better than no doctor for sure, and maybe better. That's now. Quickly illustrate it. I don't
01:23:34.460 know squat about the FDA. And so I pulled down from the FDA the adverse event reporting files.
01:23:41.180 It's a big zip file, compressed file. And I said to GPT-4, please analyze this data. And it says,
01:23:47.260 unzipping. Based on this table, I think this is about the adverse events, and this is the
01:23:51.900 locations. What do you want to know? I say, tell me what adverse events for disease-modifying
01:23:57.740 drugs for arthritis. It says, oh, to do that, I'll have to join these two tables.
01:24:02.220 And it just does it. It creates its own Python code. It does it, and it gives me a report.
01:24:08.140 Is this a part of medical education now? You're at Harvard, right? You're at one of the three best
01:24:12.620 medical schools in the United States, arguably in the world. Is this an integral part of the
01:24:17.660 education of medical students today? Do they spend as much time on this as they do histology,
01:24:23.500 where I spent a thousand hours looking at slides under a microscope that I've never once tried to
01:24:29.740 understand? Again, I don't want to say there wasn't a value in doing that. There was,
01:24:34.620 and I'm grateful for having done it. But I want to understand the relative balance of education.
01:24:40.060 It's like the stethoscope. Arguably, we should be using things other than the stethoscope.
01:24:44.540 Let me make sure I don't get fired, or at least beaten severely, by telling you that George Daly,
01:24:50.220 our dean of the medical school, has said explicitly he wants to change all of medical education.
01:24:54.940 So these learnings are infused throughout the four years, but it's going to take some doing.
01:25:01.180 Let's now move on to the next piece of medicine. So we've gone from purely the recognition image-based
01:25:09.980 to how do I combine image with voice, story, text. You've made a very compelling case that we don't
01:25:17.820 need any more technological breakthroughs to augment those. It's purely a data set problem at this point,
01:25:22.780 and a willingness. Let's now move to the procedural. Is there, in our lifetimes,
01:25:28.540 say, Zach, the probability that if you need to have a radical prostatectomy, which currently,
01:25:35.980 by the way, is never done open. This is a procedure that the da Vinci, a robot, has revolutionized.
01:25:41.580 There's no blood loss anymore. When I was a resident, this was one of the bloodiest operations
01:25:46.380 we did. It was the only operation, by the way, for which we had the patients donate their own blood
01:25:52.300 two months ahead of time. That's how guaranteed it was that they were going to need blood transfusions,
01:25:57.580 so we just said, to hell with it. Come in a couple of months before, give your own blood,
01:26:01.180 because you're going to need at least two units following this procedure.
01:26:05.260 Today, it's insane how successful this operation is on a large part of the robot.
01:26:11.180 But the surgeon needs to move the robot. Are we getting to the point where that could change?
01:26:17.340 So let me tell you where we are today. Today, there's been studies where it's collected a bunch
01:26:21.660 of YouTube videos of surgery and traded up one of these general models. So it says, oh, they're
01:26:29.020 putting on the scalpel to cut this ligament. And by the way, that's too close to the blood vessel.
01:26:35.580 They should move it a little bit to the side. That's already happening. Based on what we're seeing with
01:26:40.780 robotics in the general world, I think the da Vinci controlled by a robot 10 years is a very safe bet.
01:26:50.940 It's a very safe bet. In some ways, 10 years is nothing.
01:26:54.620 It's nothing. But it's a very safe bet. The fact is, right now, I can do a better job,
01:26:59.500 by the way, just to go back to our previous discussion, giving you a genetic diagnosis
01:27:04.780 based on your findings than any primary care provider interpreting a genomic test.
01:27:10.860 So are you using that example, Zach, because it's a huge data problem? In other words, that's obvious
01:27:17.900 that you would be able to do that because the amount of data, I mean, there's 3 billion base pairs
01:27:22.780 to be analyzed. So of course, you're going to do a better job.
01:27:25.180 But linking it to symptoms.
01:27:27.100 Yeah, yeah. But you're saying surgery is a data problem because if you turn it into a pixel problem?
01:27:33.420 Pixel and movement.
01:27:34.780 And degrees of freedom. Yeah.
01:27:36.620 That's it. Remember, there's a lot of degrees of freedom in moving a car around traffic. And by the
01:27:42.300 way, lives are on the line there too. Now, medicine is not the only job where lives are at stake.
01:27:49.740 Driving a ton of metal at 60 miles per hour in traffic is also putting lives at stake. And last
01:27:57.100 time I looked, there's several manufacturers who are saying that, or some appreciable fraction
01:28:05.020 of that effort, they're controlling multiple degrees of freedom with a robot.
01:28:08.620 Yeah. I very recently spoke with somebody, I won't name the company, I suppose, but it's one
01:28:15.260 of the companies that's deep in the space of autonomous vehicles. And they very boldly stated,
01:28:21.820 they made a pretty compelling case for it, that if every vehicle on the road was at their level of
01:28:27.100 technology and autonomous driving, you wouldn't have fatalities anymore. But the key was that every
01:28:32.780 vehicle had to be at that level. I don't know if you know enough about that field, but does that
01:28:36.620 sense check to you? Well, first of all, I'm a terrible driver. I am a better driver. It's not
01:28:42.140 for an ad, but the fact is I'm a better driver because I'm not on a Tesla, because I'm a terrible
01:28:46.300 driver. And there's actually a very good message for medicine, because I will paraphrase this.
01:28:51.900 I knew enough to know that I need to jiggle the steering wheel when I'm driving with a Tesla,
01:28:56.060 because otherwise it will assume that I'm just zoning out. But what I didn't realize is this,
01:29:01.260 I'm very bad. I'll pick up my phone and I'll look at it. I didn't realize it was looking at me and
01:29:06.060 it says, because Zach put down the phone. So I, okay, I put down. Three minutes later,
01:29:10.460 I pick it up again and it says, okay, that's it. I'm switching off autopilot. So it switches off
01:29:16.140 autopilot and now I have to pay attention, full attention. Then I get home and it says, all right,
01:29:21.740 that was bad. You do that four more times. I'm switching off autopilot until the next software
01:29:27.820 update. And the reason I mentioned that is it takes a certain amount of confidence to do that
01:29:33.180 to your customer base saying, I'm switching off the thing that they bought me for. In medicine,
01:29:38.780 how likely is it that we're going to fall asleep at the wheel if we have an AI thinking for us?
01:29:43.980 It's a real issue. We know for a fact, for example, back in the nineties,
01:29:47.740 that doses for a drug like a dancetron, where people would talk endlessly about how frequently
01:29:52.860 you should be given it with what dose. The moment you put it in the order entry system,
01:29:56.540 95% of doctors would just use the default there. And so how in medicine are we going to keep doctors
01:30:03.100 awake at the wheel? And will we dare to do that kind of challenges that I just described the car
01:30:09.340 doing? So just to get back to it, I do believe because of what I've seen with autonomy and robots
01:30:16.940 that as fancy as we think that is, controlling a dementia robot will probably have less
01:30:23.980 bad outcomes. Every once in a while, someone nicks something and you have to go into full
01:30:29.340 surgery or they go home and they die on the way home because they exsanguinate. I think it's just
01:30:34.620 going to be safer. It's just unbelievable for me to wrap my head around that. But truthfully,
01:30:41.500 it's impossible for me to wrap my head around what's already happened. So I guess I'll try to
01:30:45.740 retain the humility that says I reserve the right to be startled. Again, there are certain things that
01:30:51.980 seem much easier than others. Like I have an easier time believing we're going to be able to replace
01:30:56.140 interventional cardiologists where the number of degrees of freedom, the complexity and the
01:31:02.060 relationship between what the image shows, what the cath shows and what the input is, the stent,
01:31:09.180 that gap is much narrower. Yeah, I can see a bridge to that. But when you talk about doing a Whipple
01:31:14.460 procedure, when you talk about what it means to cell by cell take a tumor off the superior mesenteric
01:31:22.380 vessels, I'm thinking, oh my God. Since we're on record, I'm going to say, I'm talking about your
01:31:28.380 routine prostate removal. Yeah. First 10 years, I would take that bet today. Wow. Let's go one layer further.
01:31:36.940 Sure. Let's talk about mental health. This is a field of medicine today that I would also argue
01:31:42.940 is grossly underserved. Everything you've said to date resonates. I completely agree from my own
01:31:49.980 experience that the resources in pediatrics and primary care, I mean, these things are unfortunate
01:31:55.980 at the moment. Harvard has, I think, 60% of undergraduates are getting some sort of mental
01:32:01.180 health support and it's completely outdoing all the resources available to the university health
01:32:06.780 services. And so we have to outsource some of our mental health. And this is a very richly endowed
01:32:11.660 university. In general, we don't have the resources. So here we live in a world where I think the evidence
01:32:18.300 is very clear that when a person is depressed, when a person is anxious, when a person has any sort of
01:32:24.580 mental or emotional illness, pharmacotherapy plays a role, but it can't display psychotherapy.
01:32:30.100 You have to be able to put these two things together. And the data would suggest that the
01:32:35.660 knowledge of your psychotherapist is important, but it's less important than the rapport you can
01:32:41.700 generate with that individual. Now, based on that, do you believe that the most sacred, protected,
01:32:49.620 if you want to use that term, profession within all of medicine will then be psychiatry?
01:32:54.980 I'd like to think that if I had a psychiatric GPT speaking to me,
01:33:00.100 I wouldn't think that it understood me. On the other hand, back in the 1960s or 70s,
01:33:08.900 there was a program called Eliza and it was a simple pattern matching program. It would just emulate
01:33:14.740 what's called a Rogerian therapist, where I really hate my mother. Why do you say you hate your mother?
01:33:21.140 Oh, it's because I don't like the way she fed me. What is it about the way she fed you? Just very,
01:33:26.980 very simple pattern matching. And this Eliza program, which was developed by Joe Weizenbaum at MIT,
01:33:34.660 his own secretary would lock herself in her office to have sessions with this thing because it's
01:33:39.700 non-judgmental. This was in the 80s?
01:33:42.260 70s or 60s.
01:33:43.700 Wow.
01:33:44.420 Yeah. And it turns out that there's a large group of patients who actually would rather have a non-human,
01:33:51.380 non-judgmental person who remembers what they've said from last time, shows empathy verbally. Again,
01:33:58.500 I wrote this book with Peter Lee and Peter Lee made a big deal in the book about how GPT-4 is showing
01:34:05.220 empathy. In the book, I argued with him that this is not that big a deal. And I said, I remember from
01:34:11.140 medical school being told that some of the most popular doctors are popular because they're very
01:34:16.980 deep empaths, not necessarily the best doctors. And so I said, you know, for certain things,
01:34:22.100 that's just me. I could imagine a lot of, for example, cognitive behavioral therapy being done
01:34:29.380 and be found acceptable by a subset of human beings. It's not, wouldn't be for me. I'd say,
01:34:34.100 I'm just speaking to some stupid program. But if it's giving you insight into yourself and it is
01:34:39.220 based on the wisdom called for millions of patients, who's to say that it's worse? And it's certainly not
01:34:46.420 judgmental. And maybe it'll be less.
01:34:48.420 So Zach, you're born probably just after the first AI boom. You come of age, intellectually,
01:34:58.740 academically in the second. And now in the mature part of your career, when you're at the height of
01:35:05.780 your esteem, you're riding the wave of this third version, which I don't think anybody would argue
01:35:13.220 is going anywhere. As you look out over the next decade, and we'll start with medicine,
01:35:20.180 what are you most excited about? And what are you most afraid of with respect to AI?
01:35:25.780 Specifically with regard to medicine, what I'm most concerned about is how it could be used by
01:35:33.300 the medical establishment to keep things the way they are, to pour concrete over practices.
01:35:39.700 What I'm most excited about is alternative business models, young doctors who create businesses
01:35:48.260 outside the mold of hospitals. Hospitals are these very, very complex entities.
01:35:55.300 They make billions of dollars, some of the bigger ones, but with very small margins, one to 2%.
01:36:01.220 When you make, have huge revenue, but very small margins, you're going to be very risk averse.
01:36:06.180 And you're not going to want to change. And so what I'm excited about is the opportunity for new
01:36:13.380 businesses and new ways of delivering to patients insights that are data-driven. What I'm worried
01:36:19.940 about is hospitals doing a bunch of information blocking and regulations that will make it harder for
01:36:28.820 these new businesses to get created. Understandably, they don't want to be disrupted. That's the danger.
01:36:34.260 In that latter case or that case that you're afraid of, Zach, can patients themselves work around the
01:36:42.180 hospitals with these new companies, these disruptive companies and say, look, we have the legal framework
01:36:48.980 that says, I own my data as a patient. I own my data. Believe me, we know this in our practice.
01:36:54.340 Just because our patients own the data doesn't make it easy to get. There is no aspect of my practice
01:37:00.180 that is more miserable and more inefficient than data acquisition from hospitals. It's actually
01:37:06.740 comical.
01:37:07.140 Absolutely comical. And I do pay hundreds of dollars to get my data from my patients with rare and unknown
01:37:13.940 diseases in this network extracted from the hospitals because it's worth it to pay someone to do that
01:37:19.780 extraction. Yeah. But now I'm telling you it is doable.
01:37:23.860 So you're saying because of that, are you confident that the legal framework for patients to have their
01:37:29.540 data coupled with AI and companies, do you think that that will be a sufficient hedge against your
01:37:36.420 biggest fear?
01:37:37.060 I think that unlike my 10-year prostatectomy by robot prediction, I'm not as certain, but I would
01:37:44.020 give better than 50% odds that in the next 10 years, there'll be a company, at least one company,
01:37:50.100 that figures out how to use that patient's right to access through dirty APIs, using AI to clean it up,
01:37:59.300 provide decision support with human doctors or health professionals to create alternative
01:38:05.460 businesses. I am convinced because the demand is there. And I think that you'll see companies that
01:38:12.820 are even willing to put themselves at risk. What I mean by that, are willing to take the medical risk
01:38:18.100 on that if they do better than a certain level of performance, they get paid more. And if they do
01:38:24.340 worse- They don't get paid. Yeah.
01:38:26.660 I believe there were companies that are going to be in that space, but that is because I don't want to
01:38:32.420 underestimate the medical establishment's ability to squish threats. So we'll see.
01:38:38.100 Okay. Now let's just pivot to AI outside of medicine. Same question. What are you most afraid
01:38:44.340 of over the next decade? So maybe we're not talking about self-awareness and Skynet, but next decade,
01:38:51.700 what are you most afraid of? And what are you most excited about?
01:38:55.140 What I'm most afraid of is a lot of the ills of social networks being magnified by use of these
01:39:06.500 AIs to further accelerate cognitive chaos and vitriol that fills our social experiences on the net.
01:39:17.620 It could be used to accelerate them. So that's my biggest fear.
01:39:21.300 I saw an article two weeks ago that was an individual. I can't remember if they were
01:39:26.580 currently in or formerly part of the FBI. And they stated that they believed, I think it was
01:39:32.100 somewhere between 75 and 90% of quote unquote individuals on social media were not in fact
01:39:38.500 individuals. I don't know if you spend enough time on social media to have a point of view on that.
01:39:43.060 Unfortunately, I have to admit to the fact that my daughter, who's now 20 years old,
01:39:47.620 been four years ago, she bought me a mug that says on it, Twitter addict. I spent enough time.
01:39:53.380 I would not be surprised if some large fraction, our bots could get worse. And it's going to be
01:39:59.380 harder to actually distinguish reality from human beings, harder and harder and harder.
01:40:05.780 That's the real problem. We are fundamentally social animals. And if we cannot understand our social
01:40:12.980 context in most of our interactions, it's going to make us crazy, or I should say crazier. And my most
01:40:21.300 positive aspect is, I think that these tools can be used to expand the creative expression of all
01:40:28.740 people. If you're a poor driver like me, I'm going to be a better driver. If you're a lousy musician,
01:40:36.580 but have a great ear, you're going to be able to express yourself musically in ways that you could not do
01:40:41.300 before. I think you're going to see filmmakers who were never meant to be filmmakers before
01:40:46.820 express themselves. I think human expression is going to be expanded because just like printing
01:40:53.700 press allowed all sorts of... In fact, it's a good analogy because the printing press also created a
01:40:58.180 bunch of wars because it allowed people to make clear their opposition to the church and so on,
01:41:03.300 enabled a number of bad things to happen. But it allowed also expression of all literature in ways
01:41:08.420 that would have not been possible without the printing press. I'm looking forward to human
01:41:12.740 expression and creativity. I can't imagine you haven't played with some of the picture generation
01:41:17.460 or music generation capabilities of AI, or if you haven't, I strongly recommend. You're going to be
01:41:22.260 amazed. I have not. I am ashamed maybe to admit my interactions with AI are limited to really chat GPT
01:41:29.860 for and basically problem solving. Solve this problem for me. And by the way, I think I'm doing it at a very
01:41:35.860 JV level. I could really up my game there. Just before we started this podcast, I thought of a
01:41:40.900 problem. I've been asking my assistant to solve because A, I don't have the time to solve it and
01:41:45.860 I'm not even sure how I would solve it. It would take me a long time. I've been asking her to solve
01:41:49.700 it and it's actually pretty hard. And then I realized, oh my God, why am I not asking chat GPT for
01:41:54.340 to do it? So I just started typing in the question. It's a bit of an elaborate question. As soon as we're
01:41:59.620 done with this podcast, I'll probably go right back to it, but I haven't done anything creatively with it.
01:42:03.540 What I will say is what does this mean for human greatness? So right now, if you look at a book
01:42:12.560 that's been written and someone who's won a Pulitzer Prize, you sort of recognize like, I don't know if
01:42:17.800 you read Sid Mukherjee, right? He's one of my favorite writers when it comes to writing about science and
01:42:22.960 medicine. When I read something that Sid has written, I think to myself, there's a reason that he is so
01:42:30.180 special. He and he almost alone can do something we can't do. I've written a book. It doesn't matter.
01:42:38.500 I could write a hundred books. I'll never write like Sid and that's okay. I'm no worse a person.
01:42:43.860 I'm no worse a person than Sid, but he has a special gift that I can appreciate just as we could all
01:42:50.120 appreciate watching an exceptional athlete or an exceptional artist or musician. Does it mean anything
01:42:56.080 if that line becomes blurred? That's the right question. And yes, Sid writes like poetry.
01:43:04.440 Here's an answer which I don't like. I've heard many times people said, oh, you know that Deep Blue
01:43:10.620 beat Kasparov in chess, but chess is more popular than it ever was, even though we know that the best
01:43:17.580 chess players in the world are computers. So that's one answer. I don't like that answer at all.
01:43:22.720 Yeah. Because if we create Sid GPT and Sid wrote Alzheimer's, the second greatest malady,
01:43:31.180 and he wrote it in full Sid style, but it was not Sid, but it was just as empathic family references.
01:43:39.060 Right. The weaving of history with story with science. Yeah.
01:43:42.060 If it did that and it was just a computer, how would you feel about it, Peter?
01:43:45.120 I mean, Zach, you are asking the jugular question. I would enjoy it, I think, just as much,
01:43:51.840 but I don't know who I would praise. Maybe I have in me a weakness slash tendency to want to idolize.
01:44:00.180 You know, I'm not a religious person, so my idols aren't religious, but I do tend to love to see
01:44:06.340 greatness. I love to look at someone who wrote something who's amazing and say, that amazes me.
01:44:11.720 I love to be able to look at the best driver in the history of Formula One and study everything
01:44:17.900 about what they did to make them so great. So I'm not sure what it means in terms of that.
01:44:22.940 I don't know how it would change that.
01:44:24.560 I grew up in Switzerland, in Geneva. And even though I have this American accent,
01:44:28.860 both my parents were from Poland. And so the reason I have an American accent is I went to
01:44:33.180 international school with a lot of Americans. All I read was whatever my dad would get me from
01:44:37.900 England in science fiction. So I'm a big science fiction fan. So let me go science fiction on you
01:44:43.080 to answer this question. It's not going to be in 10 years, but it could be in 50 years.
01:44:48.240 You'll have idols and idols will be, yes, Greg Orovich wrote a great novel, but you know,
01:44:53.780 AI 521, their understanding of the human condition is wonderful. I cry when I read their novels.
01:45:01.160 They'll be a part of the ecosystem. They'll be entities within us, whether they are self-aware or not,
01:45:06.440 will become a philosophical question. Let's not go that narrow path, that disgusting rabbit hole
01:45:11.540 where I wonder, does Peter actually have consciousness or not? Does he have the same
01:45:15.820 processes as I do? We won't know that about these, or maybe we will, but will it matter
01:45:21.060 if they're just among us? And they'll have brands, they'll have companies around them.
01:45:26.420 They'll be superstars. And they'll be Dr. Fubar from Kansas, trained on iridivic medicine,
01:45:37.580 the key person for alternative medicine, not a human, but we love what they do.
01:45:43.260 Okay. Last question. How long until, from at least an intellectual perspective, we are immortal?
01:45:51.400 So if I died today, my children will not have access to my thoughts and musings any longer.
01:46:00.940 Will there be a point at which during my lifetime, an AI can be trained to be identical to me,
01:46:10.320 at least from a goalpost perspective, to the point where after my death, my children could say,
01:46:17.980 dad, what should I do about this situation? And it can answer them in a way that I would have.
01:46:25.940 It's a great question because that was an early business plan that was
01:46:29.940 generated shortly after GPT-4 came out. In fact, I was talking very briefly to Mark Cuban
01:46:36.240 because he saw GPT-4. I think he got trademarks or copyrights on his voice,
01:46:42.180 all his work and likeness so that someone could not create a mark who responded in all the ways he
01:46:50.440 does. And I'll tell you that it sounds crazy, but there's a company called rewind.ai. And I have
01:46:58.580 it running right now. And everything that appears on my screen, it's recording. Every sound that it
01:47:06.720 hears, it's recording. And if characters appear on the screen, it'll OCR them. If a voice appears,
01:47:13.460 and then if I have a question, I say, when did I speak with Peter Atiyah? They'll find it for me.
01:47:18.540 I'll say, who was I talking about AI and Alzheimer's? And they'll find this video on a timeline.
01:47:28.180 How many terabytes of data is this, Zach?
01:47:30.740 Amazingly small. It's just gigabytes.
01:47:33.780 How is that possible?
01:47:35.320 Because A, it compresses it down in real time with using Apple Silicon. And second of all,
01:47:40.360 you and I, you're old and you don't realize that gigabytes are not big on a standard Mac that has
01:47:46.880 a terabyte. That's a thousand gigabytes. And so you can compress audio immensely. It's actually not
01:47:54.120 taking video. It's just taking multiple snapshots every time the screen changes by a certain amount.
01:47:59.020 Yeah. It's not trying to get video resolution per se.
01:48:02.020 No. And it's doing it. And I can see a timeline. It's quite remarkable. And so that is enough,
01:48:09.240 in my opinion, data so that with enough conversations like this, someone could create
01:48:14.840 a pretty good approximation of at least public Zach.
01:48:18.780 So then the next question is, is Zach willing to have rewind AI on a recording device, his phone
01:48:26.660 with him 24 seven in his private moments, in his intimate moments, when he is arguing with his
01:48:33.400 wife, when he's upset at his kids, when he's having the most amazing experience with his postdoc. Like
01:48:40.120 if you think about the entire range of experiences we have from the good, the bad, the ugly, those are
01:48:45.800 probably necessary. If we want to formulate the essence of ourselves, you envision a day in which
01:48:51.420 people can say, look, I'm willing to take the risks associated with that. And there are clear risks
01:48:55.900 associated with doing that, but I'm willing to take those risks in order to have this legacy,
01:49:02.800 this data set to be turned into a legacy.
01:49:05.300 I think it's actually pretty creepy to come back from the dead to talk to your children.
01:49:09.600 So I actually have other goals. Here's where I take it. We are being monitored all the time.
01:49:14.740 We have iPhones, we have Alexa devices. I don't know what is actually being stored by whom and what.
01:49:20.600 And people are going to use this data in ways that we do or don't know. I feel it's us,
01:49:26.140 the little guy, if we have our own copy and we can say, well, actually, look, this is what I said then.
01:49:32.160 Yeah. That was taken out of context.
01:49:33.480 Yeah. That was taken out of context. And I can do it. I have an assistant that can
01:49:37.980 just find it and find exactly and find all the times I said it. I think that's good. I think it's
01:49:45.240 messing with your kid's head to have you come back from the dead and give advice, even though
01:49:49.820 they might be tempted. Technically, I think it's going to be not that difficult. And again,
01:49:55.080 speaking about Rewind AI, again, I have no stake in them. I think I might have paid them for a license
01:50:01.260 to run on my computer, but the microphone is always on. So when I'm talking to students in
01:50:07.100 my office, it's taking that down. So there are some moments in my life where I don't want to be
01:50:12.860 on record. There are big chunks of my life that are actually being stored this way.
01:50:16.780 Well, Zach, this has been a very interesting discussion. I've learned a lot.
01:50:20.100 I probably came into this discussion with about the same level of knowledge, maybe slightly more
01:50:25.540 than the average person, but clearly not much more on just the general principles of AI, the
01:50:30.760 evolution of AI. I guess if anything surprises me, and a lot does, but nothing surprises me more
01:50:36.780 than the timescale that you've painted for the evolution within my particular field and your
01:50:43.120 particular field, which is medicine. I had no clue that we were getting this close to that level of
01:50:53.260 intelligence. Peter, if I were you, this is not an offer because I'm too busy, but you're a capable guy
01:50:59.260 and you have a great network. If I was running the clinic that you're running, I would take advantage
01:51:03.360 of now. I would get those videos and those sounds and get all my patients with, of course, their
01:51:11.300 consent to be part of this and to actually follow their progress, not just the way to report it,
01:51:17.560 but by their gait, by the way they look. You can do great things in what you're doing and advance the
01:51:24.200 state of art. You're asking who's going to do it. You're doing some interesting things. You could be
01:51:29.180 pushing the envelope using these technologies as just another very smart, comprehensive assistant.
01:51:36.200 Zach, you've given me a lot to think about. I'm grateful for your time and obviously for your
01:51:40.640 insight and years of dedication that have allowed us to be sitting here having this discussion.
01:51:45.320 Thank you very much. It was a great pleasure. Thank you for your time.
01:51:48.900 Thank you for listening to this week's episode of The Drive. It's extremely important to me to
01:51:53.880 provide all of this content without relying on paid ads. To do this, our work is made entirely
01:51:58.860 possible by our members. And in return, we offer exclusive member-only content and benefits above and
01:52:05.500 beyond what is available for free. So if you want to take your knowledge of this space to the next
01:52:09.820 level, it's our goal to ensure members get back much more than the price of the subscription.
01:52:14.980 Premium membership includes several benefits. First, comprehensive podcast show notes that detail
01:52:21.100 every topic, paper, person, and thing that we discuss in each episode. And the word on the street is
01:52:26.940 nobody's show notes rival ours. Second, monthly Ask Me Anything or AMA episodes. These episodes are
01:52:34.820 comprised of detailed responses to subscriber questions typically focused on a single topic
01:52:39.680 and are designed to offer a great deal of clarity and detail on topics of special interest to our
01:52:45.040 members. You'll also get access to the show notes for these episodes, of course. Third, delivery of our
01:52:50.940 premium newsletter, which is put together by our dedicated team of research analysts. This newsletter
01:52:56.320 covers a wide range of topics related to longevity and provides much more detail than our free weekly
01:53:02.560 newsletter. Fourth, access to our private podcast feed that provides you with access to every episode,
01:53:09.440 including AMA's sans the spiel you're listening to now and in your regular podcast feed. Fifth,
01:53:16.360 the Qualies, an additional member-only podcast we put together that serves as a highlight reel featuring
01:53:22.560 the best excerpts from previous episodes of The Drive. This is a great way to catch up on previous episodes
01:53:28.280 without having to go back and listen to each one of them. And finally, other benefits that are added
01:53:33.100 along the way. If you want to learn more and access these member-only benefits, you can head over to
01:53:38.660 peteratiamd.com forward slash subscribe. You can also find me on YouTube, Instagram, and Twitter,
01:53:45.600 all with the handle peteratiamd. You can also leave us a review on Apple Podcasts or whatever podcast
01:53:52.160 player you use. This podcast is for general informational purposes only and does not
01:53:57.380 constitute the practice of medicine, nursing, or other professional healthcare services,
01:54:01.320 including the giving of medical advice. No doctor-patient relationship is formed. The use
01:54:07.140 of this information and the materials linked to this podcast is at the user's own risk. The content
01:54:12.920 on this podcast is not intended to be a substitute for professional medical advice, diagnosis, or treatment.
01:54:18.360 Users should not disregard or delay in obtaining medical advice from any medical condition they
01:54:23.560 have, and they should seek the assistance of their healthcare professionals for any such conditions.
01:54:28.960 Finally, I take all conflicts of interest very seriously. For all of my disclosures and the
01:54:34.080 companies I invest in or advise, please visit peteratiamd.com forward slash about where I keep an up-to-date
01:54:42.040 and active list of all disclosures.
01:54:48.360 Thank you.