ManoWhisper
Home
Shows
About
Search
TRIGGERnometry
- July 05, 2023
Richard Dawkins: God, Truth & Death
Episode Stats
Length
1 hour and 6 minutes
Words per Minute
165.08073
Word Count
10,918
Sentence Count
586
Misogynist Sentences
7
Hate Speech Sentences
22
Summary
Summaries generated with
gmurro/bart-large-finetuned-filtered-spotify-podcast-summ
.
Transcript
Transcript generated with
Whisper
(
turbo
).
Misogyny classifications generated with
MilaNLProc/bert-base-uncased-ear-misogyny
.
Hate speech classifications generated with
facebook/roberta-hate-speech-dynabench-r4-target
.
00:00:00.680
Do you not think that there is a fundamental part of us as human beings that needs religion?
00:00:06.000
You could say that there's a deep psychological need for religion.
00:00:12.400
The comfort that you get from believing a falsehood is like a drug and it's a perfectly
00:00:20.320
valid argument to say that there's everything to be said for the drug.
00:00:24.240
I just want to make a distinction between what is true and what humanity needs and it
00:00:28.240
may very well be that we do need false ideas in order to flourish and prosper but I'm also
00:00:35.200
interested in what's true.
00:00:36.880
A lot of people are zipping their mouths because of political pressure and if people for political
00:00:44.320
reasons are trying to deny that there really is a binary separation between the two sexes
00:00:49.680
then that is anti-scientific, anti-rational and is a subversion of language actually.
00:00:58.240
Hey Francis, do you think financial platforms should be apolitical and not cancel people
00:01:10.240
just because they don't agree with their politics?
00:01:12.160
I'll never forgive what those absolute f***ing b****** that Tide Bank did to us.
00:01:17.120
Now Francis, we don't know if our bank account was cancelled because of our politics.
00:01:22.000
Give me five minutes in a room with them and I'll find out.
00:01:24.800
What are you going to do? Test your new jokes on them?
00:01:26.800
Yes, and my Celine Dion impersonation.
00:01:29.200
Don't want to see that.
00:01:34.640
Moving swiftly on, if you're looking for a crowdfunding platform then you have to use GiveSendGo.
00:01:40.560
GiveSendGo is a crowdfunding platform that is available in over 80 countries
00:01:45.120
and provides a simple and easy ways to raise money online.
00:01:48.640
They are politically neutral and don't remove campaigns based on political or ideological
00:01:54.560
differences. We all know that a lot of crowdfunding platforms cancel people if they don't agree with
00:02:00.400
their politics. The more we support companies providing alternative models, the more we weaken
00:02:06.240
the power of cancel culture. What's more, GiveSendGo is a free platform powered by donations,
00:02:12.000
which means that you get to keep more of the money you raise. Other crowdfunding sites charge
00:02:17.120
between 5% and 10%. 10% of the total money raised is a huge amount. GiveSendGo can be used to raise
00:02:24.000
funds for medical expenses, business ventures, personal needs, churches, non-profits, funeral costs,
00:02:30.240
and much more. So let's take a stand. Don't give your business to companies that have made it clear
00:02:35.600
they don't want you. Choose hope, choose freedom, choose GiveSendGo. Go to www.givesendgo.com
00:02:44.400
and check out a better alternative to crowdfunding. That's www.givesendgo.com and support the people
00:02:52.000
who support freedom. Hello and welcome to Trigonometry. I'm Francis Foster. I'm Constantine
00:02:58.880
Kishin. And this is a show for you if you want honest conversations with fascinating people. Our
00:03:04.240
terrific guest today is a world-famous evolutionary biologist, prolific author, and one of the four
00:03:09.680
horsemen of the New Atheist Movement, Richard Dawkins. Professor Richard Dawkins, welcome to Trigonometry.
00:03:14.240
Thank you very much. Oh, it's such a pleasure to have you on the show. As I said, you're world-famous,
00:03:18.720
but for our audience who will already know some of your story. Tell us a little bit about the
00:03:22.720
background. Who are you? What's been your journey through life? How do you end up sitting here,
00:03:27.120
a set of huge accomplishments throughout your life, talking to us? How far do you want to go?
00:03:32.960
Back to birth? As far back as you want to go? Yes. I was born in British colonial Africa,
00:03:39.920
and my father was in the colonial service. And our family came to England when I was eight.
00:03:47.600
And then I went to boarding school, and then Oxford. And at Oxford, I was inspired by a world-famous
00:03:57.040
biologist called Nico Tinbergen, who later got a Nobel Prize. And I did my doctorate under him, and
00:04:03.680
have done, been in academic life ever since. And first of all, I went to University of California for
00:04:09.200
two years, a very junior assistant professor. And then I came back to Oxford, where I've been for
00:04:16.720
all my life. I was a tutor at Oxford, doing the rather unique Oxford tutorial system. And I started
00:04:24.800
writing books. And I wrote a lot of books, which for a general audience, which many of them sold very
00:04:29.840
well. And then I was given a professorship with a public understanding of science, the Charles
00:04:34.560
Symonia Professorship, of the public understanding of science. And then I had to, that was my job then,
00:04:39.600
was to communicate science. And pretty much that's what I've been doing, is communicating science.
00:04:44.560
I've written about 16 or 17 books. And retired, ooh, about 10 years ago now, longer,
00:04:53.040
and carried on writing books. And that's where I am.
00:04:55.280
It is indeed. And one of your first, or was it the first book, The Selfish Genius?
00:04:59.280
Yes.
00:04:59.600
That's your very first book. That was a book that revolutionized as a young student,
00:05:03.600
I remember reading it, and understanding things that I never understood. And some of the challenges
00:05:09.200
then to the idea, to the evolutionary theory, for example, altruism, were beautifully addressed in that,
00:05:14.720
in a way that even someone who's not an expert could understand. And you wrote that book almost half
00:05:20.800
a century ago now, in 1976. I'm very pleased to hear you say how much that influenced you, because
00:05:27.120
it sums up my attitude to biology. It's somewhat misunderstood, mainly by people who've read it
00:05:35.600
by title only, who think it's a book about selfishness, or even an advocacy of selfishness,
00:05:41.840
which of course it isn't, quite the contrary. And yes, it's coloured my whole
00:05:50.560
career, really.
00:05:51.680
And one of the things I wanted to ask you is, as I say, that was 1976, or nearly half a century ago.
00:05:58.480
What things, what discoveries have been made, what theories have been posited that you think,
00:06:04.480
in your field, and also more broadly, perhaps in science, have emerged in that half century,
00:06:09.440
or nearly half century, that you think are some of the most crucial things for human flourishing,
00:06:13.920
human development, that we've made in that period?
00:06:17.920
I suppose the main thing is the flourishing of molecular biology, because, I mean, it was
00:06:24.400
it was written long after Watson and Crick discovered the DNA double helix structure.
00:06:30.160
But that led to the to the unravelling of the genetic code, and the realising that biology is
00:06:37.040
fundamentally digital. And that's what's really going now in biology is dominated by this digital
00:06:44.160
view of genetics. Everything is digital. And some people have asked me, has that changed the selfish
00:06:53.120
gene? Would I would I change it if I rewrote it now because of the revolution in genomics? And I
00:06:58.800
think I wouldn't. It's still, um, it's still valid as far as evolution is concerned. I mean,
00:07:06.240
evolution is the differential survival of genes in gene pools. And that's,
00:07:11.840
could have been said in the 1930s, actually. Well, it pretty much was.
00:07:15.920
So, Richard, my question is, how much does biology dominate our lives, as in our own lives? Does it
00:07:25.680
dominate everything? Do we have free will? Or is it, are we just purely a product of our genetics?
00:07:31.600
Well, free will is a question that doesn't necessarily need a biologist to answer it.
00:07:37.600
It's a deep philosophical question. And it's not a question that I think you can answer by
00:07:44.320
looking at genetics. If we have free will, or we don't have free will, that won't be influenced by
00:07:52.000
looking at genes. You can say, um, nothing that I do is anything other than predetermined by the
00:08:01.680
things that happen in the world. The molecules in my brain, everything is predetermined. And genetics
00:08:09.600
will not just be a part of that. So there's not really a question for a geneticist. It's a question for
00:08:13.520
a philosopher, I suppose, um, or any scientist. And it's not a thing that you'll, that a geneticist
00:08:23.040
has any specific, um, input on. And is it, what, what revolution that can you see coming or is
00:08:32.480
happening at the moment that you actually get excited by? Because that's the exciting thing about
00:08:36.400
science, the changes, the innovation, the discoveries that we're seeing at the moment. Well, I mentioned
00:08:42.000
molecular genetics and, and that's happening all the time. It's not, it's not a single revolution.
00:08:46.240
It's just going on and on and on. Um, and so, um, the fact that you can actually read very swiftly
00:08:53.760
nowadays, you can actually read the genetic code of any animal. It's always the same genetic code.
00:08:58.960
And what it says is in detail different. And so you can take any animals you like and read the genetic
00:09:05.360
code, any people you like, and read the genetic code and compare it letter by letter, line by line,
00:09:11.120
just as you might compare two manuscripts. And that's an astonishing thing. I mean,
00:09:16.320
that, that would have amazed Darwin, I think. And that means that you can reconstruct the whole
00:09:23.840
tree of life, the whole pedigree, the whole family tree of all living things, um, in, in minute detail.
00:09:31.440
You can say, which is the, which pairs of animals and plants are the most of the closest cousins.
00:09:36.160
You can estimate how far back they had a shared ancestor. That's isn't, that's only an approximate
00:09:42.000
estimate, but, but you can roughly lay out the complete history of the branching tree of life
00:09:49.600
by looking at these molecular genetic, um, sentences.
00:09:55.360
Richard, and for someone like you who is a scientist and who is, uh, I, I think there's a sort of
00:10:02.000
wondrous appreciation of all these things and the beauty of the natural world and so on.
00:10:06.480
Uh, what are some, for, for our audience and, and for us actually, what are some of the practical
00:10:11.120
applications that come out of these discoveries that we may not yet have seen when we go to the
00:10:16.640
doctor? We may not yet have experienced in our own lives as a thing that's become part of it.
00:10:22.160
What is coming as a product of this, uh, steady, slow, but ongoing revolution?
00:10:27.840
Hitherto, doctors have treated us all as pretty much the same. I mean, maybe a bit different male
00:10:33.520
and female, old and young, but apart from that, if you've got a certain disease, you get a certain,
00:10:39.360
a certain medicine, a certain treatment. Um, what will come is when doctors are able to read the, um,
00:10:47.760
if they can in principle do it now, but when it becomes cheap enough that a doctor will know
00:10:52.480
the genome of each patient, then the prescription that the doctor will give for any particular disease
00:10:59.760
will vary depending upon the, uh, the genotype of the individual patient. Um, in a very sort of
00:11:07.600
crude way, we already have this when certain particular diseases are known to be caused by
00:11:12.320
certain particular genes. Um, but when everybody's patient, when everybody, every patient's genome
00:11:19.120
is on file in the doctor's computer, say, okay, well, I look at me as I see your genome, so you
00:11:24.080
better have this, this drug and the next patient, you better have this drug. Um, I think that that's
00:11:29.920
the main thing I can think of immediately off the top of my head in answer to your question.
00:11:35.200
Richard, are you worried that, no, no, for, from a medical point of view, that sounds incredible.
00:11:40.480
I'm like, of course, that, that is wonderful. That's what everybody should have access to.
00:11:44.480
But do you worry as well that this type of information could be used in a, in a way to
00:11:49.040
discriminate against people, for example, to prevent people getting health insurance because
00:11:52.880
you might have like the Baraka gene and that means that you have a 90% chance of developing breast
00:11:57.840
cancer. That's a big worry. And, uh, it's one of the things that needs to be looked at very carefully
00:12:03.360
because, um, life insurance companies at present have a very crude way of, um, doing their actuarial
00:12:11.280
calculations. They look at your age, they look at whether you smoke, whether you drink to it to
00:12:16.000
excess, that kind of thing. Um, but, um, apart from that, we're all treated pretty much equally.
00:12:23.360
And that means that, um, that those of us who are, um, who are, who are healthier in a sense subsidizing
00:12:28.880
those of us who get ill and that that's the way insurance works. If, if it came to it that, um,
00:12:37.600
genetic, uh, knowledge enabled actuarial calculators to say, okay, he's got, he's got another 25 years
00:12:47.200
to live. He's got only 15 years to live. Um, then life insurance would become impossible. I don't think
00:12:53.280
it'll come to that because it's not that predictable, but nevertheless, um, there would
00:12:58.080
have to be careful regulation. I'm not sure how that would work. I suppose to prevent insurance
00:13:04.960
companies getting access to data, which, which would have to be kept confidential. That would
00:13:10.720
be one way to do it. Uh, another where I, I'm not quite sure what, how, what the legality of, of it
00:13:15.760
would be, but that, that is a big worry. And, and also as well, what it brings into, I mean,
00:13:21.520
the term designer babies is, is used. I think it's more of a kind of tabloid description,
00:13:26.400
but we know what we mean, where people, where babies are then having screened for certain
00:13:31.680
genetics and certain genes are taking, are taken out or eliminated or played with. Do you think
00:13:37.600
that scientists playing God, or do you think that's actually where we need to get to as a species,
00:13:44.080
so thereby eliminating certain illnesses, diseases? I think it's hard to find an objection to eliminating
00:13:50.240
diseases. So I think as long as that's used, not for making designer babies, but for, um,
00:13:58.240
screening against hereditary diseases, it seems to me that that's really hard to object to that.
00:14:04.800
Richard, sorry to interrupt. Do you not think there's a moral conundrum there? Because it kind
00:14:08.800
of depends where you draw that line, doesn't it? For example, if you have a baby that's, I don't know,
00:14:13.680
has a higher risk of cancer. Now, I have a young son, I wouldn't want him to have a higher risk of
00:14:20.320
cancer. But if I could then, you know, select him from a number of fetuses and discard the others,
00:14:25.360
wouldn't that be maybe perhaps too far going in that direction? Well, I don't think necessarily so.
00:14:31.920
I suppose the way it might work, this would be something like this. In IVF, in vitro fertilization,
00:14:37.600
the woman is given hormone treatment to super ovulate. So she gets maybe a dozen eggs,
00:14:44.800
and the dozen eggs are all in a Petri dish. And what happens at present is that the doctor picks
00:14:51.200
out one of these eggs at random, well, they all get fertilized if they can, picks out one of the
00:14:56.880
zygotes at random and re-implants it in the woman. Well, um, if the, instead of picking at the
00:15:07.440
zygote at random, you, you, you examine the genes, and this can be done when it reaches,
00:15:12.560
say the eight cell stage. So you let them develop to the eight cell stage. So they are eight cell
00:15:16.720
embryos. And then you can take out one cell of the, of the eight, and look at the genes of that,
00:15:23.040
and it doesn't damage it. And you say, oh, okay, this one has the gene for haemophilia,
00:15:29.360
and this one doesn't. Why on earth would you choose at random if you know
00:15:33.840
that some, that half of them have the haemophilia gene and half of them don't?
00:15:38.960
Obviously, it makes moral sense to pick one of the half, one of the 50% that do not have the gene
00:15:45.600
for haemophilia, whatever it might be. I suppose the question that it brings into,
00:15:48.960
and this is part of the other conversation we want to have with you, um, in terms of atheism and
00:15:54.560
religion, it brings into question, uh, life. What is life? Is, is a eight cell zygote life? Can it be
00:16:01.360
discarded? Uh, is that, that, that's where the moral conundrum comes in? I have no time for those,
00:16:07.120
that kind of argument about, is this a human, human life? You don't really? No. Why not? Well,
00:16:12.400
because this eight cell zygote has no feeling, it has no nervous system, it has no capacity to suffer,
00:16:19.360
no capacity to feel pain. Um, so that there's, I mean, no, no moral difficulty at all in choosing.
00:16:25.920
You're going to pick one of them anyway. You want a dozen of them,
00:16:28.640
and the, and the 11 that you, that you don't choose are going to be flushed away anyway.
00:16:34.000
So, um, you've already taken that moral decision in a way. It's a question of whether you choose one
00:16:38.880
at random, whether you choose the one that, that does not have the lethal or sublethal.
00:16:44.960
But what I mean is once you've created that selection process, given that parents would want
00:16:49.680
their children to be happy, but also clever and charismatic.
00:16:54.000
Now you've moved on to quite a bit of a different topic. Now you've moved on to...
00:16:57.440
But that's where I was going with it, right? Okay, okay.
00:16:59.360
So, and that's, I think, the core of Francis' question too is,
00:17:01.760
if we've got the opportunity to select the optimum baby for us,
00:17:05.760
are we not going to end up in a position where parents are encouraged to go down that path?
00:17:10.240
Well, yes. Incentivized even.
00:17:11.600
But you've changed the subject, and I, I, I was talking about the removal of...
00:17:14.960
Of course. Yeah.
00:17:15.600
I'm not trying to catch you out. I'm trying to get to the bottom.
00:17:17.360
No, no, no. It's a very interesting question.
00:17:18.720
Yeah. And then you can say something like, um, if it became possible to know that of these,
00:17:25.280
of these dozen zygotes that we've got, um, some of them have a gene for musical genius,
00:17:34.080
and some of them don't, um, we're nowhere near that at the moment. But, but if, if in the future that,
00:17:39.520
that came about, then would you worry about parents who say, I want my child to, to be a great musician,
00:17:46.560
or a great composer, a great mathematician. Um, and that's very interesting. That's a more difficult
00:17:53.280
moral dilemma. Um, if, if you object to that, then I might say to you, why do you object to that
00:18:02.720
when you don't object to parents, even forcing their child to practice the piano? See, you know,
00:18:11.040
you, you haven't done your music practice today. Get on with it. You're supposed to practice,
00:18:15.200
because the only way you will become a great pianist is by practicing six hours a day, whatever it is.
00:18:20.880
Um, why would you think that is, uh, okay, when you would not think it okay to pick out
00:18:28.880
a zygote from a Petri dish, which contains a genetic predisposition to be a great musician?
00:18:34.480
I'm not answering the question, but I'm saying that, in a way, these sorts of moral questions
00:18:39.280
can best be answered by comparison with other, other things. And I'm just comparing this case,
00:18:45.120
asking sort of question a moral philosopher might ask. Um, what's the difference between
00:18:51.520
encouraging a child to practice the piano and giving the child a head start by giving it the
00:18:57.440
gene of the, of the ones that are available in the Petri dish in an IVF situation, um, to become
00:19:05.280
a musician? Well, I suppose the problem with, with the answer to that is we come back into the realm of
00:19:11.760
religion and morality, because I think at the core of it is the idea that the creation of life is some
00:19:18.480
kind of random miracle. It's supposed to be sort of, it's, it's not supposed to be is a word that I've
00:19:26.640
introduced of course, but if we, if we call it a miracle, um, it's supposed to be random to some
00:19:33.520
extent. It's supposed to be down to chance when you, when we start interfering with that selection
00:19:38.000
process. Instinctively, I, I don't, I can't explain it perhaps. I feel a certain, there's something off
00:19:44.560
about that for me. I know that's not a very rational or scientific argument, but I do think that's one a
00:19:50.560
lot of people share actually. Well, it's not rational. No, I acknowledge that. It sounds as though you're
00:19:54.800
religious. It does, it does, which I'm not interestingly, but I feel that that is probably
00:19:59.600
how a lot of people do think about it. Yes, I suppose they do. Um, I think it's more of a worry
00:20:07.440
would be if some mad dictator, some kind of Hitler, started using these techniques and, and mandating,
00:20:13.920
you know, this, the selection of blonde, blue eyed, Aryan, you know, um, the sort of thing Hitler
00:20:24.240
might have, might have selected. Um, I was talking about more giving parents the opportunity
00:20:30.160
to have their, um, to do their IVF in a, in a non-random way. I mean, to, I, I can sort of see
00:20:37.920
that both ways. I don't, I'm like, but on the other hand, if it, if you have a dictator who says,
00:20:42.800
we want to breed in this country, a race of people who are, who are of such and such a type,
00:20:48.960
um, that I think is deeply sinister. I would agree with you. What would you say as well,
00:20:53.600
Richard? Because I think about this a lot. So when I was a teacher, I taught, for instance,
00:20:58.560
uh, children who were on the autistic spectrum and they were quite far on the autistic spectrum,
00:21:04.080
so they were non-verbal, but I also saw children who were autistic and as a result of their autism,
00:21:11.360
some of them were highly gifted in mathematics and would go on to be scientists, physicists,
00:21:17.200
et cetera, et cetera. You see artists who have depression, for example, very severe depression,
00:21:23.520
but that almost gives them an insight into the human condition and human suffering.
00:21:28.240
My worry is with this, we go, okay, we're going to get rid of depression. We're going to get rid of
00:21:31.280
autism. And then you could argue, well, look, we're going to get rid of, we're going to lower
00:21:34.800
suicide rates. As a result of that, we're going to, non-verbal autism will be eradicated completely.
00:21:41.200
But also, aren't we going to get rid of the outliers, the, the, the great thinkers,
00:21:46.000
the people who look at the world in a different way that move our, our species forward?
00:21:50.240
That's an excellent point. Um, and, and it's one that I should have come onto anyway. Yeah.
00:21:54.480
Um, when you do any kind of selection for any one characteristic, you're in danger of
00:22:00.160
having side effects, which you never, which you never thought of. So if, if you select for, um,
00:22:06.240
to, to, to remove some kind of psychological problem, then it might, could well be that you're
00:22:11.520
then having a side effect of, of, of never getting any, any genius mathematicians of something of that
00:22:17.040
sort, is what you're saying. I think that's a very good point. And, um, it's what, it's one of the
00:22:22.240
problems with any kind of eugenic selection is that, is that you, you don't know what the, what the
00:22:27.840
byproducts are going to be of the selection that you're, that you're engaged upon. And that's
00:22:32.160
what I mean when, when you have a dictator who says, we're going to breed for such and such.
00:22:36.560
You see this with, with breeding animals, you, you breed for racehorses who have very, can run very
00:22:42.400
fast and, and they break their legs. I mean, they've, they've, they've, they've become more vulnerable
00:22:48.000
to, um, leg breakages because you're breeding for one thing, namely speed.
00:22:52.480
And in nature, that doesn't happen. In nature, it's, it would be counterbalanced by selection
00:22:59.360
against raking the legs. Well, you're only using a human example, which is much more serious,
00:23:04.080
um, where, um, geniuses, outliers, as you rightly said, um, are quite likely to be selected against,
00:23:14.640
if you naively, um, go for removing certain psychological problems. So there's, there's,
00:23:25.040
there is something to be said for letting nature take its course. But I think I come back to the
00:23:31.040
negative things of these, these sort of genetic diseases, which are, which are, I think anybody
00:23:36.320
would agree are undesirable, like haemophilia. Um, and, and it seemed, it seems to me that taking a moral
00:23:44.960
letting that your moral concerns overspill into forbidding the removal of genetic diseases
00:23:53.920
like haemophilia, that, that are obviously negative in all respects, um, that, that, that's going too far.
00:24:01.040
And Richard, we have a lot of scientists saying that they're worried about where science is going,
00:24:08.240
the state of science in general, free speech in universities, scientists no longer being able to,
00:24:14.560
allowed to explore certain avenues. How do you see the field of sciences at the moment? Are you
00:24:19.360
optimistic or are you a little bit more cautious? Um, what examples are you thinking of where scientists
00:24:24.560
are not allowed to? Well, we, we, for instance, when it comes to talking about, but, uh,
00:24:30.880
the sexes and saying, you know, there, there, there's differences between the sexes investigating
00:24:37.520
the differences between the sexes. Certain scientists would say, look, I don't feel able to
00:24:42.960
investigate that anymore because people will complain. People will say certain things about me
00:24:48.800
that, you know, that I'm denying the existence of trans people, for example.
00:24:53.040
Yes. I think you're right that, um, not just scientists, but, but, um, a lot of people are zipping
00:24:58.880
their mouths, um, because of, um, a political pressure of the sort that you're saying. Um,
00:25:06.560
and it, it is important, I think, for scientists to, to be honest and to, uh, use language precisely.
00:25:14.640
And in a particular case of sex that, that you mentioned, it's one of the few cases where there
00:25:18.400
really is a, a, a, a bifurcation, a binary bifurcation. There really are two sexes. And scientists
00:25:27.360
have to, have to work on, under that, that fact. And if, uh, people for political reasons are trying
00:25:33.920
to deny that there really are, there really is a binary separation between the, between the two
00:25:38.800
sexes, then that is anti-scientific, anti-rational, and is a subversion of language, actually.
00:25:45.200
Richard, uh, I, I actually didn't want to make this conversation in any way about the cultural
00:25:52.240
discussions around that issue. But since, since we've come to it, do you sit there and sometimes
00:25:58.320
have to pinch yourself that you, one of the most eminent evolutionary biologists in our society,
00:26:04.480
go on the, on, on the mainstream media and you are asked to talk about the fact that there's men and
00:26:10.800
women. Do you not, do you not feel, do you not experience that as like a gigantic regression?
00:26:15.600
Richard, I do. I, I, I, I do because, um, I, I've, um, one of the main points I, I, I like to make
00:26:23.200
is that very often, um, I, I call it the tyranny of the discontinuous mind. We're, we're far too fond
00:26:29.200
of making discriminations, um, things like, um, well, we were talking a bit earlier about, um, embryos.
00:26:37.920
When does, when does an embryo become human? Is there a particular moment when an embryo becomes human?
00:26:42.880
No, there isn't. It's a continuum. It's a sliding scale. There are sliding scales everywhere.
00:26:48.160
We, in universities, when we examine students, we give them a degree first, second, two, one,
00:26:55.040
two, two, third. We, we, we insist upon making a divide between one class and another. We know
00:27:01.760
perfectly well that the top of one class is closer to the bottom of the one above than it is to the
00:27:08.720
bottom of its own class. And yet the information about the sliding scale is actually in bell-shaped
00:27:15.200
distribution in this case, um, is thrown away when we divide. We insist upon making divide. So,
00:27:22.240
um, the tyranny of the discontinuous mind is one of my catchphrases. And if I look around and say,
00:27:28.720
is there any case where there really is a proper divide, a real, where there really is no spectrum,
00:27:34.560
spectrum. And sex is, is the one thing I can think of. There really isn't a spectrum. You really are
00:27:39.440
either male or female. And so, um, I do have to pinch myself when, um, for once, uh, it goes the other
00:27:50.080
way. I mean, there, there really is no spectrum there. And, um, yes, I, I, I do have to pinch myself.
00:27:56.320
And where do you think this regression, if you agree with my use of it, where does that come from?
00:28:02.240
Well, in certain social sciences, um, they've been very influenced by what they call postmodernism.
00:28:07.840
I don't really know what postmodernism is. In fact, I think the so-called postmodernists don't know
00:28:12.720
either. But I think it's something to do with, um, words meaning something that is determined by
00:28:24.000
politics. And you don't, you don't actually have a fixed meaning of words like male and female,
00:28:32.720
because there's an intellectual movement in, in the social sciences that says that everything
00:28:36.960
is a social construct. And male and female is a social construct. There's no real, well,
00:28:42.320
validity to the, to difference in male and female. It's all a social construct, um, or cultural
00:28:48.880
relativism, the idea that, um, um, different cultures have completely different ways of classifying
00:28:54.400
the world. And so, and so it's just our white Western way of looking at things that, um, that says that
00:29:01.680
there are, there's a divide between males and females and something like, like that.
00:29:05.600
And Richard, with this social, uh, constructivism thing, my sense is that is,
00:29:12.320
the erosion of the concept of truth itself. Is, is that too strongly put or would you?
00:29:17.440
I think it's absolutely right to say that it is erosion of the concept of truth. I mean,
00:29:21.760
it, it is valid to say that language evolves and, and what words mean today is not the same as what
00:29:27.600
they meant 200 years ago necessarily or a thousand years ago. Um, and so we have to accept the fact that
00:29:32.880
words change them, their, their meaning, but we also have to live in a world where, um,
00:29:39.680
we have to be able to communicate. And so we have to be able to say that there are certain words that,
00:29:43.680
that, that, that at present means, mean such and such. At present, black means black and white means
00:29:48.320
white and blue means blue. And it, and it, if, if that changes in a hundred years, that's okay.
00:29:53.840
But for the moment, we've, we've got to use words that everybody understands. And words like male and
00:29:58.560
female are pretty clear in everybody's mind, what that means. And, um, to come along and say,
00:30:06.000
oh, that's just a social construct, um, is subversion of language, subversion of truth. I think you're
00:30:11.280
right. Do you think that this is a religious belief, Richard? I think it has a lot in common
00:30:17.360
with religion, with religious belief. Um, it's different in that it doesn't invoke anything
00:30:22.080
supernatural, but it's very similar in the ways that heretics are hunted down. We've seen this in
00:30:30.800
Oxford recently where Kathleen Stock has been, has been, um, vigorously, not almost violently,
00:30:38.080
sometimes they are violently hunted down. Then there's actually rather detailed parallels with,
00:30:44.640
with religious, um, dogma, um, the, the doctrine of, um, original sin. Um, Christians, especially Catholics,
00:30:57.440
believe that, um, we're all born in sin. We're all, we're all in, we all inherit the sin of Adam.
00:31:02.640
Um, although they no longer think Adam ever existed, but somehow they sort of managed to
00:31:06.640
talk and go on about talking about the sin of Adam. Um, and so we all, the moment we're born,
00:31:13.040
we're born in sin. We had to be baptized to be cleansed of sin, that, that kind of thing.
00:31:17.520
And I think we see that in the collective guilt that all white people are supposed to inherit
00:31:23.840
because of slavery. And, you know, even if they, obviously they have no direct connection with
00:31:30.480
slavery, maybe even if they never had an ancestor, most probably did, but, um, and we are all
00:31:40.080
collectively guilty for what people of, of our type did in some past age. And that is original sin.
00:31:49.840
I mean, it's exactly like original sin. Francis, may I jump in just very quickly, Richard? Yes.
00:31:54.320
You said there's nothing supernatural. Do you not think that the claim that you can change your sex
00:31:59.040
by means of incantation is a claim that is supernatural? I mean, yes, I mean, you could,
00:32:04.080
you can make that case. Um, it's not quite supernatural in the same sense as, as believing in gods or believing in,
00:32:10.960
in fairies. Um, but yes, it's sort of like that. Um, let's take another religious example, the Catholic belief in
00:32:21.920
transubstantiation where the consecrated wine and bread becomes the blood and body of Christ. And the,
00:32:33.760
the, the way they put this is, is to use the Aristotelian idea of substance, essence, something
00:32:42.560
and accidental. So, so, um, um, um, Aristotle made the distinction between the, the, the real substance
00:32:52.880
and, and the, and the accidentals. And so they, they say that while the accidentals of wine is still wine
00:32:59.440
in this sort of accidental sense, but in real substance, it becomes the blood of Christ. And
00:33:04.560
that's the verbal trick that they use to, to justify this ridiculous idea that the blood becomes the,
00:33:12.240
sorry, the wine becomes the blood of Christ. Um, so yes, that's very, very similar to, you're quite
00:33:18.080
right. That's very, very similar to saying that it's an incantation that you say, I stand up and say,
00:33:23.760
I am a woman. Therefore I am a woman. Therefore everything about me is, is a woman. You're not allowed to
00:33:29.360
misgender me. You're not allowed to say anything against the idea that I'm a woman. That's very,
00:33:33.760
very similar to, to the accidental and real substance argument of Catholics in the accidentals may, may,
00:33:42.080
may say I'm a man. I still have a penis, but in my real substance, I'm a woman. And that's very,
00:33:48.160
very similar. Yes. Richard. Now you, you're obviously a very famous atheist and you've written
00:33:54.960
fantastic works talking about your atheism. Do you not think that there is a fundamental
00:34:01.040
part of us as human beings that needs religion? Yes, very possibly there is. Um, it doesn't mean
00:34:08.320
it applies to everybody. It may mean that there's just a, um, it is quite difficult to eradicate when
00:34:15.040
you have something as fundamental as that. And, and yes. Um, and I suppose a psychologist could delve
00:34:21.920
into that and say something like you, we all need, um, to believe in something higher than ourselves.
00:34:29.360
And we have a desperate need to believe that we're not going to disfizzle out when we die. Um,
00:34:34.480
so yes, uh, you could, you could say that there's a deep psychological need for religion. It doesn't
00:34:42.640
mean everybody has it. It doesn't mean that you can't get out of it. Uh, but, but there's probably
00:34:47.600
there's something like that to explain it. And because being critical of organized religion,
00:34:54.720
as you were, was there ever, did you ever think that you might, that you went too far with your
00:35:01.360
criticisms of God in particular? Or would you look back on it and go, no, I was absolutely justified in
00:35:08.720
what I said at that particular time? Um, I, I've always tried to seek the truth and, uh, I've,
00:35:19.360
I've never wanted to be, um, I never wanted to upset people. Um, but I just think that we should be
00:35:25.680
free to argue points. And, um, so, uh, I don't think I've gone too far. You could possibly dig out
00:35:37.280
cases where, where are you, you might challenge me and I might say, yes, perhaps I did go too far there.
00:35:42.240
Um, but, um, I think in general, I've always just tried to be loyal to the truth. I think that
00:35:53.440
religious claims are not trivial. And I think that they're very interesting. Um, it, it,
00:35:59.920
it's in a way is one of the biggest questions that perhaps the biggest question there is in science is,
00:36:04.720
is, is does some kind of supernatural intelligence lurk behind the universe? Is the universe a planned,
00:36:15.760
um, entity that, that, that there's a super supernatural being conceived? I mean, that's a
00:36:24.480
profound scientific question. If it's true, then we're, then the entire universe we're looking at
00:36:30.320
is a very different kind of universe from one where there's nothing behind it. Um, and so it's a very,
00:36:36.320
very deep scientific question. Um, and I think the, the answer is clear. There is none, but I'm, I, I don't
00:36:46.880
write it off as something trivial. I don't write it off as a case where you can just say, um, it's not
00:36:53.520
interesting. Of course it's interesting. Um, it's interesting. A scientist has got to be interested
00:36:58.560
in this, this suggestion that the sort of universe we're studying is a planned, designed universe.
00:37:05.680
That's a very, very different kind of universe. So it's a very, it's a very deep question.
00:37:09.440
It is a very deep question because I, I sense, and I'm not a scientist, that the more you discover
00:37:15.040
about the world, the more you learn about the beauties, the details, the intricacies, the way things are
00:37:20.000
interlocked together. Isn't maybe a part of you that thinks this can't have been an accident.
00:37:27.120
There must be something else here. It's very, very tempting to say, yes,
00:37:32.320
everything works so perfectly together. And the, the, for me as a biologist, the complexity of life,
00:37:37.680
the, the enormous panoply of, of plants and animals and forests and birds and insects,
00:37:44.160
everything working together. No, it's, it's, it's a huge, beautiful construction. And what,
00:37:52.640
for me, what is absolutely marvelous is that nevertheless, there is a perfectly decent explanation
00:37:58.400
that it did all come about without, without planning. So one of the beauties of it is precisely
00:38:04.080
that we now can explain it or we're well on our way to, to explaining it without invoking any kind of
00:38:11.200
supernatural intelligence. And it is a great temptation because, um, we are so used to the
00:38:20.240
idea looking at our own machines, looking at things that we've made like computers and cars and planes.
00:38:26.320
And, um, these are clearly the result of design, deliberate design, deliberate construction.
00:38:35.840
And when you look at something like a bird's wing, compare it with an airplane's wing,
00:38:40.640
the temptation is huge to say, oh, they must both be designed. And it was the genius of Darwin to break
00:38:48.320
away from that and say, you actually know, there is a proper explanation. There is a, a materialistic
00:38:53.600
explanation for that. Um, and so, so the flip side of, of the, of the temptation is that when you've
00:39:01.200
overcome the temptation and worked out that it is possible to explain it in
00:39:07.920
simple scientific, I, I, I really mean simple because it, the ideas, the idea,
00:39:13.520
Darwin's idea is a deeply simple idea. And yet, given enough time, the Darwinian idea of natural
00:39:20.960
selection, given enough time, can build up to prodigies of complexity and beauty
00:39:27.920
and the illusion of design. And that's a, that's a measure of the genius of Darwin to see that.
00:39:34.880
And Richard, in your book, The Blind Watchmaker, you make that point, I think, beautifully.
00:39:39.520
And I do actually think that just slightly, no disrespect to France, is probably one of the
00:39:44.160
weaker arguments against your position. What I think is a stronger, stronger argument is the one he made
00:39:48.400
earlier. And this is about the psychological need people have for religion, but also at the level
00:39:54.560
of society. And this is really something I want to get into with you. Uh, uh, Noah Yuval Harari,
00:40:00.000
for example, in his book, Sapiens, his central argument is that the reason homo sapiens were able
00:40:05.520
to out-compete other, uh, species of human, uh, human, uh, humanoids, humanids? Hominids. See, I got both
00:40:13.760
wrong, hominids, was that, um, they were able to build shared myths that allowed them to create tribes that
00:40:20.800
went beyond the, the 150 limit of sort of being able to know everybody. And when we look at the
00:40:27.680
world today, uh, our generation, I, uh, when you're not in the room, I sometimes refer to us as, you know,
00:40:33.840
the children of Dawkins and the children of Hitchens in the sense that you took away with your beautiful
00:40:39.440
books, our ability to have that illusion that the world around us is, you know, this God created
00:40:46.560
mysterious place for which there's no other explanation. But as I look around at the world
00:40:51.920
with its inability to agree on what words mean, uh, what people call now the crisis of meaning,
00:40:59.760
where a lot of people are kind of lost. They don't know what the purpose and meaning of their life is.
00:41:05.120
Do you worry that maybe the truth is that for Richard Dawkins, your worldview is perfect because
00:41:11.840
you are Richard Dawkins, you're able to be inspired by science, by the beauty of the natural world,
00:41:16.880
by the beauty of the universe. But for a lot of other people, what they actually need is not necessarily
00:41:22.960
the belief in God so much as the social function that religion used to fulfill, which is to bind us
00:41:29.520
together with a set of shared values and a set of shared morals and a set of shared ideas about what
00:41:35.920
it means to be human, what it means to relate to other people. And without that, we are lost and
00:41:41.040
therefore we create these new religions that sort of tear our society apart.
00:41:46.720
I think that's very interesting and I'm rather persuaded by Harari's argument. But notice what
00:41:55.680
you've just done. I mean, you said maybe humans need religion, maybe they need something to bind
00:42:03.920
society together and function as a unit and so on. And maybe they do. And that doesn't make it true.
00:42:09.600
Oh, I agree with you. And of course you do. How modest of you.
00:42:15.280
No, I mean, it's just that you can make an utterly watertight argument where humanity needs something,
00:42:23.520
which is false. Oh, fine. And I just want to make a distinction between what is true and what
00:42:30.720
humanity needs. And it may very well be that we do need false ideas in order to flourish and prosper.
00:42:40.240
But I'm also interested in what's true. And so, I mean, I don't want to get involved in that sort
00:42:45.920
of selfish idea that it's okay for us intellectuals, but hoi polloi need religion.
00:42:54.160
I wouldn't consider myself an intellectual. So I am the hoi polloi in this context. But I'm just
00:43:01.360
saying to you, to me, truth is a supreme value. And I agree with you. But at the same time, to me,
00:43:07.600
the cohesion of our society, the fact that people are able to live a life of meaning and purpose,
00:43:13.760
they're able to connect to other human beings and be fulfilled in their lives and know what to do.
00:43:19.040
And, you know, you don't have people who are disappointed with the way that their life has
00:43:23.440
gone because of the choices that they made, because they didn't have someone encouraging
00:43:27.680
them down a particular path. That, to me, is also very tragic, as tragic as the erosion of the concept
00:43:33.440
of truth. And I feel there's a balance to be struck there. That's what I'm saying.
00:43:36.880
Yes. I think that's a very good point. And perhaps I should say, accepting the possibility of that,
00:43:44.640
I might also say that, actually, there's something wonderful about truth itself.
00:43:48.560
Of course. And so you can lead a very, very fulfilled life in the search for truth.
00:43:56.080
And... But that's my argument. You can. You can. You are a scientist who spent his whole life
00:44:02.960
pursuing the truth with a microscope and whatever other tools you use, right? I haven't. I can't.
00:44:09.440
That's not what I can do. And that's not what a lot of people can do. It's not because they're
00:44:13.280
more stupid than you. It's just because they're not scientifically minded. Maybe they,
00:44:16.960
they're musicians or whatever, and they don't have that same mindset as you do. They don't have
00:44:21.920
the same brain as you do. That's the argument. I can't get involved in using myself.
00:44:29.120
But, but, um, I just want to repeat that, that, that you don't act, well, okay, you don't actually
00:44:34.400
have to be a working scientist. You can, you can, you can revel in the beauty of scientific
00:44:39.680
understanding, just as you can revel in music without actually be able to play an instrument.
00:44:46.080
And, um, there is, there is immense fulfillment to be had in appreciation of understanding of the
00:44:56.000
universe in which you live and why you live here, or why you exist. That, that's a wonderful thing.
00:45:02.880
And I, I don't want to accept the idea that only certain people are capable of doing that. You
00:45:09.680
want to actually have a microscope in order to do that. You don't. You can read books. You can see
00:45:16.000
the cosmos television series. You can see David Attenborough's films. Um, and there's huge
00:45:22.880
satisfaction of, of the kind of thing that I suppose religion aspired to do in past centuries. Um, and
00:45:30.560
uh, you can, you can get it by
00:45:36.000
almost worshiping, not worshiping in the sense of worshiping the supernatural, but the,
00:45:40.640
the, the, the, the, the, the wonder of your own existence and the process that has given rise to
00:45:46.160
your, to, to, to you, that has led to your existence, which
00:45:51.840
wonder of wonders, we knew, we now pretty much understand. I mean, we, we, it's a privilege to be
00:45:56.960
alive in the 21st century and to be in a position to just read a few books and see a few television
00:46:05.200
documentaries and understand why you exist. That's never happened before.
00:46:09.040
I, I think the thing that, the thing that religion gives people is a sense of safety almost. So,
00:46:21.120
for instance, if when, when somebody is very ill and maybe they have cancer and they don't know
00:46:26.000
if they're going to live or not, this knowledge that there is a supreme being looking after them,
00:46:31.920
or that there's somewhere that they're going to go afterwards, provide someone with a deep sense
00:46:37.440
of comfort in the darkest moments of their life. And whilst I agree with everything that you said
00:46:43.360
about science and discovery and the wonder of looking at the universe, I don't think that provides
00:46:50.160
that particular emotion. Do you see? No, it's probably true. Uh, and, um, again, you can say, um,
00:46:56.560
um, the, the, the, the solace that you get, the comfort that you get from a belief, even if it's
00:47:02.320
false, it nevertheless is comforting. Um, uh, you've probably read Aldous Huxley's Brave New World.
00:47:09.600
Yes. Where, um, um, the, uh, towards the end, the, the Mr. Savage, the person who's brought from the
00:47:18.800
reservation and who is, who has actually got access to reading Shakespeare and things. And he has an
00:47:22.800
argument with the world controller about whether dulling people's senses with Soma, with the drug
00:47:29.520
that they, that they all have to make them feel good. Um, comparing that with, um, the, the, um,
00:47:37.840
depth of emotion that you can get by reading Romeo and Juliet or, or, or, or Othello. Um, and the
00:47:44.400
savage is advocating Shakespeare and the world controller is advocating the pacification of people
00:47:54.560
with, with, with, with this drug that makes them all feel good. In a way, the, the comfort that you
00:48:00.640
get from believing a falsehood, um, is like a drug. And, and it's a perfectly valid argument to say that,
00:48:09.120
that there's everything to be said for the drug. And there are, of course, real drugs you can take.
00:48:13.600
It doesn't have to be a false belief. You can, you can take Soma equivalent.
00:48:19.280
Well, well, that, I mean, that is absolutely true. It's, you know, it's that drug is,
00:48:24.560
is such a powerful thing. You, you see, I can, I've got to the stage in my life, Richard, where
00:48:29.920
I can tell a lot of the time if someone is religious, because there's a lightness about them.
00:48:34.880
There's almost the, they, that kind of existential dread, which hangs off atheists. And maybe I'm
00:48:41.280
projecting somewhat they don't seem to have.
00:48:44.080
Not my experience. Um, uh, I would say you have the same lightness, by the way.
00:48:49.680
Actually, yes.
00:48:50.240
Okay. I, I, I'm not sure that that would stand up to a serious investigation.
00:48:55.920
A controlled experiment may not show that.
00:49:00.720
Well, okay. I mean, we've got nothing but anecdotes. Um, anecdotes I've heard. I, I was close to
00:49:06.640
somebody who, um, was in, was looked, looked, looked after an old people's home. And she said that,
00:49:13.440
um, the people who are really afraid of death are the Catholics. Um, and that seems surprising in a way,
00:49:23.280
but remember there's hell as well as heaven. And, and, um, there's purgatory before you get to,
00:49:30.480
to either. Um, and so, um, I think we need a bit, a bit more investigation before, before we talk about
00:49:37.520
this lightness of being that you, you get from, from, um, uh, Richard. And, and what about you?
00:49:43.040
Because, uh, you're in your eighties now, you're in extremely great shape, both physically and mentally,
00:49:48.720
but you are in, you know, you are in your eighties. And so the moment is coming, uh, for you at some
00:49:54.640
point, I hope very, very far away. Uh, an atheist, uh, deathbed conversion is not a thing that's unheard of.
00:50:01.520
Uh, how, how do you have that lightness? It, it, it's a, but often it's a myth. I mean, there's a,
00:50:07.200
there's a, there's a myth, there's a myth that Darwin hadn't had a deathbed conversion, which is,
00:50:10.960
which is utterly false. Um, I guess what I'm asking is how do you have the lightness that you have?
00:50:17.280
How do you face death? I think that, um, well, uh, I think it was Mark Twain said, um, I was dead for
00:50:26.800
billions of years before I was born and never suffered the smallest inconvenience. Um, it's
00:50:33.440
going to be just the same as before we were born. We were, we were, we were not there during the whole
00:50:40.880
of the Cambrian and the Ordovician and the age of the dinosaurs and everything. And we're going to be
00:50:46.400
not there after we're dead. So we have this brief time in the sun to have a full and fulfilled life,
00:50:56.800
which is what I am doing and intend to go on doing until I can't anymore. Um, the, the process of dying,
00:51:04.880
as opposed to being dead, the process of dying is often very disagreeable.
00:51:11.600
That is the most British understatement that's ever made.
00:51:15.040
We're not allowed to, we're not allowed the privilege that a dog has of being taken to the vet
00:51:20.400
and put painlessly to sleep. Maybe I should identify as a dog.
00:51:26.080
If I'm not allowed to go to the vet and asked to be put down or, or if the vet refuses to put me down,
00:51:37.280
I could sue him for misspeciesing. Um, so, um, okay. I mean, if there is something frightening about
00:51:47.200
being dead, it's the idea of eternity. Yes. And eternity is a sort of frightening idea,
00:51:53.280
whether it's before you're born or after you're born, it's a kind of frightening idea.
00:51:56.960
Um, and so the best way to spend eternity, therefore, is after a general anaesthetic,
00:52:02.240
which is exactly what's going to happen. Richard, I was going, you were saying that it was
00:52:08.160
under a general anaesthetic. Was there part of you when you were at the forefront of the new atheist
00:52:15.040
movement that thought you were going to defeat religion, that you, by using facts and scientific
00:52:21.040
reason and logic, you were going to, you know, defeat all of these different religions?
00:52:27.200
I was never that optimistic. Um, no, I don't, I don't, I don't think I ever thought that. And,
00:52:32.720
and there was no, no movement. I mean, would, you know, would, uh, four, four or five books came
00:52:39.280
out of rush at the same time, but by coincidence, but there was a, never an actual movement. I think
00:52:44.640
it's a journalistic invention. Were you, uh, close with Christopher Hitchens? Well, I, I didn't know
00:52:51.520
him that well. I, I, I met him from time to time. Um, I, I had a long interview with him for new
00:53:00.160
Statesman. I think it was the last interview he ever had before he died in, in, in, in Texas,
00:53:06.800
where he was being treated. And, um, so yes, but I, I wasn't one of his close circle of friends like,
00:53:14.560
like Martin Amis and Salman Rushdie. And what were your thoughts or impressions of him?
00:53:19.840
What kind of person was he? Oh, um, immensely eloquent, immensely erudite. Um,
00:53:26.240
the most eloquent speaker I ever heard, I think. Uh, and, um, and a huge loss. I mean,
00:53:34.160
wonderful intellect, wonderful command of English language, command of facts, um, command of historic
00:53:43.680
and literary reference. Richard, we, we would talk, we would, before I asked you about the
00:53:50.320
atheism question, you sounded as if you were pro euthanasia. Are you in order to alleviate
00:53:57.200
suffering at the end of someone's life? With the, with safeguards? I think you do need safeguards
00:54:05.120
against, um, you know, let, let's get rid of granny sort of thing.
00:54:11.760
You've got to get on the housing ladder somehow, Richard. Yeah. Um, I, I think there, there have to be
00:54:21.760
safeguards and, and there can be, and, and, um, um, legal scholars look into the possibility, but,
00:54:28.080
but given, given that, uh, yes, I am. I think that, um, um, um, we, we should have the right to, um,
00:54:36.960
end our lives when we want to. Um, I know of, of cases, I've personally come across cases where
00:54:46.160
somebody has committed suicide in a not very pleasant way, um, because they still had the
00:54:53.280
power to do so. And if they had waited any longer, then they would have been incapable of, of doing the
00:55:01.520
act themselves. And therefore their lives were actually made shorter by suicide because they
00:55:10.000
didn't have the comfort of knowing that if at any time later on, when they were no longer physically
00:55:17.680
capable of killing themselves, they could ask a doctor to do it for them. You understand what I'm
00:55:22.720
saying? Yes, absolutely. Absolutely. Richard, I want to, before we wrap up and ask you a final
00:55:28.640
question and go to locals where our audience get to ask you questions, I want to ask you a couple of,
00:55:33.680
uh, just like, uh, one-off questions, which is one of the things that a lot of people are talking about
00:55:39.040
now is the emergence of AI. Uh, it's not your field of expertise. Do you have any thoughts on the
00:55:45.760
development of AI as we currently have it? I, um, had a go with chat, GPT, chat? Chat GPT. Yeah. Yeah. Um,
00:55:56.400
and I was quite interested in that. Um, it's factual knowledge is lamentable. Yes. Um, I mean,
00:56:03.520
it's actually quite comic. Uh, um, I happen to be interested in JBS Haldane, um, who was a great,
00:56:12.400
great biologist. I can't remember why, why that subject came up. And, um, his wife, his third wife,
00:56:19.440
I think, was a woman called Helen Spurway, who was a geneticist. Um, and I can't remember why I asked a
00:56:27.440
question about Helen Spurway, but I did. And it said, Helen Spurway was married to Richard Dawkins.
00:56:37.680
And, um, so I was sort of aghasted by this. And, um, later on, I, about three weeks later,
00:56:44.880
I thought, well, let's check up again. So I said, um, who was Helen Spurway managed married,
00:56:51.120
who was Helen Spurway married to? And they said, Aldous Huxley, which again was false. Um,
00:56:58.160
so I don't understand quite why it's factual knowledge of silly details like that is so poor,
00:57:05.120
because anybody can Google something now. I mean, why didn't they just Google, uh, Helen Spurway and,
00:57:12.320
and come up with the correct answer, which is that she's married to,
00:57:15.440
she was married to JBS Haldane. I mean, it's a, it would take about two seconds for a human to do
00:57:21.120
it, let alone for AI to do it. So, um, that's just a, I suppose, a vaguely amusing anecdote about
00:57:27.920
factual knowledge. Um, it is said that there, that people are using it to write essays and things like
00:57:37.200
that. It is pretty impressive. Um, um, if you, if you ask it to, um, give a, to write an essay
00:57:48.560
about something in the style of somebody or other, it produces something pretty impressive.
00:57:53.600
Well, you are, I, we could probably ask it to write an essay on, um, evolution in the style of
00:58:01.280
William Shakespeare and, and it would produce a pretty good pastiche of Shakespeare's style. Um,
00:58:09.120
I'm not quite sure. I haven't really been into, as you rightly said, I haven't really been into the
00:58:14.640
things that worry people about it, but, but I, I'm, I'm alive to the possibility that there are grave
00:58:22.320
dangers in something getting out of hand, getting out of control. And, and do you have some sense of how
00:58:28.400
that might play out? Not really. I need to do more reading on the subject. I think, I mean,
00:58:34.800
I need to read some of the people who aren't seriously worried about it and whom I respect.
00:58:38.400
Well, one of the concerns, there was an article now, I don't know how accurate it was, but it is
00:58:42.400
an example whereby they were training some kind of military system of AI to choose targets. And then
00:58:49.520
a human operator had the final decision over whether that target should be struck or not with a missile.
00:58:55.600
Uh, and allegedly, according to this article, the, because the human operator sometimes denied
00:59:03.920
a strike on valid targets for other reasons, the AI decided that the human operator was the
00:59:11.840
obstacle in the way of destroying the targets that needed destroying. And in this simulation
00:59:17.840
attacked the human operator, right? That is a sort of, so the argument about AI at least seems to be that
00:59:23.680
it's a baby now and it's getting the wives of the people wrong because it's still learning how to
00:59:28.720
walk and talk, but eventually it could grow up to be like a Hitler or whatever. You know what I mean?
00:59:32.880
Yes. I think, I think that's a very valid point. And I could imagine, um, some kind of AI
00:59:41.600
asked to determine what would be the best thing for the, for the sum of happiness would be to
00:59:45.680
exterminate everybody because we're all miserable. Yeah. Um, yes, I, I, as I said,
00:59:51.280
I need to read more about it. I think it's, it's interesting and, um, I'm in favor of, uh,
01:00:00.160
looking cautiously into the future and saying for about all sorts of things, what, what kind of
01:00:05.040
dilemmas are likely to arise in the future? What kind of problems is scientific progress? And this is one,
01:00:11.680
AIs is one example. Scientific progress likely to, to raise in future. We need to be prepared in
01:00:17.360
advance for what's going to happen. And I, I certainly have read enough to know that, um,
01:00:25.200
um, AI is capable of doing things which are really beyond our dreams at the moment. And,
01:00:31.840
and who knows where that will lead. And now, other question I wanted to ask you
01:00:36.480
in this part of the interview is this, on the balance of probability, are we alone in the universe?
01:00:41.920
And second part to that question, if we're not, is it wise to seek contact with those other sentient
01:00:48.000
life forms? I think the balance of probability is that we are not alone. Um, if we are alone in the
01:00:54.240
universe, then that has interesting implications. It means that the origin of life on our planet was
01:00:59.440
a supremely improbable event. So improbable that we're probably wasting our time trying to work
01:01:04.720
out how it happened. Um, but I don't believe that. And I, I do actually believe that we are
01:01:10.880
one of many, um, uh, life forms in the universe. And I would love to know what, what the others are like.
01:01:19.680
I mean, I'd, I'd love to know how unique we are. I'd love to know, I, I could make a few predictions.
01:01:24.400
I mean, I think I could predict that, um, there's going to, it's going to be Darwinian. Um, it's going
01:01:30.320
to be, um, it's going to have some kind of digital genetics. Um, there's got to be, there's got to be a
01:01:36.240
very accurate genetic system. We could make predictions of that, of that sort. Um,
01:01:41.040
um, is it wise to, uh, well, first of all, there's a very, there's a big distinction between life and
01:01:49.360
sentient life. Um, uh, intelligent life. Uh, the, um, intelligent life is a big step further.
01:01:59.760
And so if there's, there's probably a lot more life around than there is intelligent life around.
01:02:04.240
It's a big barrier to get through there. Um, and intelligent life, if we ever encounter it,
01:02:09.920
we'll almost certainly encounter it, not physically. We won't actually meet them
01:02:14.160
because the distances would be too great. Um, but, um, we, we're most likely to meet them through
01:02:20.960
radio waves. Um, SETI, the search for extraterrestrial intelligence where they have
01:02:26.480
actual dishes pointing out, looking for signs of intelligent life. Um,
01:02:30.880
Um, I don't think it's likely to be unwise to respond to any such messages that we find,
01:02:39.200
that we hear because the distances, again, are so great. And, um, you, we won't be able to have
01:02:45.680
a conversation. I mean, even, even the nearest star, um, is, is, is so far away that you could,
01:02:55.280
you could only send a message and then you all be dead before, before you get.
01:02:59.200
With current technological levels of development, wouldn't we be a little bit like the Navajo
01:03:03.760
sending a message in a bottle to the, to Spain in the 1450s?
01:03:07.280
Well, if you, but, but if you believe Einstein, um, then, then there is a limit to, to, to,
01:03:12.800
to how fast information can, can travel. It's limited to the speed of light.
01:03:18.160
Um, and so, um, the, the, I mean, the two science fiction books that I've read, which, um, come to
01:03:26.240
grips with that, that problem is you, you, um, Fred Hoyle's, um, Ava Andromeda and, uh, Carl Sagan's
01:03:34.800
contact where, um, both authors face up to the fact that, um, the distances are too great for direct
01:03:42.320
control of humans, for direct manipulation of humans. We don't have to fear that they actually
01:03:47.280
come in flying saucers and, and run our lives. And you can't run people's lives by radio unless
01:03:55.520
both books face this, come to the same conclusion. The, the instructions are build a computer
01:04:01.920
computer which will then control humanity. Both authors, I don't know whether they independently
01:04:09.360
thought of it or whether one of them, I forget which of, which of those books came first. Um,
01:04:13.760
so the, the extraterrestrial intelligence sends information
01:04:22.080
telling people to build a computer which will do certain things.
01:04:26.800
And the, the original senders of the information may be long dead
01:04:31.440
because it takes so long for the information to get here.
01:04:34.880
But once the information is here, then the computer can work in short-term time
01:04:41.840
and can, um, manipulate us. And, and in both cases, um, that, that, that's what,
01:04:48.480
a very interesting science fiction idea. Uh, and that's, I think, the only way we need to be afraid.
01:04:55.280
We're not going to be visited by people in flying saucers. That, I think, is too improbable.
01:04:59.360
Richard, what an absolute pleasure. Thank you so much for coming on the show. The final question
01:05:03.760
we always ask is, what's the one thing we're not talking about as a society that we really should be?
01:05:09.120
Well, I suppose we've sort of touched on it, really, the, looking into the future, we don't know
01:05:14.800
what's going to happen. And we know, looking into the past, um, that we are horrified by certain
01:05:21.120
things that we did in the past, like slavery. Um, and maybe what we should be doing is looking into
01:05:28.160
the future and imagining what will our descendants look back at our time, uh, and shudder with horror
01:05:34.960
at what we did, um, in the same ways we look back two or three centuries and shudder with horror.
01:05:42.320
Richard, thank you so much for coming on the show. It's been an absolute pleasure.
01:05:46.080
And please make sure to join us on our Locals where we'll be carrying on with this conversation.
01:05:52.400
Elon Musk is keen to escape Earth and build on Mars. As an evolutionary biologist, how do you feel about
01:05:58.160
that? And what do you think establishing a successful second planet means for the human race in our
01:06:02.960
evolution?
Link copied!