Juno News - August 26, 2023


On C2C: Will AI be the end of critical thinking?


Episode Stats

Length

21 minutes

Words per Minute

178.90642

Word Count

3,802

Sentence Count

153

Misogynist Sentences

7


Summary


Transcript

00:00:00.000 I have not been as fearful of artificial intelligence as some people have, because I don't think
00:00:13.620 that human intelligence has often served as well, but that's a bit of a glib joke to start
00:00:18.080 off what is a serious discussion, which is what AI is doing to discourse and to thought.
00:00:24.260 Now, we haven't talked a lot about AI on this show.
00:00:27.280 I've kind of been waiting for the right angle and the right opportunity, and I should say
00:00:31.560 I've been one of these people that has sort of enjoyed the novelty of it.
00:00:34.900 When ChatGPT has come up and you get the ability to just have a quick conversation with this
00:00:40.760 thing and have it give you some response to a question, and there's a program that I've
00:00:46.320 had some fun with called MidJourney, which will create AI-generated images, and you can
00:00:52.720 give it a whole bunch of prompts.
00:00:54.140 I've had a lot of fun with this one.
00:00:55.340 The one that I did, I won't show you because I wasn't thrilled with it, but I asked for
00:00:58.620 like childhood photos of Fidel Castro pushing young Justin Trudeau on a swing, but the AI
00:01:04.560 was getting Fidel Castro and Justin Trudeau's faces mixed up, which maybe makes it smarter
00:01:08.940 than humans.
00:01:09.520 Who knows?
00:01:10.380 And then I also had some fun this morning, and I asked to get like some photographs of
00:01:14.760 Krista Freeland driving, so maybe we can throw those up.
00:01:19.060 Yeah, there we go.
00:01:20.180 These are the samples it gave me.
00:01:21.500 I thought that speed demon, Krista Freeland, fresh off the heels of getting her ticket
00:01:25.580 for going however many kilometers over.
00:01:27.760 That's her basically road racing down some Alberta highway.
00:01:31.620 I like the one on the bottom right myself, although it looks a little terrifying.
00:01:36.000 That one, it looks to be in Ottawa, though.
00:01:37.780 You can see in the back right there, it looks to be a center block that she's just like leaving
00:01:42.640 in the dust there.
00:01:43.980 The one on the top right is good.
00:01:45.120 It's a little aspirational.
00:01:46.080 She's really flying there so much that she needs the space helmet.
00:01:49.860 She's putting on so many miles and going so fast, she has ascended off the ground.
00:01:54.540 So take from that what you will.
00:01:56.420 But for all the fun that AI offers, and yes, there is some, it also has very serious implications.
00:02:03.620 And those implications we have not really fully explored because despite the fact that this
00:02:08.320 technology has been in development for many, many years, it really seems it's only been
00:02:12.940 in the last year that people have started to grapple with the real world implications
00:02:17.960 of it.
00:02:18.500 And, you know, we see this in academia where universities which have had to focus on detecting
00:02:23.940 plagiarism now have this new problem, which is did students just create something original
00:02:29.380 by entering a few prompts for their essay assignments into chat GPT or whatnot.
00:02:36.180 There's a great piece in C2C Journal by Christopher Snook about this called AI,
00:02:41.280 The Destruction of Thought and the End of the Humanities.
00:02:45.000 He is a lecturer with Dalhousie University and a contributor to C2C Journal.
00:02:49.800 And he joins us now, not an AI-generated version, but the man himself.
00:02:54.480 Good to talk to you, Christopher.
00:02:55.480 Thanks for coming on.
00:02:57.040 Thank you, Andrew.
00:02:57.900 Thank you so much for the invitation.
00:02:59.740 So let's start first off with where your issue is with this.
00:03:04.060 Why are you concerned about AI in the context here?
00:03:07.600 Yeah, I can, I suppose I can answer that in a fairly simple way.
00:03:11.280 As you've already indicated, there's a great deal of joy maybe to be had with playing with
00:03:17.000 sort of AI applications.
00:03:18.800 But at the simplest level, I suppose maybe I could say two things.
00:03:22.940 One would be that AI introduces, I'm a humanities teacher, so AI-generated content introduces
00:03:29.520 into the university and into students' lives very easy possibilities of escaping from a certain
00:03:37.240 kind of reflection that may be essential to their development within the context of the
00:03:41.300 humanities historically.
00:03:43.040 But secondly, I think I have a pretty significant concern that AI is actually indicative in many
00:03:48.320 respects of a much longer trend in humanities education in Canada that has fairly uncritically
00:03:55.820 assimilated new technological developments without reflecting on their consequences for pedagogy
00:04:01.720 and education.
00:04:04.440 That's quite an interesting approach to this.
00:04:06.760 And, you know, one thing that I always recall, even from my own time in university, is that
00:04:11.280 essays were very challenging.
00:04:14.060 I would do better at them now, but they were very challenging because you can't really cheat
00:04:17.900 your way through an essay unless you're actually cheating and plagiarizing and whatnot,
00:04:22.720 because it's not just about knowing the facts.
00:04:25.400 You can't Google the answer to the question when you basically have to show your work
00:04:29.280 and show how you arrived at something.
00:04:31.340 And certainly in an academic context, AI has huge implications for that, because all of
00:04:37.040 a sudden someone else could do the thinking with you.
00:04:39.660 I could just give this machine a bunch of different data points and say, formulate an argument
00:04:45.320 for me.
00:04:45.920 And that's something, I mean, I've talked to professors who have already been complaining
00:04:51.100 about the decline in critical thinking in universities.
00:04:53.440 And now we've added this other tool, which maybe can be used for good, but also can further
00:04:58.920 erode people having to come up with these skills on their own.
00:05:03.900 Yeah.
00:05:04.200 What I've tried to do in the article, maybe if I kind of talk about some of the points
00:05:08.000 in the article, that may be helpful for at least giving people a sense of where my concern
00:05:13.020 lies.
00:05:13.480 So my concern really grew out of two things that I saw in the university last year.
00:05:17.640 So the first was, I mean, a remarkable amount of energy and anxiety around the appearance
00:05:22.200 of things like chat GPT, right?
00:05:25.020 Sort of large language models that can produce texts fairly competently, increasingly competently
00:05:32.080 for students with very, very little to no work on their side.
00:05:36.020 So there's a huge amount of anxiety, as you pointed to, Andrew, earlier in your introduction
00:05:39.920 to this conversation.
00:05:42.840 It's a different, it's not even plagiarism in any recognizable sense.
00:05:46.120 So it's just allowing AI to generate texts from the information it's kind of gathered
00:05:51.980 through its internet, through its chat box on the internet.
00:05:54.840 So there was a huge conversation about this in the university.
00:05:57.880 And what I noted was that primarily that conversation was focused on questions of use.
00:06:02.200 And so I spend some time teaching engineers, though I teach humanities to engineers, and they
00:06:05.860 very patiently kind of tolerate this course.
00:06:09.180 I was going to say, that seems like a very, that seems like a very difficult challenge for
00:06:12.180 you.
00:06:12.380 It's a hard sell, it's a hard sell, but they're very patient.
00:06:14.840 And they tolerate this sort of required course on, effectively on the history of technologies.
00:06:20.080 And one of the key things that I've been thinking about since teaching this course is
00:06:24.440 what Neil Postman simply observes, which is that the introduction of every new technology
00:06:28.980 doesn't simply give a new tool to humans.
00:06:31.620 But he sort of coined the idea, or helped kind of articulate the idea that every technology
00:06:37.220 shifts the world ecologically in much the same way that an ecosystem is changed if a new
00:06:43.340 species is introduced, right?
00:06:44.900 So it's not just that we have all of a sudden AI, but rather that the whole world shifts
00:06:49.280 around the availability of these new technologies.
00:06:52.180 And embedded in these technologies are certain assumptions about what it is to be human.
00:06:56.940 And so it's a bit of a rambling response to your previous comment.
00:07:01.800 But what I noticed in the university over the last six or seven months is that the conversation
00:07:07.040 has been almost exclusively focused on use.
00:07:10.240 There have been some people who are sort of diametrically opposed to the appearance of
00:07:15.880 AI in any form in the university.
00:07:17.780 I kind of tend in that direction, certainly for the humanities.
00:07:21.100 Others who are much more supportive of the use of AI in various ways to facilitate writing.
00:07:26.980 But regardless of where one stood on, it stands on the use of AI, I've noticed that very few
00:07:32.360 people are asking deeper questions, such as what kind of world does AI produce and what
00:07:37.420 kind of worldview or what sort of assumptions are built into the technology?
00:07:42.180 And it's there that I think universities need to really be careful about the implementation
00:07:46.000 of AI, partly because I think AI actually reveals, it's a bit apocalyptic, that is, it kind of
00:07:52.900 reveals something about the nature of higher education in Canada that's been developing
00:07:56.980 for years.
00:07:57.740 And we could talk about that sort of narrowing a viewpoint diversity to different aspects
00:08:01.860 of the university life.
00:08:03.560 But also, I think the other thing that I think has been missed is that AI, AI generated texts
00:08:11.000 pushes against certain proclaimed positions, moral positions that the university has adopted
00:08:17.720 in the last, maybe in the last decade.
00:08:19.620 One aspect that this springs to mind, and you address it in the piece, in one section anyway,
00:08:25.640 notably, is the idea of bias.
00:08:27.660 And, you know, facts, in theory, are neutral, and they do not have a political persuasion.
00:08:33.040 It's the assembly of facts and the composition of various facts that you can use to sort of
00:08:38.140 demonstrate something that is a bit more biased.
00:08:40.740 And one thing we've seen in AI is how it's providing, it's doing the thinking for you in theory.
00:08:47.600 But the problem with that, among others, is that it is producing a biased outcome.
00:08:53.660 It's producing a biased response.
00:08:55.520 I mean, I had once, when I was first playing around with ChatGPT, a debate with a machine.
00:09:00.060 So the joke was on me about what a woman is.
00:09:02.540 And it was interesting seeing this machine twist itself into all of these many logical
00:09:06.620 knots about trying to answer this question.
00:09:09.400 But it was actually quite terrifying how it started giving me the talking points I would
00:09:13.980 expect if I were having this with some university diversity administrator.
00:09:17.980 And it started telling me about inclusivity and tolerance and women can come in any forms.
00:09:22.720 And there is something there in which AI is basically telling people that there is one
00:09:29.360 way to construct a thought when it does this, that you aren't actually able to assemble facts
00:09:35.320 into different worldviews.
00:09:37.560 Yeah, I mean, I think that that's true.
00:09:39.900 And certainly the studies have varied and in some cases disagree a little bit with one
00:09:44.140 another about where the biases are found in AI technologies.
00:09:48.180 Though some people have, there's been some studies that have tried to argue that there's,
00:09:51.980 because some of the early scraping of the internet focused primarily on Reddit sites,
00:09:57.760 that there was a kind of conservative or male bias somehow in the technology.
00:10:03.700 But it seems to me fairly clear now that the technology seems to be biased fairly clearly,
00:10:08.480 I think, in the other direction in terms of the kinds of sources that it's recycling when
00:10:15.560 it's when it's producing kind of mashup texts.
00:10:18.120 So I began the article and this is one of the things, maybe a way of thinking about AI in a
00:10:22.200 broader context or sort of moving back from the technology to think about what it has to
00:10:26.180 say about universities.
00:10:27.740 I began the article in C2C really just by reflecting on the fact that
00:10:31.760 a kind of formulaic response to the work of pedagogy has become characteristic of universities
00:10:39.660 generally in Canada.
00:10:41.200 And it's in part indicated through the demand that applicants for university positions
00:10:46.220 complete diversity, equity, and inclusion statements, diversity statements as part
00:10:51.040 of their application packages.
00:10:52.280 And kind of famously, or if you follow these sort of stories infamously, depends on what
00:10:58.500 one thinks about all these things, a professor in the United States asked ChatGPT to produce
00:11:03.840 a diversity document just last year.
00:11:05.860 And he was astounded to see, just as you've described, the speed with which ChatGPT was
00:11:12.060 able to reproduce all of the talking points and all of the assumptions of a fairly kind of
00:11:17.720 middle-of-the-road Canadian higher education position on issues of diversity, inclusion,
00:11:24.340 and equity.
00:11:25.880 To my mind, what that revealed was that we're sort of beginning to operate in the university
00:11:30.140 in a world that is fairly, at the very least, formulaic in its expectations of whatever diversity,
00:11:38.320 inclusion, and equity may be about, or diversity, equity, and inclusion.
00:11:41.920 So it was from there, really, in the article that I wanted to try to see or to explore how
00:11:48.440 is it that the technologies that are available to us now in the university and that have been
00:11:52.740 slowly growing in their implementation in the university over the last 15 or 20 years
00:11:59.760 may actually be both accelerated by the advent of AI, but may also be, in a certain sense,
00:12:08.340 pointing towards AI, that is to say, pointing towards a world in which a kind of formulaic
00:12:13.980 regurgitation of information becomes a kind of normative expectation of students, even
00:12:19.040 in the context of their degrees.
00:12:20.480 So that's sort of where the article began.
00:12:24.120 So just to sort of maybe make a connection with what you found when you poked and prodded
00:12:27.800 ChatGPT, that it tends to kind of produce fairly predictable results relative to certain
00:12:34.100 questions.
00:12:34.720 Well, there is also to this, I mean, the most, I guess, cogent defense I hear of AI is that
00:12:42.860 AI is little more than a mirror to the existing world.
00:12:46.600 I mean, AI is not really formulating its own materials that it's not drawing from the trove
00:12:53.340 of inputs.
00:12:54.400 Now, obviously, individual inputs can be manipulated, and we also have terms of use that govern it.
00:13:00.480 I'm trying to bring us away from the use discussion of this that you were talking about earlier,
00:13:04.560 but I guess in that sense, is this just reflecting an existing problem, or is this making it worse?
00:13:12.980 Right, yeah, that's a very good question.
00:13:14.520 I would say from, there might be two things to say about that.
00:13:17.220 On the one hand, from my perspective, and this was maybe my concern with the conversation so far
00:13:22.900 in higher education in Canada about AI, its preoccupation with questions of use has really prevented
00:13:28.620 people from asking a much slower and more difficult question, which is to say, is this actually a
00:13:34.500 benefit, or is it simply a reflection of the world we're in, or is it making things worse?
00:13:40.640 So I think that that deeper question about the kind of ecosystem consequences or cultural
00:13:46.140 consequences of AI is not really being asked.
00:13:50.780 So that is to say, AI at some level is a kind of metaphor, in much the same way one might think
00:13:57.300 about COVID as a kind of, we can think about it as an illness, but we can also think about
00:14:02.840 COVID response, at least, as a bit of a metaphor of our contemporary cultural moment.
00:14:08.060 So there is that mirroring back.
00:14:10.160 But I would say from the perspective of pedagogy, AI raises some very deep questions that, to
00:14:15.700 my mind, intensify problems that were already present.
00:14:18.620 So it's not so much that it simply introduces a newness that's radical, but intensifies certain
00:14:24.920 very particular problems.
00:14:26.800 So one of those problems, I think, is connected to the use of devices generally for humanities
00:14:34.760 education in particular.
00:14:36.680 So one of the things I think many of us have experienced is the extent to which screens and
00:14:41.880 screen reading and iPhones or cell phones, the extent to which they actually produce in us
00:14:46.500 kind of habits of scanning, a kind of hyper-attention, what one scholar calls forms of hyper-attention,
00:14:51.140 not focused attention or contemplative reflection, but a kind of hyper-attention that actually
00:14:57.100 tends to kind of lead us towards a certain kind of rashness in our decision-making.
00:15:02.260 So that's one deep and profound concern I have, especially when institutions seem to be dominated
00:15:07.380 by certain sets of political commitments that ought themselves to be subject to serious
00:15:12.780 reflection and consideration, right?
00:15:14.460 So if we're in an environment where there's certain assumptions about what political positions
00:15:18.580 are normative, it needs to be the case that those can be thought about deeply and reflectively.
00:15:23.920 And if we're using technologies that limit that capacity, then we're in a little bit of
00:15:27.800 danger.
00:15:28.660 But the other one, I think for me, from the perspective as a teacher, is that, and this
00:15:33.260 would be, I mean, I'm affiliated with the classics department and I spend a lot of time in the last
00:15:38.000 years teaching Augustine, a sort of famous foundational voice for the Western world.
00:15:43.340 And Augustine is one of many thinkers who highlights the fundamental role of memory in the
00:15:49.140 constitution of our personalities, the sort of crucial role that memory plays.
00:15:53.040 And it's kind of essential to technologies like Chachi BT that we offload or offshore the
00:15:59.580 faculty of memory to the technological device, right?
00:16:03.460 It does the work for us, right?
00:16:05.140 So I don't struggle with Augustine or Dante or Homer or any of those things.
00:16:09.680 I let Chachi BT do the struggle in a certain sense.
00:16:12.660 I mean, it's not really struggling, but I let it do the amalgamation of opinion making,
00:16:16.440 information, and I'm left passive in that response.
00:16:20.100 So in that sense, I think the technologies actually inhibit the kind of interior dialogue
00:16:25.320 that's fundamental to education, but that's also fundamental to being a free person in the
00:16:30.300 world, right?
00:16:30.800 Hannah Arendt, I think, points this up very, very powerfully in her reflections on totalitarianism.
00:16:35.740 If we can't have a dialogue with ourselves, if we're pulled out of ourselves endlessly
00:16:40.200 and offshore, even our memory, we lose the ability to actually be free agents in the world.
00:16:47.180 So these are some of the things that I'm very concerned about at the level of pedagogy,
00:16:50.740 which is why I tend to a pretty puritanical, I suppose, relationship to AI when it comes to
00:16:56.440 at least to humanity's classrooms.
00:16:58.040 I recognize AI has different applications in different contexts.
00:17:02.260 It's funny, at the risk of oversimplifying it, I think of, you know, a movie that I've
00:17:06.900 watched that, you know, say is two hours long.
00:17:09.240 I could find out what happens in that movie in about 60 seconds by just reading a plot
00:17:14.340 synopsis on Wikipedia.
00:17:15.760 But I don't do that.
00:17:16.720 I watch the movie because there is something in that process.
00:17:19.800 You feel, you see, you learn, you get insights.
00:17:22.800 It's the same as why, you know, despite the fact that I may not have taken this advice when
00:17:26.120 I was in high school, reading the Coles notes of something is not the same as reading the
00:17:31.420 thing itself.
00:17:32.120 I mean, I could get chat GPT to say, you know, give me some bullet points that I can bring
00:17:36.240 up in tutorial about, you know, the cave or something.
00:17:38.980 But that doesn't mean I've done that.
00:17:40.460 So you're quite right.
00:17:42.100 And I also wonder, I mean, to appeal to your department, the classics, if you were to input
00:17:46.660 into chat GPT, the most beautiful works of literature of classics that you'd ever seen
00:17:52.940 and said, create something like this, could it do that in your view?
00:17:57.240 Could it create the beauty that we have seen from all of these people thousands of years
00:18:02.440 ago?
00:18:03.600 Yeah, that's a very interesting question.
00:18:04.780 That's a very hotly debated topic, as you may know.
00:18:06.980 Of course, in the world of visual arts, someone recently was awarded a prize, right,
00:18:11.220 in the visual arts for an artificially created, produced image, all kinds of, of course, very
00:18:19.400 deep ethical questions around AI and its accumulation of information and how that happens.
00:18:25.520 But, you know, from my perspective, for me, the answer to that question was really given
00:18:30.000 quite beautifully by Nick Cave recently.
00:18:31.760 Nick Cave, the Australian singer-songwriter, was asked this question.
00:18:36.980 A fan sent him a poem that ChatGPT had written when he asked ChatGPT, write me a poem or a
00:18:44.180 song in the style of Nick Cave.
00:18:46.480 And Nick Cave's response was to say that even if it were a good song, which Nick Cave refused
00:18:52.600 to concede that it was a good imitation, he said that the problem is that ChatGPT, artificial
00:18:59.060 intelligence, has been nowhere and suffered nothing.
00:19:02.020 And to be human in the world at all, as someone like Jordan Peterson is constantly reminding
00:19:07.720 us, is to suffer and out of that suffering either to sort of produce meaning in the world
00:19:12.340 and in our lives.
00:19:14.180 And I've been fairly persuaded by Nick Cave that no matter how close the approximation one
00:19:19.780 might be able to artificially reproduce, the fact that the technology has itself been nowhere
00:19:25.440 and suffered nothing means that that material can have very little consequence for me as
00:19:30.680 someone who lives in the world with all of its fragility.
00:19:33.300 So, yeah, so I guess that would be my answer, which isn't, I mean, yeah, maybe not the best
00:19:39.180 answer.
00:19:39.720 No, it is interesting.
00:19:41.040 And now I'm like geeking out on this topic myself.
00:19:43.860 So I think we'll have to have you back on in another show.
00:19:46.160 But, you know, I remember when I did tutorials in various classes in university, the one thing
00:19:52.060 that was always so critical when you were understanding a work was to understand the author and the
00:19:57.300 context in which they wrote a particular work.
00:19:59.940 And even if the author is some professor who's still alive, understanding how that professor
00:20:04.560 came about, you know, you read, for example, a, you know, a dissertation and you say, oh, well,
00:20:09.060 this was an environmental historian.
00:20:10.680 Why were they writing about this issue?
00:20:11.960 And with ChatGPT, that context is eroded because there is no human context or it's an amalgamation
00:20:19.460 of, you know, 150 human contexts that you don't actually know about and can't see.
00:20:23.680 So I think that's a less elegant way of describing what you've shared from Nick Cave there.
00:20:28.660 And I thank you for it.
00:20:30.120 The piece in C2C Journal is AI, the Destruction of Thought and the End of the Humanities by Christopher
00:20:36.340 Snook.
00:20:36.700 And they also have another part of this series written by Gleb Lysik, who we had on the show
00:20:41.080 a couple of weeks ago about something else entirely.
00:20:44.200 Christopher, thanks so much for coming on.
00:20:45.960 Good to talk to a real human in this day and age.
00:20:48.840 Thank you, Andrew, very much.
00:20:49.880 I appreciate it.
00:20:50.600 Thanks for listening to The Andrew Lawton Show.
00:20:53.180 Support the program by donating to True North at www.tnc.news.
00:20:58.640 Thank you, Andrew.