Juno News - August 26, 2023


On C2C: Will AI be the end of critical thinking?


Episode Stats


Length

21 minutes

Words per minute

178.90642

Word count

3,802

Sentence count

153

Harmful content

Misogyny

7

sentences flagged


Summary

Summaries generated with gmurro/bart-large-finetuned-filtered-spotify-podcast-summ .

In this episode, we're joined by cognitive scientist and lecturer at Dalhousie University and contributor to C2C Journal, Christopher Snook, to talk about the dangers of artificial intelligence (AI) in the real world.

Transcript

Transcript generated with Whisper (turbo).
Misogyny classifications generated with MilaNLProc/bert-base-uncased-ear-misogyny .
00:00:00.000 I have not been as fearful of artificial intelligence as some people have, because I don't think
00:00:13.620 that human intelligence has often served as well, but that's a bit of a glib joke to start
00:00:18.080 off what is a serious discussion, which is what AI is doing to discourse and to thought.
00:00:24.260 Now, we haven't talked a lot about AI on this show.
00:00:27.280 I've kind of been waiting for the right angle and the right opportunity, and I should say
00:00:31.560 I've been one of these people that has sort of enjoyed the novelty of it.
00:00:34.900 When ChatGPT has come up and you get the ability to just have a quick conversation with this
00:00:40.760 thing and have it give you some response to a question, and there's a program that I've
00:00:46.320 had some fun with called MidJourney, which will create AI-generated images, and you can
00:00:52.720 give it a whole bunch of prompts.
00:00:54.140 I've had a lot of fun with this one.
00:00:55.340 The one that I did, I won't show you because I wasn't thrilled with it, but I asked for
00:00:58.620 like childhood photos of Fidel Castro pushing young Justin Trudeau on a swing, but the AI
00:01:04.560 was getting Fidel Castro and Justin Trudeau's faces mixed up, which maybe makes it smarter
00:01:08.940 than humans.
00:01:09.520 Who knows?
00:01:10.380 And then I also had some fun this morning, and I asked to get like some photographs of
00:01:14.760 Krista Freeland driving, so maybe we can throw those up. 0.99
00:01:19.060 Yeah, there we go.
00:01:20.180 These are the samples it gave me.
00:01:21.500 I thought that speed demon, Krista Freeland, fresh off the heels of getting her ticket 1.00
00:01:25.580 for going however many kilometers over.
00:01:27.760 That's her basically road racing down some Alberta highway. 1.00
00:01:31.620 I like the one on the bottom right myself, although it looks a little terrifying.
00:01:36.000 That one, it looks to be in Ottawa, though.
00:01:37.780 You can see in the back right there, it looks to be a center block that she's just like leaving 0.98
00:01:42.640 in the dust there.
00:01:43.980 The one on the top right is good.
00:01:45.120 It's a little aspirational.
00:01:46.080 She's really flying there so much that she needs the space helmet. 0.50
00:01:49.860 She's putting on so many miles and going so fast, she has ascended off the ground.
00:01:54.540 So take from that what you will.
00:01:56.420 But for all the fun that AI offers, and yes, there is some, it also has very serious implications.
00:02:03.620 And those implications we have not really fully explored because despite the fact that this
00:02:08.320 technology has been in development for many, many years, it really seems it's only been
00:02:12.940 in the last year that people have started to grapple with the real world implications
00:02:17.960 of it.
00:02:18.500 And, you know, we see this in academia where universities which have had to focus on detecting
00:02:23.940 plagiarism now have this new problem, which is did students just create something original
00:02:29.380 by entering a few prompts for their essay assignments into chat GPT or whatnot.
00:02:36.180 There's a great piece in C2C Journal by Christopher Snook about this called AI,
00:02:41.280 The Destruction of Thought and the End of the Humanities.
00:02:45.000 He is a lecturer with Dalhousie University and a contributor to C2C Journal.
00:02:49.800 And he joins us now, not an AI-generated version, but the man himself.
00:02:54.480 Good to talk to you, Christopher.
00:02:55.480 Thanks for coming on.
00:02:57.040 Thank you, Andrew.
00:02:57.900 Thank you so much for the invitation.
00:02:59.740 So let's start first off with where your issue is with this.
00:03:04.060 Why are you concerned about AI in the context here?
00:03:07.600 Yeah, I can, I suppose I can answer that in a fairly simple way.
00:03:11.280 As you've already indicated, there's a great deal of joy maybe to be had with playing with
00:03:17.000 sort of AI applications.
00:03:18.800 But at the simplest level, I suppose maybe I could say two things.
00:03:22.940 One would be that AI introduces, I'm a humanities teacher, so AI-generated content introduces
00:03:29.520 into the university and into students' lives very easy possibilities of escaping from a certain
00:03:37.240 kind of reflection that may be essential to their development within the context of the
00:03:41.300 humanities historically.
00:03:43.040 But secondly, I think I have a pretty significant concern that AI is actually indicative in many
00:03:48.320 respects of a much longer trend in humanities education in Canada that has fairly uncritically
00:03:55.820 assimilated new technological developments without reflecting on their consequences for pedagogy
00:04:01.720 and education.
00:04:04.440 That's quite an interesting approach to this.
00:04:06.760 And, you know, one thing that I always recall, even from my own time in university, is that
00:04:11.280 essays were very challenging.
00:04:14.060 I would do better at them now, but they were very challenging because you can't really cheat
00:04:17.900 your way through an essay unless you're actually cheating and plagiarizing and whatnot,
00:04:22.720 because it's not just about knowing the facts.
00:04:25.400 You can't Google the answer to the question when you basically have to show your work
00:04:29.280 and show how you arrived at something.
00:04:31.340 And certainly in an academic context, AI has huge implications for that, because all of
00:04:37.040 a sudden someone else could do the thinking with you.
00:04:39.660 I could just give this machine a bunch of different data points and say, formulate an argument
00:04:45.320 for me.
00:04:45.920 And that's something, I mean, I've talked to professors who have already been complaining
00:04:51.100 about the decline in critical thinking in universities.
00:04:53.440 And now we've added this other tool, which maybe can be used for good, but also can further
00:04:58.920 erode people having to come up with these skills on their own.
00:05:03.900 Yeah.
00:05:04.200 What I've tried to do in the article, maybe if I kind of talk about some of the points
00:05:08.000 in the article, that may be helpful for at least giving people a sense of where my concern
00:05:13.020 lies.
00:05:13.480 So my concern really grew out of two things that I saw in the university last year.
00:05:17.640 So the first was, I mean, a remarkable amount of energy and anxiety around the appearance
00:05:22.200 of things like chat GPT, right?
00:05:25.020 Sort of large language models that can produce texts fairly competently, increasingly competently
00:05:32.080 for students with very, very little to no work on their side.
00:05:36.020 So there's a huge amount of anxiety, as you pointed to, Andrew, earlier in your introduction
00:05:39.920 to this conversation.
00:05:42.840 It's a different, it's not even plagiarism in any recognizable sense.
00:05:46.120 So it's just allowing AI to generate texts from the information it's kind of gathered
00:05:51.980 through its internet, through its chat box on the internet.
00:05:54.840 So there was a huge conversation about this in the university.
00:05:57.880 And what I noted was that primarily that conversation was focused on questions of use.
00:06:02.200 And so I spend some time teaching engineers, though I teach humanities to engineers, and they
00:06:05.860 very patiently kind of tolerate this course.
00:06:09.180 I was going to say, that seems like a very, that seems like a very difficult challenge for
00:06:12.180 you.
00:06:12.380 It's a hard sell, it's a hard sell, but they're very patient.
00:06:14.840 And they tolerate this sort of required course on, effectively on the history of technologies.
00:06:20.080 And one of the key things that I've been thinking about since teaching this course is
00:06:24.440 what Neil Postman simply observes, which is that the introduction of every new technology
00:06:28.980 doesn't simply give a new tool to humans.
00:06:31.620 But he sort of coined the idea, or helped kind of articulate the idea that every technology
00:06:37.220 shifts the world ecologically in much the same way that an ecosystem is changed if a new
00:06:43.340 species is introduced, right?
00:06:44.900 So it's not just that we have all of a sudden AI, but rather that the whole world shifts
00:06:49.280 around the availability of these new technologies.
00:06:52.180 And embedded in these technologies are certain assumptions about what it is to be human.
00:06:56.940 And so it's a bit of a rambling response to your previous comment.
00:07:01.800 But what I noticed in the university over the last six or seven months is that the conversation
00:07:07.040 has been almost exclusively focused on use.
00:07:10.240 There have been some people who are sort of diametrically opposed to the appearance of
00:07:15.880 AI in any form in the university.
00:07:17.780 I kind of tend in that direction, certainly for the humanities.
00:07:21.100 Others who are much more supportive of the use of AI in various ways to facilitate writing.
00:07:26.980 But regardless of where one stood on, it stands on the use of AI, I've noticed that very few
00:07:32.360 people are asking deeper questions, such as what kind of world does AI produce and what
00:07:37.420 kind of worldview or what sort of assumptions are built into the technology?
00:07:42.180 And it's there that I think universities need to really be careful about the implementation
00:07:46.000 of AI, partly because I think AI actually reveals, it's a bit apocalyptic, that is, it kind of
00:07:52.900 reveals something about the nature of higher education in Canada that's been developing
00:07:56.980 for years.
00:07:57.740 And we could talk about that sort of narrowing a viewpoint diversity to different aspects
00:08:01.860 of the university life.
00:08:03.560 But also, I think the other thing that I think has been missed is that AI, AI generated texts
00:08:11.000 pushes against certain proclaimed positions, moral positions that the university has adopted
00:08:17.720 in the last, maybe in the last decade.
00:08:19.620 One aspect that this springs to mind, and you address it in the piece, in one section anyway,
00:08:25.640 notably, is the idea of bias.
00:08:27.660 And, you know, facts, in theory, are neutral, and they do not have a political persuasion.
00:08:33.040 It's the assembly of facts and the composition of various facts that you can use to sort of
00:08:38.140 demonstrate something that is a bit more biased.
00:08:40.740 And one thing we've seen in AI is how it's providing, it's doing the thinking for you in theory.
00:08:47.600 But the problem with that, among others, is that it is producing a biased outcome.
00:08:53.660 It's producing a biased response.
00:08:55.520 I mean, I had once, when I was first playing around with ChatGPT, a debate with a machine.
00:09:00.060 So the joke was on me about what a woman is. 1.00
00:09:02.540 And it was interesting seeing this machine twist itself into all of these many logical
00:09:06.620 knots about trying to answer this question.
00:09:09.400 But it was actually quite terrifying how it started giving me the talking points I would
00:09:13.980 expect if I were having this with some university diversity administrator.
00:09:17.980 And it started telling me about inclusivity and tolerance and women can come in any forms. 1.00
00:09:22.720 And there is something there in which AI is basically telling people that there is one
00:09:29.360 way to construct a thought when it does this, that you aren't actually able to assemble facts
00:09:35.320 into different worldviews.
00:09:37.560 Yeah, I mean, I think that that's true.
00:09:39.900 And certainly the studies have varied and in some cases disagree a little bit with one
00:09:44.140 another about where the biases are found in AI technologies.
00:09:48.180 Though some people have, there's been some studies that have tried to argue that there's,
00:09:51.980 because some of the early scraping of the internet focused primarily on Reddit sites,
00:09:57.760 that there was a kind of conservative or male bias somehow in the technology.
00:10:03.700 But it seems to me fairly clear now that the technology seems to be biased fairly clearly,
00:10:08.480 I think, in the other direction in terms of the kinds of sources that it's recycling when
00:10:15.560 it's when it's producing kind of mashup texts.
00:10:18.120 So I began the article and this is one of the things, maybe a way of thinking about AI in a
00:10:22.200 broader context or sort of moving back from the technology to think about what it has to
00:10:26.180 say about universities.
00:10:27.740 I began the article in C2C really just by reflecting on the fact that
00:10:31.760 a kind of formulaic response to the work of pedagogy has become characteristic of universities
00:10:39.660 generally in Canada.
00:10:41.200 And it's in part indicated through the demand that applicants for university positions
00:10:46.220 complete diversity, equity, and inclusion statements, diversity statements as part
00:10:51.040 of their application packages.
00:10:52.280 And kind of famously, or if you follow these sort of stories infamously, depends on what
00:10:58.500 one thinks about all these things, a professor in the United States asked ChatGPT to produce
00:11:03.840 a diversity document just last year.
00:11:05.860 And he was astounded to see, just as you've described, the speed with which ChatGPT was
00:11:12.060 able to reproduce all of the talking points and all of the assumptions of a fairly kind of
00:11:17.720 middle-of-the-road Canadian higher education position on issues of diversity, inclusion,
00:11:24.340 and equity.
00:11:25.880 To my mind, what that revealed was that we're sort of beginning to operate in the university
00:11:30.140 in a world that is fairly, at the very least, formulaic in its expectations of whatever diversity,
00:11:38.320 inclusion, and equity may be about, or diversity, equity, and inclusion.
00:11:41.920 So it was from there, really, in the article that I wanted to try to see or to explore how
00:11:48.440 is it that the technologies that are available to us now in the university and that have been
00:11:52.740 slowly growing in their implementation in the university over the last 15 or 20 years
00:11:59.760 may actually be both accelerated by the advent of AI, but may also be, in a certain sense,
00:12:08.340 pointing towards AI, that is to say, pointing towards a world in which a kind of formulaic
00:12:13.980 regurgitation of information becomes a kind of normative expectation of students, even
00:12:19.040 in the context of their degrees.
00:12:20.480 So that's sort of where the article began.
00:12:24.120 So just to sort of maybe make a connection with what you found when you poked and prodded
00:12:27.800 ChatGPT, that it tends to kind of produce fairly predictable results relative to certain
00:12:34.100 questions.
00:12:34.720 Well, there is also to this, I mean, the most, I guess, cogent defense I hear of AI is that
00:12:42.860 AI is little more than a mirror to the existing world.
00:12:46.600 I mean, AI is not really formulating its own materials that it's not drawing from the trove
00:12:53.340 of inputs.
00:12:54.400 Now, obviously, individual inputs can be manipulated, and we also have terms of use that govern it.
00:13:00.480 I'm trying to bring us away from the use discussion of this that you were talking about earlier,
00:13:04.560 but I guess in that sense, is this just reflecting an existing problem, or is this making it worse?
00:13:12.980 Right, yeah, that's a very good question.
00:13:14.520 I would say from, there might be two things to say about that.
00:13:17.220 On the one hand, from my perspective, and this was maybe my concern with the conversation so far
00:13:22.900 in higher education in Canada about AI, its preoccupation with questions of use has really prevented
00:13:28.620 people from asking a much slower and more difficult question, which is to say, is this actually a
00:13:34.500 benefit, or is it simply a reflection of the world we're in, or is it making things worse?
00:13:40.640 So I think that that deeper question about the kind of ecosystem consequences or cultural
00:13:46.140 consequences of AI is not really being asked.
00:13:50.780 So that is to say, AI at some level is a kind of metaphor, in much the same way one might think
00:13:57.300 about COVID as a kind of, we can think about it as an illness, but we can also think about
00:14:02.840 COVID response, at least, as a bit of a metaphor of our contemporary cultural moment.
00:14:08.060 So there is that mirroring back.
00:14:10.160 But I would say from the perspective of pedagogy, AI raises some very deep questions that, to
00:14:15.700 my mind, intensify problems that were already present.
00:14:18.620 So it's not so much that it simply introduces a newness that's radical, but intensifies certain
00:14:24.920 very particular problems.
00:14:26.800 So one of those problems, I think, is connected to the use of devices generally for humanities
00:14:34.760 education in particular.
00:14:36.680 So one of the things I think many of us have experienced is the extent to which screens and
00:14:41.880 screen reading and iPhones or cell phones, the extent to which they actually produce in us
00:14:46.500 kind of habits of scanning, a kind of hyper-attention, what one scholar calls forms of hyper-attention,
00:14:51.140 not focused attention or contemplative reflection, but a kind of hyper-attention that actually
00:14:57.100 tends to kind of lead us towards a certain kind of rashness in our decision-making.
00:15:02.260 So that's one deep and profound concern I have, especially when institutions seem to be dominated
00:15:07.380 by certain sets of political commitments that ought themselves to be subject to serious
00:15:12.780 reflection and consideration, right?
00:15:14.460 So if we're in an environment where there's certain assumptions about what political positions
00:15:18.580 are normative, it needs to be the case that those can be thought about deeply and reflectively.
00:15:23.920 And if we're using technologies that limit that capacity, then we're in a little bit of
00:15:27.800 danger.
00:15:28.660 But the other one, I think for me, from the perspective as a teacher, is that, and this
00:15:33.260 would be, I mean, I'm affiliated with the classics department and I spend a lot of time in the last
00:15:38.000 years teaching Augustine, a sort of famous foundational voice for the Western world.
00:15:43.340 And Augustine is one of many thinkers who highlights the fundamental role of memory in the
00:15:49.140 constitution of our personalities, the sort of crucial role that memory plays.
00:15:53.040 And it's kind of essential to technologies like Chachi BT that we offload or offshore the
00:15:59.580 faculty of memory to the technological device, right?
00:16:03.460 It does the work for us, right?
00:16:05.140 So I don't struggle with Augustine or Dante or Homer or any of those things.
00:16:09.680 I let Chachi BT do the struggle in a certain sense.
00:16:12.660 I mean, it's not really struggling, but I let it do the amalgamation of opinion making,
00:16:16.440 information, and I'm left passive in that response.
00:16:20.100 So in that sense, I think the technologies actually inhibit the kind of interior dialogue
00:16:25.320 that's fundamental to education, but that's also fundamental to being a free person in the
00:16:30.300 world, right?
00:16:30.800 Hannah Arendt, I think, points this up very, very powerfully in her reflections on totalitarianism.
00:16:35.740 If we can't have a dialogue with ourselves, if we're pulled out of ourselves endlessly
00:16:40.200 and offshore, even our memory, we lose the ability to actually be free agents in the world.
00:16:47.180 So these are some of the things that I'm very concerned about at the level of pedagogy,
00:16:50.740 which is why I tend to a pretty puritanical, I suppose, relationship to AI when it comes to
00:16:56.440 at least to humanity's classrooms.
00:16:58.040 I recognize AI has different applications in different contexts.
00:17:02.260 It's funny, at the risk of oversimplifying it, I think of, you know, a movie that I've
00:17:06.900 watched that, you know, say is two hours long.
00:17:09.240 I could find out what happens in that movie in about 60 seconds by just reading a plot
00:17:14.340 synopsis on Wikipedia.
00:17:15.760 But I don't do that.
00:17:16.720 I watch the movie because there is something in that process.
00:17:19.800 You feel, you see, you learn, you get insights.
00:17:22.800 It's the same as why, you know, despite the fact that I may not have taken this advice when
00:17:26.120 I was in high school, reading the Coles notes of something is not the same as reading the
00:17:31.420 thing itself.
00:17:32.120 I mean, I could get chat GPT to say, you know, give me some bullet points that I can bring
00:17:36.240 up in tutorial about, you know, the cave or something.
00:17:38.980 But that doesn't mean I've done that.
00:17:40.460 So you're quite right.
00:17:42.100 And I also wonder, I mean, to appeal to your department, the classics, if you were to input
00:17:46.660 into chat GPT, the most beautiful works of literature of classics that you'd ever seen
00:17:52.940 and said, create something like this, could it do that in your view?
00:17:57.240 Could it create the beauty that we have seen from all of these people thousands of years
00:18:02.440 ago?
00:18:03.600 Yeah, that's a very interesting question.
00:18:04.780 That's a very hotly debated topic, as you may know.
00:18:06.980 Of course, in the world of visual arts, someone recently was awarded a prize, right,
00:18:11.220 in the visual arts for an artificially created, produced image, all kinds of, of course, very
00:18:19.400 deep ethical questions around AI and its accumulation of information and how that happens.
00:18:25.520 But, you know, from my perspective, for me, the answer to that question was really given
00:18:30.000 quite beautifully by Nick Cave recently.
00:18:31.760 Nick Cave, the Australian singer-songwriter, was asked this question.
00:18:36.980 A fan sent him a poem that ChatGPT had written when he asked ChatGPT, write me a poem or a
00:18:44.180 song in the style of Nick Cave.
00:18:46.480 And Nick Cave's response was to say that even if it were a good song, which Nick Cave refused
00:18:52.600 to concede that it was a good imitation, he said that the problem is that ChatGPT, artificial
00:18:59.060 intelligence, has been nowhere and suffered nothing.
00:19:02.020 And to be human in the world at all, as someone like Jordan Peterson is constantly reminding
00:19:07.720 us, is to suffer and out of that suffering either to sort of produce meaning in the world
00:19:12.340 and in our lives.
00:19:14.180 And I've been fairly persuaded by Nick Cave that no matter how close the approximation one
00:19:19.780 might be able to artificially reproduce, the fact that the technology has itself been nowhere
00:19:25.440 and suffered nothing means that that material can have very little consequence for me as
00:19:30.680 someone who lives in the world with all of its fragility.
00:19:33.300 So, yeah, so I guess that would be my answer, which isn't, I mean, yeah, maybe not the best
00:19:39.180 answer.
00:19:39.720 No, it is interesting.
00:19:41.040 And now I'm like geeking out on this topic myself.
00:19:43.860 So I think we'll have to have you back on in another show.
00:19:46.160 But, you know, I remember when I did tutorials in various classes in university, the one thing
00:19:52.060 that was always so critical when you were understanding a work was to understand the author and the
00:19:57.300 context in which they wrote a particular work.
00:19:59.940 And even if the author is some professor who's still alive, understanding how that professor
00:20:04.560 came about, you know, you read, for example, a, you know, a dissertation and you say, oh, well,
00:20:09.060 this was an environmental historian.
00:20:10.680 Why were they writing about this issue?
00:20:11.960 And with ChatGPT, that context is eroded because there is no human context or it's an amalgamation
00:20:19.460 of, you know, 150 human contexts that you don't actually know about and can't see.
00:20:23.680 So I think that's a less elegant way of describing what you've shared from Nick Cave there.
00:20:28.660 And I thank you for it.
00:20:30.120 The piece in C2C Journal is AI, the Destruction of Thought and the End of the Humanities by Christopher
00:20:36.340 Snook.
00:20:36.700 And they also have another part of this series written by Gleb Lysik, who we had on the show
00:20:41.080 a couple of weeks ago about something else entirely.
00:20:44.200 Christopher, thanks so much for coming on.
00:20:45.960 Good to talk to a real human in this day and age.
00:20:48.840 Thank you, Andrew, very much.
00:20:49.880 I appreciate it.
00:20:50.600 Thanks for listening to The Andrew Lawton Show.
00:20:53.180 Support the program by donating to True North at www.tnc.news.
00:20:58.640 Thank you, Andrew.