The Art of Manliness - February 07, 2017


#276: Utopia is Creepy


Episode Stats

Length

44 minutes

Words per Minute

166.26862

Word Count

7,449

Sentence Count

320

Hate Speech Sentences

1


Summary

In his new book, Utopia is Creepy, Nicholas Carr argues that the future envisioned by many in Silicon Valley is, well, kind of creepy. And he suggests a middle path toward technology that doesn t reject it fully but simultaneously seeks to mitigate its potential downsides.


Transcript

00:00:00.000 Brad McKay here and welcome to another edition of the Art of Manliness podcast. A few weeks
00:00:19.240 ago, I had futurist Kevin Kelly on the podcast to discuss the technological trends that are
00:00:23.660 shaping our future from driverless cars to artificial intelligence that will make new
00:00:27.560 scientific discoveries. Kevin paints a pretty rosy picture of what's to come. My guest today
00:00:32.480 sees a different side of the coin and argues that the future envisioned by many in Silicon Valley
00:00:36.740 is, well, kind of creepy. His name is Nicholas Carr, and he's the author of several books that
00:00:41.340 critique the wide-eyed utopianism of technologists. In his book, The Shallows, he reported on the
00:00:45.440 research that shows how Google is making us dumber. In The Glass Cage, he explored the science and why
00:00:50.220 outsourcing our work and chores to computers and robots might actually make us miserable and
00:00:54.560 unsatisfied in life. And in his latest book, Utopia is Creepy, Carr pulls together all the
00:00:59.860 essays he's written over the years on how the rapid changes in technology we've seen in the past few
00:01:04.060 decades might be robbing us of the very things that make us human. Today on the show, Nicholas and I
00:01:08.980 discuss why he thinks our utopian future is creepy, how the internet is making us dumber, and why doing
00:01:14.060 mundane tasks that we otherwise would outsource to robots or computers is actually a source of
00:01:18.440 satisfaction and human flourishing. We finish our discussion by outlining a middle path approach
00:01:23.140 technology, one that doesn't reject it fully, but simultaneously seeks to mitigate its potential
00:01:27.900 downsides. After the show's over, check out the show notes at aom.is slash utopiaiscreepy,
00:01:33.260 where you can find links to resources where you can delve deeper into this topic.
00:01:42.920 Nicholas Carr, welcome to the show.
00:01:45.360 Thanks very much.
00:01:46.720 So I've long been a fan of your work, The Shallows, The Glass Cage. Your new book is Utopia is Creepy,
00:01:52.820 which is a collection of blog posts slash essays you've written over the years about
00:01:57.840 technology's influence on our cognition, how we think, our culture, our autonomy, the gamut.
00:02:05.260 So let's start off with this broad question. One of the criticisms you make at Silicon Valley
00:02:09.140 in particular is that they're not just selling us gadgets and software, and that's what we think
00:02:13.140 they're selling us, but you argue they're also selling an ideology. What is that ideology,
00:02:19.780 you know, and why do you think it's bad for human flourishing?
00:02:24.520 It's an ideology that has deep roots in American culture and American history. There's a strain of
00:02:32.380 technological utopianism that runs through U.S., United States thinking going back a couple of
00:02:40.760 years. And it assumes that technological progress will bring us to, will solve all our problems and
00:02:54.340 bring us to some kind of paradise on earth. And second, and more insidious, I think it assumes that
00:03:00.900 we should define progress as technological progress rather than, and I think this is a better way to do
00:03:07.440 it rather than looking at technology as a tool that gets us to some broader, that moves us forward to
00:03:14.880 some broader definition of progress, cultural, social, economic, or whatever. With Silicon Valley,
00:03:21.200 I think it gives this long tradition of techno-utopianism a new twist and kind of a new ideology that is all
00:03:30.180 about venerating the virtual over the real. So I think, on the one hand, Silicon Valley is very
00:03:39.180 materialistic. It wants to measure everything. And what can't be measured is kind of not even worth
00:03:44.960 keeping track of or giving any value to. But at the same time, it hates the materiality of the world
00:03:52.120 and even the materiality of the human body. It wants, it believes that by virtualizing everything,
00:03:57.860 by running the world with software, and perhaps even creating a new virtual world out of software,
00:04:04.560 will solve the kind of messiness, the emotion, emotionalism, and so forth that characterizes
00:04:12.240 human beings, and also the messiness that characterizes the physical world. So I think
00:04:18.760 Silicon Valley has this kind of misanthropic ideal that physicality is the problem that we need to
00:04:27.660 solve. And if we can turn everything into algorithms and even turn ourselves into artificial intelligence,
00:04:33.460 we'll be better off. Right. They're all about making things frictionless.
00:04:38.760 Right. And it turns out, I would argue that friction is what gives interest and fulfillment
00:04:46.440 and satisfaction to our lives. It's coming up against the world and figuring it out, figuring out
00:04:52.300 how to act with agency and autonomy, gaining talents and skills, all these things that emerge from coming
00:04:59.900 up against hard challenges and coming up against friction. This is, I think this is what gives interest
00:05:04.960 to our lives. And I think the tech industry sees all of this as something to get rid of. The less friction
00:05:15.060 there is, the more, you know, everything will run efficiently and we won't come up against challenges
00:05:21.020 or hard work or things that might make us fail. But it seems to me that that's a recipe for removing
00:05:29.340 satisfaction and fulfillment from our lives. Right. And that's why you think utopia is creepy,
00:05:34.360 or at least how Silicon Valley envisions it? Well, I think beyond Silicon Valley, I would argue that
00:05:41.480 all kinds of visions of utopia tend to be creepy. There's this famous concept of the uncanny valley
00:05:50.320 in robots. So what that says is that the more humanoid a robot becomes, the creepier it becomes,
00:05:59.840 because we're very, very good at, we're social beings, we humans. And so we're very good at picking
00:06:07.760 up signals from other living things. And it becomes immediately clear when there's a robotic
00:06:15.720 attempt to mimic a human being that this is not a human being and we get creeped out. And that's one
00:06:21.500 of the big problems that roboticists have as they try to create humanoid robots is that these always
00:06:28.180 seem creepy to us. And that's the uncanny valley that it's very hard for roboticists to cross.
00:06:34.180 And I think something very similar happens in portrayals of utopias. Because almost by definition,
00:06:42.140 there's no tension in a utopia, no friction. Everybody behaves very well. Everybody is on their
00:06:50.620 best behavior all the time. And when you see that, all of a sudden you realize that people begin in
00:06:58.000 utopias, people act very robotically. They're very efficient. They have no messy emotions. They don't
00:07:04.740 get angry. And I think that that characterizes utopias in general, which is one of the reasons
00:07:11.460 that in science fiction, we're much more drawn to dystopian portrayals of the future as this horrible
00:07:17.380 mess. Whereas attempts in fiction or in movies or whatever to create a vision of utopia always turns
00:07:24.700 out to be more repellent than the dystopias because we don't see any human qualities there.
00:07:31.900 And I think the ideal that Silicon Valley has, the utopian ideal where everything is very efficient
00:07:37.780 and runs on software, is very much this kind of creepy ideal that in order to achieve it,
00:07:44.500 you have to drain human beings of what makes them human.
00:07:49.060 And why do you think they still have that drive? I mean, why don't they see the creepiness of it?
00:07:52.780 Like, I look at them like, man, that's weird. Like, why don't they see it?
00:07:57.240 Well, I, you know, I think some of it is, comes from their personalities that I don't think in,
00:08:04.000 you know, this might, this is a generalization, but I think it's some, to some extent true. These,
00:08:10.360 a lot of the leaders in Silicon Valley have spent most of their lives interacting with the world
00:08:16.540 through computer screens and that suits their personality. They're not necessarily, uh, open to
00:08:22.160 ambiguity, ambiguity, ambiguity, or to, uh, messy emotions or the kind of social conflicts that,
00:08:30.580 that come whenever you engage, um, with people face to face and with the world. So I think they're,
00:08:37.460 I think they come from, you know, I think their ideals reflect a personality that is very comfortable
00:08:45.980 with computers and very comfortable with software and programming. But when things aren't programmed
00:08:53.300 and things happen unexpectedly and perhaps inefficiently and ambiguously, they're, they draw
00:09:01.300 away from those things. So, so I think in, in some sense, what the, the world that Silicon Valley
00:09:07.020 wants to create, and it's, uh, to me, it's a very robotic world is the world that these people
00:09:13.440 actually want to live in. That's interesting, which is, it's just strange because like, you know,
00:09:18.200 the internet, the, the, one of the promises of the internet would like, it would be this sort of,
00:09:22.660 uh, it would be sort of utopia, you know, it's sort of utopia where you could see have different
00:09:26.440 types of viewpoints, different types of ideals all together that anyone can access. But the
00:09:31.260 way it's worked out is we have these people at the top who are actually structuring the internet
00:09:36.200 in a way that suits their personality and the way they like things. And we have to go along with it.
00:09:41.240 Right. And, you know, one of the things I try to, I've, when I put together Utopia is Creepy,
00:09:48.200 this collection, one of the things I did is read through my blog going back a dozen years. And,
00:09:52.900 and I realized that a lot of, a lot of the dreams about the internet and, and ones that a lot of us
00:09:59.940 held when, when we first started going online have not only haven't panned out, but in many cases,
00:10:06.020 the opposite has happened. So we, we thought that the, you know, by going online, we'd bypass
00:10:12.480 centralized, uh, centralized hubs of power, whether it's political power, media power, economic power.
00:10:21.280 And we'd have this great democratization where everyone would have their, their own voice. And,
00:10:26.420 and we'd listen to lots of different viewpoints. And what's happened is, is really, we've seen
00:10:31.720 a, an incredible centralization of power, power over information. So you get a handful of, of
00:10:38.820 companies like Google and Facebook, uh, and, and Amazon and so forth that control now huge amounts
00:10:45.500 of the time people spend online and huge amounts of the information that they get. And, and, and so
00:10:51.760 more and more of our experience is being filtered by these companies. And need, needless to say,
00:10:58.460 they're, they're motivated not only by their ideology, but by their desire to make money. These
00:11:03.840 are profit-making companies, of course. Um, and so I, I think a lot of the, a lot of the feelings about
00:11:10.880 democracy, democratization of information about people broadening their viewpoints, uh, has not
00:11:17.920 panned out. And, and I think what we're, what we're learning is that when we're bombarded by
00:11:22.980 information, the way, way we are these days, we actually become less open-minded and more polarized
00:11:29.160 and more extreme in our views. And I think we saw that, you know, in the recent election, and I think
00:11:34.140 we see it in political discourse that, that, that our hopes for, for the, our hopes about how society
00:11:41.860 and ourselves would adapt to having this constant bombardment of information just haven't panned out.
00:11:47.480 And, and now we have to struggle with consequences that we didn't foresee.
00:11:52.060 Right. And I think one of the insidious things about the internet, or at least the way it's
00:11:55.900 structured, is that it gives us the illusion that, um, we have freedom, right? Like we can spout our
00:12:01.600 opinion on Facebook or Twitter, and we think we're participating in democracy and that we're, you know,
00:12:07.360 expanding our viewpoints, but you argue, I mean, I think you just made the point that it actually is,
00:12:12.720 is an illusion. Like it actually, it reduces our autonomy and it reduces our agency.
00:12:17.660 I think, I think that's right. It, in some of that is, is simply because we become more reflexive
00:12:24.440 when we have, when we have to process so much information, so many notifications and alerts and,
00:12:29.640 and messages so quickly that we have to, we have to deal with it in a superficial fashion. And so we may
00:12:35.860 think we're, we're, you know, being participative if we click a like button or retweet something,
00:12:42.080 but really this is a very superficial, uh, way of being, uh, being involved and participating.
00:12:49.080 And it's on the terms, uh, determined by the technology companies, by the social media platforms.
00:12:55.700 Um, you know, it's in their interest to get us kind of superficially taking in as much information
00:13:02.660 as possible because that's how they collect behavioral information. That's how they get
00:13:06.680 opportunities to show us ads. Um, and, you know, I would argue that in this, that on the one hand,
00:13:14.560 you know, there's the great benefit of the internet, which is that it does give us access to,
00:13:19.000 to information and to people that used to be hard to access or, or impossible to access.
00:13:24.460 On the other hand, what it's stolen from us is kind of the contemplative mind, the calm mind
00:13:32.000 that takes this information, takes our experiences and our conversations in quietly make sense of
00:13:38.620 them. And I think that, I think that ultimately, you know, you need that space in which to think
00:13:48.120 by yourself and without interruption, without distraction, in order to develop a rich point of
00:13:53.700 view and hence, you know, express yourself and express yourself and communicate yourself in a,
00:13:59.900 in a rich way, rather than this reflexive way that we've adapted to online, which does give us this
00:14:06.960 illusion that we're constantly participating, constantly part of the conversation, but really
00:14:12.060 kind of ends up narrowing our perspective, makes us more polarized, makes us quicker to reject,
00:14:20.340 you know, information that doesn't fit with our existing worldview. Um, so I do think there's
00:14:25.760 this kind of illusion of thinking and illusion of participation that often isn't the reality of
00:14:34.100 what's going on. Right. And along that lines of participation, um, you know, one of the benefits
00:14:39.920 that technologists tout about the internet is that it makes, it democratizes the, the ability to create
00:14:45.040 content, right? We're no longer just consumers. We're also creators, but what people forget is
00:14:49.780 that like, you're creating that for the company that you're kind of working for the company for
00:14:53.960 free, right? When you create Amazon reviews or create YouTube videos or create content on Facebook
00:14:59.180 or Twitter. That that's right. And this is something I, I coined a term digital sharecropping
00:15:05.960 to a kind of an ugly term, but I think it, it describes this, that, that what, you know,
00:15:12.100 whether it's Google search engine or whether it's Amazon reviews or whether it's the entirety of
00:15:16.560 Facebook or Twitter, essentially the content that these companies use to make money off of is the
00:15:23.380 content we create. So, so similar to a sharecropping system where, you know, a plantation owner would
00:15:28.640 give a, a poor farmer, a little plot of land and some tools, and then would take most of the economic
00:15:33.580 value of any crops that were grown, uh, were given by these social media platforms, these little,
00:15:39.100 uh, little plots of land to express ourselves and to develop our profiles and, and to share and so
00:15:47.180 forth. But all of that creativity and that goes into that is monetized by the company. So, so we become
00:15:54.620 these people who, who create the content without getting any compensation for it, any monetary compensation
00:16:02.180 for it that allows companies like Facebook and Google to become enormously rich. Um, and that's
00:16:08.420 not, you know, that's not to dismiss some of the opportunities that the, the internet, the web does
00:16:14.640 really give us to express ourselves. I mean, I've, as I say, I've written a blog for a long time and I
00:16:19.560 enjoy that. And I, I feel like I've, you know, been able to clarify my own thoughts as well as speak to
00:16:26.300 an audience that I might not have, but I do think we need to recognize the kind of economic dynamic
00:16:32.480 that underlies a lot of what we do online and how in a very real sense, even though we, we don't notice
00:16:39.900 it, we are being exploited and manipulated, um, and kind of outside of our own consciousness, sometimes
00:16:48.520 we're kind of employed without pay to create these huge, very, very powerful and very rich companies
00:16:55.140 in, in, as well as very rich owners of them. Right. I thought another criticism you made was funny is that
00:17:01.180 the idea that, Oh, if you democratize that, well, I suddenly have all this great stuff, this great content,
00:17:06.160 but like most of the content that's sell, you know, amateur created is crap. Like, I mean, it's, I know
00:17:11.800 it's mean to say, but like Instagram comedians are the worst. I don't understand why people think that's
00:17:19.240 funny, but apparently it's a thing. Yeah. And it, it kind of, sometimes it shows that the audience,
00:17:27.720 you know, when they get free stuff, well, and they, they'll, they'll look for the most superficial
00:17:32.960 kind of buzz and, and that'll be enough because they don't, that they're, nobody's encouraged to
00:17:38.600 spend time kind of developing taste or, or thinking too carefully about things. Um, and you know,
00:17:44.260 this is, this kind of dream that everybody, you know, is going to be a great writer or a great
00:17:51.620 filmmaker, um, or a great musician, you know, unfortunately it's just not true. And, and I
00:18:00.360 think a lot of people understand that. Uh, I certainly understand that, you know, I'm not
00:18:05.040 going to be a great songwriter and so forth. And so, so again, this is another illusion that the web
00:18:10.340 sometimes gives us that, that it's always better. What the web tells us is this kind of myth of
00:18:17.940 participation that, you know, if you're just passively reading something or watching a movie
00:18:22.980 or a TV show or listening to a podcast, there's something wrong with that. I would argue that
00:18:27.600 that's exactly the opposites that a lot of the greatest satisfactions come from being a member of
00:18:34.320 an audience of good stuff in that we, we shouldn't feel that if we're not, we should, we shouldn't
00:18:41.060 mistake kind of a rich experience of other people's creative work as a passive experience.
00:18:47.520 It's actually very active, um, as anybody who's read a great novel or watched a great play or anything
00:18:53.160 knows. And so, you know, the web and, and a lot of the companies on the web kind of encourage us to,
00:19:01.060 uh, to think that we have to be active and participative and creative all the time. Well,
00:19:07.140 that's very important, but let's not lose sight of, uh, of the great pleasures of being an audience of,
00:19:13.300 of really good stuff. So in the shallows, um, you, you take a look at how the web has changed our
00:19:19.500 brains and you, you talked about, you began the book talking about how you've noticed your brain
00:19:24.300 change over the years. Like you can't focus as much. Um, it's hard to sit through and read a book
00:19:30.000 for a long period of time. Um, and one of the arguments you make in that book is that every
00:19:34.400 information technology, we're talking in the alphabet, uh, read the book itself and the internet
00:19:40.720 carries with it an intellectual ethic. Um, what's the ethic of the internet, of the internet?
00:19:46.880 Yeah. And what I mean by that is that, that all of these media technologies incorporate,
00:19:53.280 encourage us to think in a particular way. Um, and, and also not to think in other ways that they
00:20:00.680 don't support and that this is the ethic. I think what the, what the digital technologies in general
00:20:07.140 and the internet specifically, it, it values information gathering as an end in itself. Um,
00:20:16.100 and so what it says is the, the more information that we have access to, the faster that we are able
00:20:21.740 to process it, the more, you know, the more intensively it bombards us, the better that
00:20:28.440 more information is always better. And what, what's lost in that, I think is what everyone
00:20:34.680 used to understand, which is that information gathering very, very important, but it's only
00:20:39.660 the first stage in developing knowledge. Um, and, and certainly the first, uh, in early stage
00:20:45.720 in developing wisdom, if we ever get to that, but knowledge isn't about just gathering information.
00:20:51.300 It's about making sense of information, uh, going out, uh, having experiences, learning stuff,
00:20:57.740 finding, reading the news, uh, taking in information, but then backing away from the flow of information
00:21:04.140 in order to weave what you've learned together into personal knowledge. And this is, this is what's
00:21:10.980 lost, I think, in the ethic, the intellectual ethic of the internet. This sense that there are times
00:21:17.060 when you have to back away from the act of information gathering, if you're going to think
00:21:23.280 deeply and if you're going to develop a rich point of view, if you're going to develop a rich store
00:21:27.780 of knowledge, uh, you can't do it when you're actively distracted and interrupted by incoming
00:21:33.740 information. So I think, I think the, the internet is very, very good as a tool for information
00:21:41.800 gathering. But what it encourages us to do is to think that we should always be in the process of
00:21:47.180 gathering information. And I think that's the danger that, that the web presents.
00:21:52.420 And how do you counter that?
00:21:54.660 Not very well sometimes. I mean, and this is also something I talk about in the shallows. I think as
00:22:01.500 human beings, we are there, we have this primitive instinct to want to know everything that's going
00:22:07.300 on around us. And I think this goes back to, you know, caveman and cavewoman days when you wanted,
00:22:13.100 you wanted to scan the environment all the time because that's how you survived. Um, and so we
00:22:18.340 bring this deep information gathering compulsion into this new digital world where there's no end
00:22:25.040 to information. And as a result, and I think we see, all of us see this in ourselves. We become very
00:22:30.780 compulsive about wanting to know everything that's going on, you know, uh, on Facebook or, or in news
00:22:38.420 feeds or through notifications and so forth. And so we, we kind of constantly pull out our phone or our
00:22:44.260 computer and look at it, even if it's completely trivial information. So I think there's this deep
00:22:49.620 instinct that, that the net in, in, in technology companies tap into, uh, that can, that can become
00:22:57.140 kind of counterproductive that, that keeps us gathering information and glued to the screen.
00:23:04.000 And so to me, the, the only way I found to combat this is to resist some technology. So for instance,
00:23:14.460 I, I don't have, I'm not active on Facebook or on most social media. And it's not because I don't
00:23:21.100 see the benefits of social media. It's because I know that these systems are designed,
00:23:27.140 to tap into this instinct I have to want to be distracted and interrupted all the time.
00:23:32.520 And in order to avoid that, I just have to say, no, I don't, uh, I'm going to lose the benefits of
00:23:38.020 Facebook. I mean, one thing you, you realize when you're not on Facebook is for instance, nobody wishes
00:23:43.040 you happy birthday anymore because you're not on Facebook. Um, but nevertheless, it does seem to me
00:23:48.060 that in order, if you value kind of the contemplative mind, the introspective mind, the ability to follow
00:23:55.740 your own train of thought without being interrupted sometimes, then you have to shut down some of
00:24:00.900 these services and make the social sacrifices that are inherent in shutting down services that
00:24:07.640 increasingly are central to people's social lives. So it's, it's a very, at this point, it's a very
00:24:13.900 difficult challenge to kind of bring more balance into your behavior, into your mind, into your intellect.
00:24:21.320 And, but to me, at least my hope is that I can raise awareness that there are sacrifices that
00:24:28.740 are being made here and we should be a little more diligent, I hope, in figuring out which of
00:24:33.960 these technologies are really good for us or making us more intelligent and happier and more fulfilled
00:24:39.600 in which are simply tapping into this compulsive behavior that we often evidence.
00:24:44.600 So, you know, one assumption that technologists have is that we'll, we'll be able to find technology
00:24:50.600 to fix problems, even problems caused by technology. I mean, do you think someone in Silicon Valley
00:24:56.240 will come up with something to fix the problem of the distractibility of the internet or?
00:25:02.080 Um, I think, I think there are technologists who are trying to do that. I mean, I think we've seen
00:25:07.900 an increasing awareness among the public and, and among, you know, people in Silicon Valley or in
00:25:15.780 other, uh, technology companies outside of Silicon Valley that this is a problem that, that we have
00:25:22.600 created a system that, you know, has huge benefits and huge potential, but increasingly it is keeping
00:25:30.200 people distracted and thinking superficially and often, you know, polarized and unable to, uh,
00:25:38.780 you know, you know, give credence to people's points of view that don't fit their own. And so I think you
00:25:44.540 see kind of attempts to create, uh, apps or other software programs that reduce the intensity, the flow of
00:25:52.740 information that, that, that kind of, uh, vary the flow of information, turn off, you know, some feeds
00:25:59.940 at some, at, at times when people might be, might, might get more out of thinking without distraction
00:26:09.080 and being alone with their thoughts than, than looking into a screen, um, kind of creating a more
00:26:14.720 unitasking, uh, environment where there aren't lots of windows and lots of tabs and lots of notifications
00:26:22.140 going. The problem is that, that these are a hard sell because we've adapted ourselves very, very
00:26:31.260 quickly to, um, to this kind of constant bombardment of information in this sense that we're missing out
00:26:39.120 if we're not on top of everything that's happening all the time. So I, I do think, you know, and I think
00:26:45.640 we, we see this historically that, that often technology rushes ahead and creates problems that
00:26:52.560 were unforeseen and you can solve some of those problems with, with new technology. We certainly
00:26:57.620 see it in, in driving, for instance, with the creation of seatbelts and all sorts of technologies
00:27:03.740 that kind of make cars safer and, and so forth. But it can be, there's always this kind of tension
00:27:12.660 between the momentum a technology gains as it moves ahead and as we adapt ourselves to it and,
00:27:19.980 and, and the need to sometimes back up a little bit to redesign things, uh, to, to better fit with
00:27:27.920 what, you know, makes us interesting, well-rounded people. So, so I'm, I think, I think there are ways
00:27:36.320 to, to deal with some of these problems technologically through better design of systems, better
00:27:42.560 design of software. The question is, will we, as the public, accept those, uh, those, those
00:27:51.940 technological advances or are we stuck in this pattern of behavior that, that has been inspired
00:27:57.700 by the technology and the companies that are dominant in the technology?
00:28:02.620 Right. The other issue is there's, there's really no money in that, right?
00:28:05.040 That's, yeah, that's, as long as, I mean, one of the big issues is that we have set up the web and
00:28:13.740 social media and so forth as an advertising based system. Um, if you, if we were paying for these
00:28:19.680 things, you know, there, there was a time, uh, in the, in the error of the personal computer where if
00:28:26.480 you wanted to do something with your, with your PC or your Mac or whatever, you'd go out and you'd buy a
00:28:31.380 piece of software and you'd install it and then you'd use it for whatever you wanted to accomplish.
00:28:37.040 And that was actually a pretty good model. And we've abandoned that model for a model of, you know,
00:28:42.860 give it to me free, but distract me with ads and, and collect information about me. Um, and getting
00:28:49.000 away from that, you know, would mean actually having to pay for stuff. And, and we've, we've so
00:28:56.020 adapted ourselves to the idea that, you know, everything is free, uh, that boy, it's getting
00:29:02.600 people to pay for something that you, that they could get for free. It's a really, really a hard
00:29:06.620 sell.
00:29:07.420 All right. So, so in the glass cage, you take a look at artificial intelligence and this is the
00:29:13.360 stuff that creeps me out, uh, the most is AI. Um, I had Kevin Kelly on the podcast last week,
00:29:20.580 talked to him and he's pretty like, he's gung ho about this and it's great, but you, AI gives you pause.
00:29:26.020 Uh, why, why is that?
00:29:28.420 Well, for, for a number of reasons. And, and again, you know, I, I don't want to, I don't
00:29:35.540 want to come off as just reactively against, uh, progress in computing and progress in AI,
00:29:42.440 because I, I think there are ways that we can apply, um, artificial intelligence that would,
00:29:47.660 that would be very good. And that, that would help us out and would help us avoid some of the,
00:29:52.560 uh, some of the flaws in our own thinking and our own perspectives. Um, but I, first of all,
00:30:00.220 the AI has, the, the definition of AI has gotten really fuzzy. Um, so it's hard to know, you know,
00:30:06.640 these days technology companies call pretty much everything AI, but where I, where I see the,
00:30:12.340 the problem with artificial intelligence, as it begins to substitute for human intelligence
00:30:19.060 in analyzing situations, making judgments about situations, making decisions about it,
00:30:26.260 is that it begins to steal from us, our autonomy, our agency, and also steals from us opportunities
00:30:33.520 to build rich talents of our own. Um, and I, I think we can see this in a, in a simple way with,
00:30:41.200 um, uh, navigational technologies, um, you know, Google Maps or, or, or GPS systems in your car
00:30:50.860 that on the one hand, they, they make it very, very easy and kind of mindless to get from one place
00:30:57.260 to another. But as a result, we don't develop our own navigational skills. And also we don't
00:31:03.040 pay attention to our surroundings. And so don't develop a sense of place. And it turns out that
00:31:08.620 those types of things, the ability to make sense of, of space and of place and to be attuned to your
00:31:14.400 surroundings is really pretty important to us. I mean, we are physical creatures in a physical world
00:31:20.320 and we have evolved to be part of that world. And so what we don't in, in our drive to, to make
00:31:26.300 everything more convenient and easier, often we sacrifice the things we take for granted,
00:31:33.400 which are all about learning to navigate the world and have agency in the world and have autonomy in
00:31:39.320 the world. We kind of take those for granted. And so we're very quick to lose them in order to gain a
00:31:44.860 little bit more efficiency or a little bit more convenience. And it does strike me that, you know,
00:31:49.480 beyond the kind of doomsday scenarios or the utopian scenarios of the singularity and, you know,
00:31:55.260 computers overtaking human intelligence at a practical level, the danger is that as computers
00:32:01.880 become more able to sense the environment, to, to analyze things, to, to make decisions that will
00:32:09.760 simply become dependent on them and we'll lose our own skills and our own talents in those regards,
00:32:16.380 you know, our own ability to make sense of the world and to, and to overcome difficult challenges,
00:32:21.840 we'll simply turn on the machine and let the machine do it. And unfortunately that that's a
00:32:27.260 scenario that gives us great convenience and great ease, but also I think, and this goes back to
00:32:34.660 something we talked about earlier, also steals from us the opportunity to, to be fulfilled as human
00:32:41.140 creatures in a physical world. Yeah. Like with self-driving cars, like I, I still don't get it.
00:32:46.960 Cause like, I enjoy driving. Like, I don't know why I'd want to give that up. Everyone's like, well,
00:32:51.500 it's safer. You can be more productive. It's like, I actually enjoy driving.
00:32:55.600 And that's true of the, I completely agree with you. And you know, the last thing, even though I
00:33:01.080 realized that there are ways, and this has been a long story with, with automobiles, there are ways
00:33:05.780 for technology to make driving safer. And I think that's very, very important. The fact is that,
00:33:10.780 you know, most people in, it's like 70 to 80% of people actually enjoy driving. And, and, and it's
00:33:19.940 not like they're blind to, you know, the dangers and, and to traffic jams and to road rage and all
00:33:26.640 the horrible things that come with driving, but there's something very pleasant about driving,
00:33:31.440 about being in, it's actually one of the rare times that, that we as individuals are actually
00:33:38.080 in control of a very sophisticated machine. And, and there's pleasure that comes with that.
00:33:45.060 And there's this sense of autonomy and a sense of agency. And in, in, in, in some ways, this is a
00:33:51.840 kind of a, a microcosm of the Silicon Valley view. Silicon Valley dismisses, I think Silicon Valley
00:33:58.400 is totally unaware or of the pleasure that people get from things like driving. Um, and so that
00:34:07.440 leads them, that leads them to simply see driving as a problem that needs to be solved because there
00:34:13.620 are accidents, because there are inefficiencies, because there are traffic jams. Um, all of that
00:34:18.960 is what they focus on. Um, and so they, you know, their desire is to relieve us of what to them is this
00:34:25.880 horrible, horrible, horrible chore of driving a car. And so they, they don't realize that for a lot
00:34:33.200 of people. And that driving is really a great pleasure and, and owning a car and all of, all of
00:34:39.460 that. And, and to me, that, that kind of, uh, puts in a nutshell, the, the, the tension between the
00:34:48.180 Silicon Valley ideal and how people actually live and how they get some satisfaction out of life.
00:34:53.400 Right. So again, it's, uh, this idea that they're giving us freedom, but in the process there,
00:34:58.920 we have to give up freedom to get that freedom, right? They're giving us freedom. They're freeing
00:35:05.100 us from that, which makes us feel free. I think you could, you could say. And then, but then we find
00:35:10.540 out that we actually enjoyed those burdens when it's finally taken away. We feel existentially
00:35:15.520 empty and we're like, Oh, I don't do anything. Right. And, and, and I do think that there is,
00:35:20.540 you know, some evidence. And I think this both comes from psychological studies, but also from
00:35:25.940 our own experience that, that when we're freed of labor and freed of, of effort, we actually become
00:35:34.060 more anxious and more nervous and more unhappy. And, you know, it turns out that it's, it's the,
00:35:40.240 it's the chores that's, that software frees us from that are often the things that bring us
00:35:46.340 satisfaction that in our, in our life, that, that the experience of facing a difficult challenge
00:35:52.480 in developing the talents required to overcome that challenge, that's very deeply satisfying.
00:35:57.760 And yet, if you look at the goal of, of software programmers these days, it's to find any place
00:36:04.920 where human beings come up against hard challenges and have to spend lots of time overcoming them
00:36:09.780 in kind of automating that process. Um, uh, so there, you know, that, that's why in, in many cases,
00:36:17.600 we think our lives are going to be better when we hand over something, some chore or some task or
00:36:23.980 some job to a machine, but actually we just become more nervous and anxious and unhappy.
00:36:30.780 Right. What, what's your take on virtual reality? Cause I mean, it's crazy. Like I remember back in
00:36:35.880 the nineties, like I'd go to the science museum and they had the VR thing. Um, you could go through
00:36:41.140 the human digestive system. That was the thing. Um, and it was like, and then I thought this is
00:36:46.080 the future. This is amazing. And then like it died, didn't go anywhere. Now we're seeing this
00:36:49.960 resurgence. Um, is, does virtual reality give you pause? Do you think it's going to catch on this
00:36:55.560 time? I think it's going to, well, I mean, there's the question, there's the kind of physical
00:37:00.300 question of how long can people be in a, you know, virtual environment without getting nauseous or
00:37:07.300 dizzy or whatever. And so let, let's assume that that will be solved, that, that we'll figure out
00:37:13.880 how to create, you know, systems of, of virtual reality that actually don't, that, that are actually
00:37:20.400 pleasant to be in. Well, I, I think it, I think it will have successful applications. I mean, I can see
00:37:28.420 it in gaming. I can see it in certain marketing aspects. You know, if you want to, uh, if you're
00:37:34.800 looking to buy a house or something or rent an apartment, you'll be able to put on your virtual
00:37:38.780 reality goggles and walk through the space. Uh, you can certainly see applications in pornography,
00:37:43.900 which will probably be one of the first to come along. But what I don't think will happen, I,
00:37:48.560 I think there's, there's this belief, Mark Zuckerberg, I think from Facebook has, has stated it,
00:37:54.900 that virtual reality goggles will become kind of the next interface for computing. So we'll,
00:38:00.760 we'll spend lots of time with our goggles on or some kind of virtual in, in some kind of virtual
00:38:07.120 reality, um, in order to, you know, socialize. And that's, that's what social media will become
00:38:14.240 and what personal computing will become. I don't think that that's going to happen because,
00:38:17.980 and I think you see signs of this from like the failure of Google Glass, that there's something
00:38:23.100 about, I think there's some deep instinct within us and within probably all animals that resists
00:38:31.480 having something else control our senses, something else control, something get in the way of our field
00:38:38.640 of vision. I think we can do it for brief periods. We can do it when we're playing a game or, or when
00:38:43.460 we want to accomplish a particular task that, that can be accomplished through virtualization of,
00:38:49.400 of space or whatever. But I don't think we're going to see people walking around, uh, with virtual
00:38:55.700 reality goggles or even with simpler, uh, device projection devices. I think, uh, I think I,
00:39:03.120 I'm very dubious about kind of smartphones being replaced, for instance, by virtual reality systems.
00:39:11.420 Right. Because it looks goofy and looks creepy.
00:39:13.100 It looks goofy and it looks creepy. And it also, it feels, you feel vulnerable. You feel weird
00:39:18.900 when, when you're, when you're seeing, you know, something that some, some, when you're cut off from
00:39:26.260 the actual real world and embedded in a world that somebody else is manipulating. I, I mean,
00:39:33.020 it, it's disorienting and it's also, I think we're repulsed by it after a while.
00:39:38.480 Well, Nicholas, people who are listening to this and they're, they agree with you. Like,
00:39:41.740 yeah, utopia is creepy. Like, I don't want any, I'll, I'll take some of this, uh, utopia that
00:39:46.920 they're offering because there's some benefits to it, but there's parts of it. I just know,
00:39:50.600 I don't want to go. Is it possible to opt out? Um, I guess it is possible. You, you, you don't do
00:39:55.860 social media. Um, yeah. Any other ways to opt out?
00:40:00.020 I mean, I do a little, for instance, I'm on Instagram, but I have a private account with,
00:40:05.180 you know, a handful of close friends and family members and it's really good. I mean, it's,
00:40:10.080 you know, if you restrict, if you place certain restrictions on social media, I think it can be
00:40:16.080 very, very valuable and very fun. So I'm not, you know, I'm not arguing for total opting out if,
00:40:23.160 as if that were even possible. I, I mean, I think one thing we know about the internet and computers
00:40:29.540 and smartphones is that it's actually very, very hard to live these days without, without those
00:40:36.200 kinds of tools because society as a whole, our social systems, our work systems, our educational
00:40:42.080 systems have rebuilt themselves around the assumption that everybody's pretty much always
00:40:48.240 connected, or at least has the ability to connect very frequently. So I don't think, you know, some
00:40:53.400 people will, will opt out totally just as some people opted out of television totally and so forth.
00:40:58.740 But those, I think those will be people on the margins. For most people, I think it's really,
00:41:04.780 the challenge is really more about developing a, developing a sensibility of resistance rather
00:41:14.240 than a sensibility of rejection. And, but, you know, often tech, techies will, will quote Star Trek and
00:41:21.780 say, resistance is futile. You know, the Borg of the internet is going to take us all over. So just give
00:41:26.580 into it. I think that's absolutely the wrong approach. I, I think it's valuable to resist
00:41:33.500 these kinds of powerful technologies. And, and this is a powerful technology. It's a media technology
00:41:41.060 that wants to become the environment in which we exist. And I think it's important to resist.
00:41:47.160 And by resist, I mean, instead of, instead of being attracted to whatever's new, to the latest novelty,
00:41:53.880 to the latest gadgets, the latest app, always pause and say, you know, how am I going to use this? How
00:42:00.180 do other people use this? Is this going to make my life better? Am I going to be, am I going to be
00:42:06.540 happier? Am I going to feel more fulfilled and more satisfied if I adopt this technology? Or am I just
00:42:13.520 going to be more distracted, more dependent on, uh, technology companies, uh, more, less able to
00:42:21.560 follow my own train of thought, uh, less well-rounded. And I think if we just start asking these
00:42:27.640 questions, um, and everybody's going to have different answers to these, but if we start asking
00:42:33.060 to these questions, I think we can be, become more rational and more thoughtful in what technologies
00:42:39.340 we adopt and what technologies we reject. And ultimately I think that's the only way to kind of
00:42:45.760 balance the, the, the benefits and the good aspects of, of the net and all, all related technologies
00:42:53.120 with the bad effects. And, and by now I think we all know that there are bad effects, that this isn't
00:43:00.360 just a story of, you know, uh, everything getting better. It's, it's a, it's a story about costs and
00:43:08.100 benefits and we have to become better at balancing those. And that really does mean becoming more
00:43:13.980 resistant and more skeptical about the technology in the promises being made about the technology.
00:43:19.500 Well, Nicholas, this has been a great conversation. Where can people learn more about your book and
00:43:23.220 your work? Um, well, you can go online. Yeah. I have, uh, my personal site is nicholascar.com where
00:43:31.280 you can find out information about my books and links to my articles and essays that I've written
00:43:35.840 in, in my blog, um, which I, I still write though, not as intensively as I used to is called rough
00:43:41.660 type. And you can find that at rough type.com. Awesome. Nicholas Carr. Thank you so much for your
00:43:47.080 time. It's been a pleasure. The pleasure was all mine. Thank you. My guest today was Nicholas Carr.
00:43:51.840 He's the author of several books, including the shallows, the glass cage and utopia is creepy. All of
00:43:56.900 them are available on amazon.com and bookstores everywhere. You can find more information about
00:44:00.320 Nicholas's work at nicholascar.com. That's car with two R C A R R. Also check out our show notes
00:44:06.380 at aom.is slash utopia is creepy. We can find links to resources where we can delve deeper into this
00:44:11.560 topic. Well, that wraps up another edition of the art of manliness podcast. For more manly tips and
00:44:27.660 advice, make sure to check out the art of manliness website at art of manliness.com. The show is
00:44:31.980 recorded on clearcast.io. If you're a podcaster who does remote interviews, it's a product that I've
00:44:36.380 created to help avoid the skips and static noises that come with using interviews on Skype. Check it
00:44:41.660 out at clearcast.io. As always, we appreciate your community support. And until next time, this is Brett
00:44:45.980 McKay telling you to stay manly.