00:00:00 Asha Curran:

Good evening, everyone. My name is Asha Curran , and I have the privilege of being the director of the Center for Innovation and Soc ial Impact here at the 92nd Street Y. On behalf of everyone here at 9 2Y welcome. We are very glad to have you here tonight.

Tonight's debate is part of our third annual Seven Days of Genius Festival, which invites leading thinkers to explore all aspects of genius, how we define it, how it emerges across communities and cultures, why it matters, and what the future of genius looks like. You can find out more at 9 2Y.org/genius. And please go online and share your impressions of to night's event and all of our Genius Festival using the hashtag # thats genius . The festival would not be possible without the generous support from various individuals and organizations, including the John Templeton Foundation. Please join me in thanking them.



[laughter]

00:01:01 Also, a big thank you to our partners at Inte lligence Squared U.S. for bringing this amazing event here tonight, and to our four debaters for what I am sure will be an incredible and fascinating look at many implications of artificial intelligence. Now,

please join me in welcoming John Donvan, the h ost and moderator for Intelligence Squared U.S.[applause]





John Donvan:

Hi, everyone. Thanks so much. It's great to be crossing town tonight to be part of the Genius Festival. I personally have always wanted to speak from this particular stage, so i t's an honor. Then, when I saw they had the word "Genius," in the title, I got a little worried that I wasn't going to be allowed onstage. But I slipped in through the back door. For those of you who don't know our program -- and I know actually a lot o f you have crossed town with us -- I just want to take a couple of minutes to explain your role as members of audience, because it is not merely to be passive observers. You are active participants in a couple of ways.

00:02:00 The most important way is that at a certain point in the debate, we ask you to register your vote on the motion before us: Don't Trust the Promise of Artificial Intelligence.

And the way that we have you vote -- if you look at your seat, there's a keypad attached to it. And when the time comes, I'll ask you to pick it up and use keys 1, 2, 3, to vote for -- that's 1 -- against -- number 2 -- or remain undecided -- position yourself as undecided on the motion, that will be position number 3. Again, a point will come in the debate when I explain to you that that's what we're doing. And it'll be pretty clear.



But when the time comes to vote, the way this particular keypad works is I need you to hold down the key that you choose until you see the number r egister in the little window. And then you'll know that your vote has registered. Also, in the middle of the debate, I actually will come to you for questions. And I need to make clear that I'm pretty tough about the questions. I need them to be short. I need them to be under 30 seconds.

00:03:01 I will ask you please not to debate with the debaters. Let them debate with each other.

But ask a question that gets them to debate with one another better on the motion itself. And again, put it in the form of a question. I'm fine if you want to put a little statement at the beginning. But then, form it as a question. And you'll know it's a question if a question mark naturally fits at the end of whatever you've said, then you've hit the tone correctl y. One last thing. We are creating and we always do create a podcast out of these debates. So, you'll get to see the sausage being made. There will be a number of times when I say things like, "We'll be right back." And you'll see that I haven't gone anywhere. I'll still be here. And that's to let you know that we're working for breaks in the podcast. And for the same reason, I'll ask you from time to time to explode into a round of spontaneous applause. I think it'll be obvious , it's when Iintrod

uce the debaters, et cetera. But in case you miss it, I'll give you one of these, and that'll be the gesture to applaud.

00:04:00 The -- we do like the podcast audience to know that you're here because ultimately you're going to be picking the winner s. Let them know you're here. It's fine with us to applaud points that you like. We discourage -- very much we discourage booing and hissing. We want to keep it positive, not negative. So, the most negative that we would be comfortable with would be s ort of sardonic chuckle or something like that if you don't like a point.

00:05:10 But feel fine -- feel fine to applaud points that you like. We always start these debates by talking a little bit about the relevance. Why we're doing them now. And the way we do that is I invite to the stage the chairman of Intelligence Squared, Bob Rosenkranz .

And we have a little bit of a chat. But this time, we have two guests to bring to the stage. So, let's bring them to the stage, and I'll introduce everybod y when we're sitting down.



[applause]



Hi Bob. Hi Gail. So, you can grab a mic. So, I'm joined by Bob Rosenkranz and also Dr.

Gail Saltz, who is an author and a psychiatrist. And she is the chair of the 92nd Street Seven Days of Genius Festival. Gail, thanks so much for joining us on this one. And Bob,

it's a pleasure to be here, as I just said, with 92nd Street Y. And talking, in terms of relevance, we have a little bit of news on our topic today.

00:06:03 Robert Rosenkranz:

Well, we do indeed. Where the Google -created program beat the world champion at the game of Go. And I think it's fascinating, from an AI perspective, because the program actually learned the game just by observing hundreds of thousands of professional games. And this is a very different and much more advanced way of learning then, let's say, the chess program that 20 years ago was able to beat champions. But -- where it was basically following algorithms that had been generated by expert human players.



John Donvan:

Well, I want to come back to you in just a second. Let's talk a little bit about your personal experience with AI. But Gail, I also want to bring you into the conversation totalk about this festival and what we mean by "Genius," and what we mean by "Seven Da

ys of Genius."

00:07:00 Gail Saltz:

Right. Well, first of all, for those who don't know, the 92nd Street Y is a cultural center and a community center. And the mission is to really look at big ideas. One of them is genius. So, the Seven Days of Genius -- this is the third year. And it's been to explore the idea of genius, which doesn't mean the same thing to everyone.



John Donvan:

[affirmative]



Gail Saltz:

Of course we think of extraordinary -- the ability. We think of extraordinary ideas that change humanity. But in this case, this year, we're looking at the way genius ideas can impact world challenges for social good. So, we're here in New York all this week with amazing programming. We're in 50 communities around the world this year.



John Donvan:

Wow. Wow.



Gail Saltz:

Issues like inequality, food shortage, sustainability. So, this is different this year. And tonight, we're looking at artificial intelligence. And in that same vein, the question of --

is it for social good or does it imperil social good?

00:08:00 John Donvan:

And that's where I want to bring it back to Bob, because, as an investor, you've dabbled a bit in the artificial intelligence realm. What was your experience?



Robert Rosenkranz:

Well, it was interesting. And I think it illuminates maybe a couple of aspects of tonight's debate. So, I funded for a period of about four or five years a group of computer scientists who were trying to do something that people have not been very successful in doing, which is outperfor ming the stock market. And the ultimate result -- and it used all of the techniques of statistical analysis, of correlation, of machine learning. This was a very advanced effort. And we ultimately got to the point where we could do maybe 50 or 75 basis points, or hundredths of one percent -- better than the averages, which will put us maybe in the top 10 or 15 percent of professional money management organizations.

00:09:02 It wasn't enough of an advantage to build a real business around. But the interesting thing is that I had six people who were basically outperforming teams that typically would have hundreds --



John Donvan:

[affirmative]



Robert Rosenkranz:

-- of analysts, of traders, or portfolio managers. And that can -- I mean, I -- to me, it shows both the difficulty of achieving a result with artificial intelligence, particularly the result that humans are not very good at.



John Donvan:

[affirmative]



Robert Rosenkranz:

And secondly, the potential for job loss, because the kinds of jobs that we would be displacing are jobs held by people who were business school graduates, and highly educated, and very cognitive , and very hardworking. And you know, it can all be replaced by a handful of guys running a computer.



John Donvan:

And it's the depth of these questions, Gail, that exactly sound like what you think this festival, and this particular debate, should be about.

00:09:59 Gail Saltz:

I think this year -- I mean, this is the important thing. Many genius ideas can be a double -edged swo rd. And you have to look at implemen ting them within the context of our real world and look at both the ups and downs, and I'm excited tonight to hear this debate and --



John Donvan:

All right.



Gail Saltz:

-- in the light of that.



John Donvan:It's the double

-edged sword and we have two edges coming out to the stage, two teams on both sides. So, let's thank you, Gail, you Bob, and let's bring our debaters to the stage. Let's welcome them.



[applause]



Okay. So, that was one kind of start, and now we're going to do the start for the podcast, which will once again begin with your voices through your applause to get us all launched.

00:11:04 So if you could give another round of applause. Thank you.



[applause]



Let's talk for a moment abo ut how Hollywood sees robots, has machines operating autonomously using artificial intelligence, also known as AI, and in the movies these AI running machines are usually human like, usually also either very good like R2D2 in Star Wars or very, very evil like Schwarzenegger in The Terminator, and you might say that that is a false dichotomy, that it's a Hollywood thing, but in fact, the potential for artificial intelligence, to make the world either a very much better or a very much worse place for people is at the center of some pretty furious discussions taking place in places like Silicon Valley where companies like Google and Facebook and Apple are investing massively in AI technology these days, such that the technology is bound to continue to touch our lives, even more than it is already.

00:12:06 And the question is : I s that a good thing or is it a bad thing ? Is it even a real thing? All of which makes the sounds -- sounds like the makings for a debate, so let's have it. Yes or no to this motion : Don't Trust the Promise of Artificial Intelligence, a debate from Intelligence Squared U.S. I'm John Donvan. We are at the 92nd Street Y as part of their S even Days of G enius Festival with four superbly qualified debaters who will argue for and against the motion: Don't Trust the Promise of Artificial Intelligence. As always, our debate will go in three rounds and then our audience will vote to choose the winner and only one side will win. And we want to have your preliminary vote on this motion right now. So let's go to the keypads at your seats. Take a look again at this motion and be careful because it's got a negative in it. The motion is Don't Trust the Promise of Artificial Intelligence.

00:13:01 If you agree with the Don't Trust, push num

ber one. If you disagree, don't Don't or Trust, push number two, and if you're undecided push number three. Hold down that key until you see the number register in the little window and then you can release. That will be your vote locked in. The other keys are not live, and we'll give this about 15 seconds and then we'll lock it out. Okay. It looks like everybody is good. I just want to check off stage, are we good to move on? We're good. Okay. So we're going to move on, but here's what I want to explain. The way that we decide the winner in our Intelligence Squared debate is by the difference between two votes, the votes that you just made, and then again at the end of the debate, after you've heard all of the arguments, we have you vote a second time and it's the difference between the two votes in percentage point terms that will determine our winner.

00:14:10 So you are picking the winner and you're picking the winner with your second vote, and I want to be clear. It's the difference between the first and second vote that will declare our winner. So, our motion is this: Don't Trust the Promise of Artificial Intelligence. Let's meet the team first arguing for the motion. Ladies and gentlemen,

welcome Andrew Keen.



[applause]



So, Andrew Keen, you are an internet entrepreneur. You're currently executive director of the FutureCast salon . You're host of the internet chat show, "Keen On." You're author of the book, "The Internet is Not the Answer." Your second time debating with us. The first time the motion was Smart Technology is Making Us Dumb. You argued for the motion. The debate ended in a tie, the first and only time we've had that happen in more than 100 debates.

00:15:00 So, is smart tech still making us dumb and do you expect a better outcome than a tie tonight?



Andrew Keen:

Well, I certainly expect a better outcome, because our voters are humans rather than machines . So I'm trusting you to do a good job tonight, all of you out there.



John Donvan:

All right, an appeal to the home crowd team here. Ladies and gentlemen, Andrew Keen.



[applause]



And, Andrew, please tell us who is your partner in this debate?Andrew Keen:



My partner is the great philosopher and poet of Silicon Valley, Jaron Lanier.



John Donva n:

Ladies and gentlemen, Jaron Lanier.



[applause]



Jaron, you are also arguing for the motion, "Don't Trust the Promise of Artificial Intelligence." You are a former goat herder and midwife, an accomplished musician,

and artist, a computer scientist, the father of virtual reality, and an interdisciplinary scientist at Microsoft Research.

00:15:59 You are author of the book, "You Are not a Gadget." You are often described as a pessimist when it comes to our digital future. Is that fair?



Jaron Lanier :

No, it's absolutely unfair and I would never describe myself as a pessimist. A pessimist is somebody who believes things are worth improving and attempts to do so. The -- it's the Panglossian optimist is the lazy bum who's just like, "Oh, it's all great." And I am not that person.



John Donvan:

All right, well we'll get to know you better as the evening goes on.



[laughter]



Ladies and gentlemen, the team arguing for the motion.





[applause]



And we have two debaters arguing against the motion, "Do n't Trust the Promise of Artificial Intelligence." That means they are arguing, "Trust it." First, please let's welcome James Hughes.



[applause]



James, you are a sociologist, a bioethicist, a former Buddhist monk. You are the executive director of the Institute for Ethics and Emerging Technologies, author of,"Citizen Cyborg: Why Democratic Societies must Respond to the Redesigned Human of the Future."

00:17:05 You are also a proponent of something called democratic transhumanism which means what?



James Hughes:

Transhumanism is the belief that our descendants will be strange and wonderful, that humanity is a work in progress, that we can use technology to be smarter and happier and live longer and healthier.



John Donvan:

You're the optimist?



Jame s Hughes:

And I'm the optimist, yes.



John Donvan:

Okay, and so please tell us who your partner is.



James Hughes:

My partner is the renaissance woman, Martine Rothblatt.



Martine Rothblatt:

Thank you.



John Donvan:

Ladies and gentlemen, Martine Rothblatt.



[applause]



Martine also arguing against the motion, "Don't Trust-- " that means trust, "The Promise of Artificial Intelligence." You're author of the book, "Virtually Human: The Promise and the Peril of Digital Immortality." You launched the GPS navigation system GeoStar in the '80s. You founded Sirius Satellite Radio. You have founded a biotechnology company.

00:17:59 You founded a religion. And then in 2010, you commissioned a robotic clone of your wife called Bina48. You have had a pretty good track record for success. But will Bina48 therefore someday be a fully conscious being?Martine Rothblatt:



So I think she will be someday a fully conscious being. It's not going to be in the business planning cycle or an election cycle. But if you take a look at the time frame of social movements or technology paradigms, 30 years ago as you mentioned, there was no GPS. Today there's GPS in three billion people's hand phones. So I think in a similar time frame we'll go from having like no GPS to lots of GPS. We'll go from no cyber consciousness to there'll be billions of cyber consciousness.



John Donvan:

Okay, sounds like another optimist. Ladies and gentlemen, Martine Rothblatt and the team arguing against the motion.



[applause]



Now, I want to repeat, "This is a debate." It's a contest, a contest of ideas and logic,

presentation, a little humor is also allowed, a little bit of charm.

00:19:01 But, ultimately, these debaters are trying to persuade you to vote for their side in the second vote. I want to remind you, once again, you have voted already. You'll be asked to vote again at the end of the debate and it's the team whose numbers have moved the most in percentage point terms between the first and the second vote will be declared our winner. We go in three rounds. Let's move onto round one. The motion is this, "Don't Trust the Promise of Artificial Intelligence." Speaking first for the motion from the lectern, Jaron Lanier. He's a computer scientist, a composer, author of the book, "Who Owns the Future?" Ladies and gentlemen, Jaron Lanier.



[applause]



Jaron Lanier:

Hey. I can't possibly be talking about artificial intelligence right now without remembering Marvin Mi nsky who is my dear mentor who just passed away very recently.



[applause]

And Marvin was infinitely sweet to me.

00:19:59 And decades ago we would have the same argument, more or less. The spirit was a little different though because back then it seemed a little clearer where the line was between the philosophy and the technology. So Marvin would argue the hard AI position and he was largely the author of a lot of the ideas that I suppose I find myselfarguing against now. Oh, but he loved to ar

gue. And I just remember these arguments with such joy. And I wish that people who have come along since retained that sense of open -mindedness and didn't take it all so seriously. I feel as though the ideas about AI have evolved into something of an o rthodoxy, where they started off as an element of good humor and camaraderie. And I regret seeing that. And I wonder if similar progressions happened with some of our great religious and philosophical traditions.

It's been interesting to see it. So, fr om my perspective, as a practitioner, I must add, I --

my friends and I sold the Machine Vision company to Google.

00:21:01 I'm in the thick of it. I'm not anti -AI algorithms. I make them, you know ? And I'm fascinated by them. So, from my perspective, there has to be a division made between the work itself -- the engineering and the science on the one hand, and then on the other, the storytelling about it, the narrative that we have about it, the fantasy life of it -

- perhaps the religion of it. These are two distinct things. It doesn't mean one is good and one is bad, but they're just different sorts of beasts. So, to say, "What is the promise of an area of research?" We fundamentally don't know. It's research. It's basic research. We just ob served gravity waves for the first time. Does that mean we'll suddenly have anti -gravity devices ? Well, you know, maybe someday. We have absolutely no clue what we're going to discover. And in the same sense, we currently don't know what a thought is, in terms of scientific description. We can kind of find collections of neurons that seem to active at certain times. It's provocative.

00:22:00 We can replicate certain functions. It's provocative. Do we understand how brains work? No. And yet, the work is fascinating, the work is important. A lot of our existential threats as a species involve great complexity. We wouldn't even know about climate change if it weren't for masses of sensors networked together that would allow us to get a big picture. We would be blind to our perils without it. So, to say that there isn't promise in pursuing science information systems and sensing systems, and algorithms to understand it all, like the -- it would be crazy to argue that. I can't imagine any serio us person taking that position. So, the only form of this proposition that I can possibly argue against is this other realm, th e fantasy life, the culture. And here, I do find myself not enjoying it. And I'd like to just say a few things about where it fails. And this is rather personal, perhaps , but I will share it with you.

00:23:01 One issue is, as an engineer, if I say, "Oh, I'm making this algorithm into something that's intelligent," or conscious, or whatever -- or cultured, or whatever it is -- form an engineering point of view, I can't define those terms. So, it makes me nuts. I have no baseline. And this is a really crucial, crucial point. A lot of the systems you call smartsystems are kind of derailed from the empirical process. As an

example, I frequently review student work. And some people say, "Well, here I've made an intelligent system.

For instance I saw a machine that makes beverages." And it had quite fancy software.

And it's to choose the beverage for a person. People found it hard to use. And what I'm saying is nobody cares how fancy your algorithm is. The only thing you can measure is how well the machine works, in the end. So, if you don't define a baseline that's measurable, you're off in fantasy land. And this -- t his might seem like a bit of a wonky point to you, but it's absolutely crucial.

00:24:04 Engineering without clear concrete grounding in reality goes off the rails and does become dangerous. So, for a more important example, if you have some sort of d rone that goes around killing the wrong people, whether it did so because of an intelligent algorithm that made the wrong decisions, or just because it's malfunctioning -- who cares ? That distinction is not actionable or meaningful. The only thing that m atters is whether people can use machines that we design responsibly, with intention. That is the meaningful question. Everything else is fantasy. And so, adding this whole layer about power transcending and everything, it's -- it just confuses matters. There is an economic angle to it. I love automatic machine translation. I love that you can go online and get something converted to German automatically. My own lab does that.

00:24:59 We have real -time Skype translation now. But the only way we do it is by scraping the efforts of millions of translators who don't even know what's happening to them to get the examples. And in order to have the fantasy that this thing is a free -standing creature, we're pretending t hese other people don't exist and we're creating potentially a massive wave of technology -driven unemployment that doesn't need to happen. We shouldn't be shrinking the economy over a fantasy, if we just acknowledge that the people are just contributing in new ways. Their Go games are informing a Go algorithm.

Pay those people instead of having to resort to some sort of weird socialist solution.

You know, this fantasy of these artificial creatures makes us ignore our own lives, our own contributions. And so, it creates a needless eco nomic peril -- absolutely silly. But then I have to say something else. I understand that for many people, these ideas of AI have become very tender and dear to them.

00:26:03 They've become part of the way they think about their lives, and their loved ones, and their future. They think they might become immortal or something. I have absolutely no interest in ridiculing or opposing that. I absolutely believe in religious free dom, and I would never, never, never speak against somebody's beliefs. I respect them. All I askfor is the separation of church and state. Without a separation of church and state,

there can be no religious freedom. Never more true than when it comes to AI.





John Donvan:

And that's it. Thank you. Jaron Lanier.



[applause]



The motion is Don't Trust the Promise of Artificial Intelligence. And here to make her opening statement against the motion, Martine Rothblatt. She is chairman and CEO of Unit ed Therapeutics and author of the book, "Virtually Human." Martine, you can make your way to the lectern. Ladies and gentlemen, please welcome Martine Rothblatt.



[applause]

00:27:00 Martine Rothblatt:

Thank you. It's my pleasure to speak on -- in f avor of the motion that we should trust the promise of artificial intelligence. And in thinking about this question, I thought the major theme sort of orbited around three key words. And these words came to me when I asked myself, "Well, what do we mean by the promise of AI?" As Jaron said so well, there are so many different words thrown around. Define your terms. So, to me,

the promise of AI revolves around the three terms: replication, and application, and then fascination. So, with regard to replic ation, the promise of AI is that we will be able to replicate the human mind. And that is a -- you know, a startling statement.

00:28:01 It's a -- kind of an awe -inspiring statement, because, I think, for most of us, for all the beauty and credulousne ss that we have about life around us, human minds are just like the most awesome things anybody can imagine. So now we're talking about replicating.

The promise of AI is that we can replicate the human mind. But I believe that this promise will be fulfilled because we're not talking about replicating the structure of the human brain with the hundreds and billions of different subtle neural connections, of which, as Jaron says, we have slight knowledge. Instead, we're talking about replicating the functi on of a human mind, much in the way that we're not able to replicate a bird,

that we're able to replicate flight. And certainly, the flight that we have with an airplane or helicopter is not the same as the flight of a hummingbird, but it's flight nonethe less.

00:28:59 And the promise of artificial intelligence is not that we will replicate every little nuance of a biologically -human mind, but it will be a replication of human consciousness,nevertheless. The second term that to me kind of embraces the concept of the promise of artificial intelligence is the term application. So, from replication to application.

What do we mean? The promise of AI is that this stuff will be good for things, that we will really have uses for artificial intelligence. Otherwise, why bother developing it

?

And in fact, lots of people all over the world are enthralled with the promise of AI. We talk about it being helpful in helping to navigate our airspace, our ground space, our traffic. We talk about it being helpful in terms of health care, discovering cures for new diseases -- for -- new cures for diseases.

00:30:00 The application of artificial intelligence that I find most amazing is the application of it to diseases of the mind, specifically diseases such as dementia and Alzheimer's. And if we are able to go ahead and de velop artificial intelligence so that it can serve as a kind of mental wheelchair, something that a person that has lost a lot of the faculties of their mind can instead rely upon because it provides a pretty damn good replication of their thoughts, of their responses, of their ability to recognize loved ones and respond to one,

the ability to talk and form sentences and have an interior sense of reality. That promise of artificial intelligenc e, I believe, will come to pass because there's an enormous demand for it. Everybody who thinks about it wants to contribute to it in some part, small or large, depending on their capability and their skills. So I think it's all of the incredible applica tions of AI that will pull AI forward into the future and realize its potential.

00:31:06 Finally, once we've replicated our minds and once we've been able to develop applications that makes these AIs so useful to us and such an important part of our l ife,

far more than a personal digital assistant, far more than a great educational tool. In fact, it's a mind. It's something that we develop a relationship with. The next thing that comes to me is fascination. We will love these AIs. We'll love them in the same way that we love our cats and our dogs, that we love our friends, ones that we see distantly or the ones that we see frequently, because if it is a replicated human mind it'll have all the cool features of human minds, being able to answer ques tions, be able to really frame the rest of the sentence before a sentence is finished, be able to feel empathy and when we're sad help us feel better and we're happy join in that joy.

00:32:04 Now part of the fascination of AI is that I believe the fas cination part will only attach to the friendly AI, and this is why I believe when people offer a promise of AI as it being friendly and not be kind of the scary AI that you see out of Hollywood, I think that is a reality that we can expect to pass. Our fascination with AI will be for the AI that is friendly to us. We have no fascination with cars that don't stop when you put on yourbrake.

In fact, we run from them. We have no fascination with stoves that when you turn on the stove they explode. We don't buy those. AI will arise in a natural environment in which humans are the agents of selection. We will select for the friendly AI and we will stamp out the unfriendly AI. So I believe that the promise of AI will be a good one and we should believe it because the environment in which AI evolves will be a human selection environment and the mass activities of hundreds of millions of people will select for the friendliest AI.

00:33:12 This does not mean that there will never be bad AI. This does not mean that there will never be broken applications and that some replicated minds will be as messed up as people who are suffering from something like an antisocial personality disorder, but ultimately it will be the efforts of millions of individual hacke rs throughout the world in a decentralized process drawn forward by the promise of the positive applications of AI,

such as helping to save relatives and loved ones that will result in a beautiful AI that will be the fascination of all of us and therefore will be a definite survivor in the future to come. Thank you.



[applause]



John Donvan:

Thank you, Martine Rothblatt.



[applause]



And a reminder of where we are.

00:33:59 We are halfway through the opening round of this Intelligence Squared U.S. debate. I'm John Donvan. We have four debaters, two teams of two, arguing it out over this motion: Don't Trust the Promise of Artificial Intelligence. You have heard the first tw o opening statements and now onto the third. Making his way to the lectern, Andrew Keen. He is executive director of F utureCast and author of the book, "The Internet is Not the Answer." He will argue for the motion: Don't Trust the Promise of Artificial Intelligence. Ladies and gentlemen, Andrew Keen.



[applause]



Andrew Keen:

Thank you, John. So this is a big deal. V ery, very big deal. You just heard Martine given an extremely good speech in which she talked about replicating the human mind. Nobody laughed. You all sat there. You took that seriously, the notion that wecan replicate our minds, that they'll live fo

rever in the cloud or some other digital space is not outrageous. Jaron might argue it's not practical at this very moment, but it isn't outrageous.

00:35:05 This is a huge issue. The notion that we can replicate ourselves, our intelligence, our bein g, our identities, our souls, whatever other word you want to use. So this is not a debate irrel evant to the future. Indeed, it is probably the central debate of the 21st century for better or worse. Important to remind ourselves of this debate, we are not talking about trusting artificial intelligence. That is not what Jaron and I are arguing against. We're arguing that we shouldn't trust the promise of artificial intelligence. We're not against the technology in itself. That is not the subject of the debate tonight. Jaron talked about the Separation of Church and State. We talked about this over lunch today.

00:36:00 What I think he means is that we need to separate the science from the belief. And, all too often in this enormously important debate about whether or not we should trust the promise of artificial intelligence, of smart machines that are indeed so smart that they are able to replicate us as a species both collectively and individually, Jaron is arguing that we're muddling them up. Jaron is arguing that there's too much state in the church and too much church in the state. And the philosophers are intervening. The philosophers are telling us that this science is good, is moral, is liberating. And that's, I think, the argument that our team is making, or certainly that I am making. It is that there's a problem not with the technology but with its promise,

with the ideology around that technology.

00:37:04 We already heard a little bit about it from Martine who talked about so mething called "good versus bad " AI. I have no idea who she can make those moral judgments. I have no idea how you moralize I -- AI unless you humanize it, which itself is deeply problematic. But the problem today, I think, is that the philosophers ha ve got hold of this technology and they're presenting it as liberation theology. At least some of them are. We've been through this, of course, before in the middle of the 19th century in the midst of industrial revolution. The philosophers got hold of that. Marx thought the Industrial Revolution would free us from work, would free us from inequality. He was of course entirely wrong. And we have a similar kind of discourse, a similar kind of philosophical debate emerging today about the promise of AI.

00:38:05 We're told that this AI will liberate us from our bodies.

That's really what Martine is saying, that we'll live forever, that our minds will be replicated, they'll be up in the cloud. We'll never die. If we go insane, if we get Alzheimer's , somehow our essence,

our kernel will remain. We're told that AI will liberate us from work because we'll have these smart machines which will do our labor for us. They'll drive our cars. They'll do our medicine. We'll go to an artificial intelligent agent, an algorithm to be taught. Perhaps one benefit is we'll avoid lawyers, but apart from lawyers, I worry deeply about the impact of AI on the expert professions of the 19th and 20th century.

00:39:00 T he very professions that represent the backbone of our economy. The problem with this promise is it's not being thought through. It's being thought through philosophically, idealistically. We're not thinking about it in the context of the real world. We're not thinking about it in the way in which these technologies -- and we've already had a great deal of example of this during the digital revolution, that in spite of all the great promise of the Internet, and I've written extensively about this -- in spite of all its great promise of democracy and egalitarianism and opportunity, it's actually created new elites. The promise of artificial intelligence, these grand philosophical frameworks, they forget about the realities. Jaron already talked about the art -- the AI of language being owned and peo ple of labor being appropriated. Who is going to own these platforms ? Will it be Google ?

00:40:00 We all love Google, of course, in some ways. And yet, Google is the largest, the most valuable, the most powerful company in the world. Google are the ones who just made the breakthrough when it comes to this artific ial agent being able to play Go. Google is pouring billions of dollars into AI, as is Facebook, as is Amazon, as are the other giants of Silicon Valley. Do you trust these guys to benefit mankind? Do you think they care about us ? These aren't bad people or bad companies, but they're focused on profit.

They're focused on monopoly. They're focused on owning that technology. We haven't t hought this stuff through. The promise is scary. The promise is that the technology is moving way faster than we are -- politically, culturally, existentially. We're not ready for this yet.

00:40:59 We're not ready to replicate ourselves, whether it 's in 50 or 100 years. We haven't thought it through. We can't imagine the cultural, the social, the economic impact. We haven't thought about AI and the way it will destroy labor, destroy jobs. What are we going to do when we have these machines that do everything for us ? How are we going to create value ? How are we going to feed and clothe ourselves in a world where all this technology will be owned by increasingly monopolistic companies ? So, we can't, at themoment, trust AI. We can't trust its p

romise. One day, perhaps. But there are a lot of problems still to be resolved. And for the moment, I would strongly argue that you should not trust the promise of AI. Thank you.



[applause]



John Donvan:

Thank you, Andrew Keen. And that is our moti on: Don't Trust the Promise of Artificial Intelligence.

00:42:00 And here to make his opening statement against the motion, James Hughes. He's executive director of the Institute for Ethics and Emerging Technologies and author of the book "Citizen Cyborg." Ladies and gentlemen, James Hughes.



[applause]



James Hughes:

So, it's been very useful. We've narrowed down to, I think, what is an intelligible question, because obviously, Martine and I don't want to defend the most extreme hyperbole in this debate. But there is an ideological question around promise. And it's an ideological question that I would frame as the promise of human intelligence itself.

And that is that artificial intelligence is a crystallization, a condensation, a manifestation of human intelligence, an extension of our capacity. And the promise is the same promise that we have been suggesting since the Enlightenment, that human beings, by taking control, by understanding how we think, and understanding the world -- that we can take control of the natural, and political, and social circumstances that determine our affairs and make a better world.

00:43:08 Now, the anarchist philosopher John Zerzan , I think, has the most trenchant critique of this. He thinks we went off the rails when we invented symbolic thought. As soon as we began to have lan guage, we were downloading the contents of our minds onto external storage media and uploading them again through our eyes -- we became cyborgs. That was the beginning of the end of the human project. I would argue that if you share my intuition that it' s better to be alive in the 21st century than to have been one of our hunter -gatherer or peasant ancestors, an intuition that it's better not to have a gutful of intestinal worms, better to live to 80, better to have literacy, better to be less likely to b e killed in a violent death, less likely to live in a life of slavery and oppression --

there is something about modernity, and this promise of the Enlightenment, that I think we should explore.

00:44:02 And artificial intelligence, I would argue -- I have a slightly different definition that Martine does. I would argue that artificial intelligence has been with us for a while.

Artificial intelligence can be reframed as the codification of the way that we do things together. Civilization is a form of artificial intelligence. It's what allows us to build cathedrals, and aqueducts, and banking systems. It's what allowed us to create laws and universities, and standard operating procedures. And today, that crystallization is taking the form of workers figuring out how to make things faster, and better, and smarter. And yes, change in the nature of work and ways that we should discuss.

Today, artificial intelligence is applied to our genomes, to our healthcare systems, to trying to figure out how to diagnose and treat disease in ways that yes, no one physician or nurse will ever be able to comprehend, that we will be able to put in the hands of every healthcare provider and in the hands of ourselves.

00:44:59 Those tools which will allow us the empowerment of understanding our own bodies.

And this AI today is allowing us to understand the ecosystem, allowing us to understand the myriad of consequences that we are wrought -- that we have wrought on the ecosyste m, to mitigate, to predict, and to reverse those trends. It is also armies figuring out who to kill and how to kill the most efficiently. It is also advertisers and totalitarian governments figuring out how to suppress dissent and how to manipulate opini ons. It is stockbrokers trying to figure out the best ways to exploit workers and accumulate wealth into the hands of the .1%. So, reason and technology can be applied to all these different ends. And the question of the difference between one end and t he other is a value difference. It is a difference that comes out of the values, as I said, of the Enlightenment. If we fight for free and equal societies in the future, the applications of technology -- including artificial intelligence, will be applied in free and equal ways.

00:46:00 But our decisions to be pessimistic about artificial intelligence will have no effect on the application by China or North Korea, or other authoritarian regimes. It is our own embrace in liberal democracy of these pow erful tools, making our society as strong and as effective as possible that will determine its future. So, future AI will allow us to understand the complexity of the genome, unlock health and longevity for our children.

It will not determine whether the re's universal access to health care. That is on us.

Future AI will allow us to displace routine labor and make possible abundance and leisure for all. But it will not tax the rich. It will not determine if we create a safety net,

and universal basic income so that we can all benefit from that universal abundance.

That is on us. Future AI will allow us to make better collective decisions, to understand the consequences of our actions. But it will not determine whether we have a totalitarian governmen t or a democratic one.

00:47:02 That is on us. Focusing on AI as either a panacea or a cause of social ills is a distraction from the political project that will allow us to use AI and all technologies for good versus ill. It's the flip side. Technol ogy itself does not determine these outcomes. We need to focus on creating the political contexts so that these powerful tools, the productive tools that we've been working on for tens of thousands of years, will be applied in the best possible way. So, I urge you to vote against this proposition, the proposition that we not trust the promise of AI, because if you do vote for it, you are voting against the promise of human intelligence itself. Thank you.



[applause]



John Donva n:

Thank you, James Hughes. And that concludes Round 1 of this Intelligence Squared U.S.

debate, where our motion is Don't Trust the Promise of Artificial Intelligence.

00:48:00 Now we move on to Round 2. And in Round 2, the debaters address one ano ther in turn and they take questions from me and from you, our live audience here in New York, at the 92nd Street Y. The motion is Don't Trust the Promise of Artificial Intelligence.

We've heard Jaron Lanier and Andrew Keen argue that "don't trust" side. They're saying that the central debate of the 21st century will be this one -- that they are not against the technology itself, but they are against the promise, as it has been laid out. They just say that the implications of this world of AI are not be ing thought through, that -- it's less about the technology, actually, than about a belief system -- and a self -deluding one, they say, at that. They also point out that most of us -- or those of us who are in the expert classes are very likely to be out of a job, that the whole issue has been oversold. The threat to expert professions is real, and that the promise itself is actually scary.

00:49:00 The team arguing against the motion -- and that is to say, they are arguing to trust it --

Martine Roth blatt and James Hughes -- they are actually describing artificial intelligence,

when all is said and done, really, as no more than a set of tools -- tools equal to other tools that we have used throughout history. They say that the promise of artificial i ntelligence itself is obvious. It's the promise that the human being can create technology that can make a better world, not a new story. They laid out the vision of replicating a human mind, not in its structure, but in its function. And they say that an artificial intelligence -- if managed by us, and the choice to manage is ours, has the potential to be good, to be useful, to navigate, to cure -- especially to cure diseases ofthe minds, and that it can evolve in an organic and friendly way. So, what

we have here is obviously already a discussion, and quite a complex one that is as much about physics as it about metaphysics, and as much about technology as it is about philosophy.

00:49:59 And we're going to cut through -- cut some of the arguments that have been made into smaller pieces and present them again to take some of what each debater has said to the other side. And I want to take first to Jaron Lanier -- your opponents have said,

quote unquote, "We will love the artificial intelligences in our lives." A really powerful assertion of their bottom line that what's there has great promise. It's going to do great good. We're going to become comfo rtable. We're going to become familiar. They will not be alien. They will be part -- something that is part of our world such that we are glad they are there and they are not unnatural. Will we love the artificial intelligences ?

Or at least can you respond to your opponents ’ assertion to that point?



Jaron Lanier:

You know, people are -- we're social. We want to be decent and if we're presented with an artificial character, even the ones that exist today, the Siri or the Cortana, we'll be deferent to them. We'll give them a shot. We find it funny. They're cute. And it can be harmless. If you take it too seriously the problem is that you kind of lower yourself to make the computer seem smarter and we see zillions of examples of this.

00:51:07 A big one going on now in education i s teaching to the test because the -- you know, you teach to make yourself look good to the algorithm rather than the actual teaching and the algorithm doesn't quite capture it, and that's just one example of many. I'd like to respond in sort of an unusua l way. If you're interacting with Siri or Cortana you might think well, I'm not that different. I'm just a more complicated version of a Siri, in virtual reality this other technology you might be familiar with, I've had exactly the opposite experience. When you're in VR you can turn into some weird creature. You can turn into some weird crab or something and the whole world can change and yet you remain there hanging and I feel as if you notice that your consciousness is this real thing. And I find th at experience just amaz ing. When everything becomes mutable suddenly you realize wow, there is some consciousness thing that's not j ust mechanism, at least to me, and I understand this is a personal belief and we can argue ourselves into the ground round and round about this thing.

00:52:01 But what I want to say is that if what technologists are doing is telling people hey, you're not so special, our machines are just like you, we shouldn't be surprised if people then respond by saying well, we don't trust your medicine. We don't trust your modernity ,and I see kind o

f a unified backlash against arrogance in everything from the anti -

vaccine movement to fundamentalist --



John Donvan:

All right. You made a lot of points. Let me let Martine Rothblatt --



Jaron Lanier :

Sorry.



John Donvan:

No. That's fine. It's my job to jump in a little bit early , which I'll do, but let's let Martine Rothblatt respond to the point.



Martine Rothblatt:

Thanks. I have no doubt that we will love o ur AIs just as much as we love our pets,

which we also create and end up making themselves into a relationship with us. I could not disagree more with Andrew. I mean, I just imagine that he's creating AIs , this fearful creature that will take our jobs away and it reminds me of somebody standing up in the middle of the 15th Century and saying you know, don't trust the promise of the printing press.

00:53:01 The scribes shall have nothing to inscribe.



[laughter]



And instead -- and you could paint this whole same parade of horribles, you know, the church will control all the printing presses. Instead we've had a [unintelligible ] of cognization. We have a new type of person called a bibliophile that loves books and we will have AI philes that will love our AIs. This is, as James said, an enormously empowering and liberating force of AI, and what's most important is that we the people make sure that access to AI is available to everyone and since it will be the hackers,

millions of them, d ispersed throughout the world that create this AI out of open source software, I think there is nothing to fear and I would argue against the motion.



John Donvan:

All right. Andrew Keen -- was your position fairly charact erized there?



Andrew Keen:

Sorry. Say that again.



John Donvan:Was -- were you fairly characterized there ? Are you against printing presses ? Or would you have been ?

00:54:00 [laughter]



Andrew Keen:

As a writer absolutely.



[laughter]



So, yeah, this is a typical kind of argument that we always fall into in these kinds of debates. Jaron and I will say we have to worry about this stuff. We have to worry that it's not doing what it says it's going to do. And then someone like Martine will come along and say oh, you're just whiners, pessimists. Just look at history. Just look at the history, for example, of the printing press. I don't know how many jobs, by the way, the invention of the printing press caused and one of the consequences , of course, of the printing press was the reformation and the Hundred Years War and a lot of other kind of suffering. So, it's a rather muddy consequence firstly.



[laughter]



That's the first problem. The second problem is more substantial. Because he's making an evangelical spiritual argument. His argument is hopeful. He's saying well, in the past it's always worked out okay, so it will work out again in the future.

00:55:04 And he's absolutely wrong when it comes to technology and jobs. The fact is most major economies, most researchers, most people who spend their living figuring out what we're going to do in this world cannot figure out what people are going to do. He talks about cuddly AIs. That's not a job. There's no labor there. He talks about open source , absolutely nonsense. Open source technology has had no successful in today's digital world. The four largest companies in the world today, the four most valuable companies in the world today are Facebook, Google, Microsoft, and A pple.



[laughter]



And these four companies are the owners of the platforms of our networked age. Where's open source at ? Open source is just another ideological dream. It never happens and it -- I doubt it will ever happen.

00:56:00 John Donvan:



All right. Let me bring in James Hughes. You can --



[applause]



-- you can vent. You can let that go. Let me bring in James Hughes to respond to some of what your opponent just said.



James Hughes:

Sure. Well, see, I believe in technological unemployme nt. It's actually -- I've been trying to make the argument for a decade now that -- of the inevitability of technological unemployment and that we need to start anticipating it. It's actually a hard argument to make right now because employments beginning to pick up again. But I believe in this inevitability. But I'm one of those lefty folks who I've heard the last 200 years think that eventually freeing us all from wage slavery might be a good idea. And that if, in fact,

people started to see that -- the inevitability of the elimination of work, that we might all wrap our minds around the concept that there might be something better to do than all have wage slave jobs.



[applause]



Just as we wrapped our mind around social security and Medicare and Med icaid and the British National Health Service and so forth, the progress of social welfare legislation is,

yes, that bad stuff happens like the industrial revolution and then we responded.

00:57:06 So I think that we have to sort of imagine that we'll be able to do that again because people are already talking about what the necessary response is to technological unemployment should be.



John Donvan:

Okay, your opponent, Jaron, put forth the idea of translators, professional translators,

people who know multiple languages, that gradually machine translation, while stealing from their body of work to build their algorithms and f eed their algorithms, will be put out of work eventually. And are you saying that, that's one of those jobs that you would like t o see go away?



James Hughes:

Like to see go away ? I don't -- I think there is probably no definition of futility of it other than -- no worse definition than knowing that the machine next to you could do the job that you're doing faster, better, and safe r, but that you're forced to do it because somebody says, "You can't use that machine." So, yes, I think in the future there will beall kinds of machines that do all the jobs.

I basically -- I think that everything that we do will eventually be done bet ter and faster and safer by machines.

00:58:00 But the things that are most immune right now are the creative jobs. So, you know, my daughter's an opera singer. She's probably relatively immune. I -- you know, if opera could be replaced, it would ha ve been replaced by record players and radio a long time ago. So I think, ye s. I think there will be many things that will be replaced. And what it will do is free us up like Andrew proposed to the Marxian vision that we'll be able to be farme rs in the morning and poets in the afternoon if that's what we want to do, a life of true choice. That's the vision of the future I have.



John Donvan:

Okay. Jaron Lanier.



Jaron Lanier:

So, look, there are two problems with this. First off, there's a pheno menon I call,

"premature mystery reduction," which is when we pretend we have something working that we really don't. So right now, most AI actually depends on scooping up things that people do, including the Go program today which is looking at people's Go games, and the machine translation example. Now -- we want to pretend that there's this AI behind the curtain that's freestanding. But, actually, there's millions of people there, too. Now,

the problem with saying, "Well, we'll ju st pay everybody a basic income and then pretend they're not valuable," when they are is -- it's -- first of all, it's a lie. Secondly, it -

-

00:59:01 John Donvan:

Why ? Why is it a lie?



Jaron Lanier:

It's a lie because they're still needed. I mean, this is like crazy.



John Donvan:

But he's saying that they're not needed. He's saying the opposite.



Jaron Lanier:

Well, he's wrong, technically. I'm sorry I have to pull that on you. The truth is that in order to make machine language work, you have to scrape millions and millions of real translations every single day to keep up with current events and frameworks. So what --

so you're factually wrong to say that they're not needed. Now, if someday -- you could say, "Well, somebody maybe they wo n't be needed." But, you know, the problem ofthat is this premature mystery reduction.

I mean, yeah, someday we might be able to float because we understand gravity better o r something. I am an optimist. I don't want to believe that we'll always be stuck with our current level of knowledge. So I'm --

in a sense, I'm a transhumanist. But I just believe that it's a fundamental unknown what the timeline is. That's what science is. Science is what we don't know. So we don't know. A nd so we shouldn't pretend we know. That's lying to ourselves. It's undignified. It's petty. It's silly. It's childish. But also -- I'm sorry, but -- you know, but --



[laughter]



-- the other -- the other --



James Hughes:

But there are things --



J aron Lanier:

But, but, but, but, but, but, but no, no, no. There's a peril.

00:59:59 T here's also great peril in what you're suggesting. If we say, "Instead of taking the dignified path and admitting that people are still needed, if we pretend they're not needed and then we have some agency that doles out basic income to them as if they weren't needed, you really think that thing's not going to be a magnet for corruption?

Every history lesson teaches us that, that's a huge peril.



J ohn Donvan:

Okay, I'm going to break in there.



Jaron Lanier:

Don't step into that problem.



John Donvan:

I'm going to break into there to let Martine come in.



[applause]



Martine Rothblatt:

[unintelligible] we’re down talking about facts and figures here, because I don't think that the facts support the idea that new technological innovation, such as AI, result in mass unemployment. In fact, there are today, with our global population of almost 8 billion people, vastly more pe ople employed doing things than there ever were in the annals of history. You could imagine that before the turn of the 20th century, back in the 1900s, over 95 percent of all people were on farms.

01:01:00 And somebody would make a very sensible argu ment, saying, "We can't allow in tractors and things that steal our techniques for harvesting vegetables and food,

because then all these farmers would be out of work." Instead, now we have less than 10 percent of the global population raising food. The average nutritional content of everybody, on average, is much higher. And there's vastly more people doing vastly more interesting jobs and passions than ever before. We are an intelligent, creative species. It is in our DNA to solve problems . We wouldn't be here today if we weren't super good problem solvers. In the past, we were farmers, and then carpentry. Now,

coding is the new carpentry. We will figure out new and amazing things to do.



John Donvan:

I want Andrew Keen -- Andrew Keen. Now, Andrew, you're also anti -tractor, as opposed to -- and anti -printing press. What about them? You are hearing -- you accused your opponents of optimism, and they certainly do sound optimistic. What's wrong with that?



Andrew Keen:

No I [unintelligible].

01:02:00 Look, there's nothing wrong with -- John, there's nothing wrong with optimism. I mean,

as long as you're optimistic about -- and you're realistic. And by the way, Jaron , I didn't know you were transhumanist. If I'd have known, I wouldn't be on your team.



[laughter]



Jaron Lanier

I meant, you know, within the context of how unknown it all is, you know?



Andrew Keen:

But the unknown unknown. Right? That's transhumanism.



Jaron Lanier

Yeah . I’m a Rumsfeldian transhumani st.



[laughter]



Andrew Keen:

[unintelligible] now. And by the way, I have no idea --[speaking simultaneously]





Andrew Keen:

-- what our te am -- our t eams seem split . On the one han d, we have one guy saying that he's celebrating the elimination of work. And then Martine is saying that actually, we're going to innovate so much that everyone will have new jobs. So, you guys have got to make up your mind, whether you're for or against --



John Donvan:

Well, we'll stop you right there.



Andrew Keen:

-- jobs.



[laughter]



John Donvan:

Has he pointed to a contradiction in your t eam, James?



James Hughes:

I -- well, it partly depends on what a job is in the future. You know, I'm -- wha t I want to see eliminated is wage slavery. I think we'll all have occupations. Aristocrats who didn't have to work in the past all had occupations of one sort of another that kept them from,

you know, putting guns in their mouths.

01:03:00 And I think that we will find out those too.



[laughter]



John Donvan:

But they had lots of serfs taking care of things for them.



James Hughes:

And we will have artificial intelligence taking care of us.



John Donvan:

Jaron.



Andrew Keen:

But can --



John Donvan:All right. Andrew --





Andrew Keen:

Can I just say one thing? That -- we're sort of talking -- in these kind of debates,

everyone's throwing, you know -- we know this from the presidential debates --

everyone's claiming the facts are on their side. But the reality is is when you look at the numbers, when you look at the serious research, you will find that the vast majority of economists are deeply worried with this, Martine. What -- give me some example s of economists who say, "Yeah, there's going to be millions of jobs in the future. We don't need to worry about this?"



Martine Rothblatt:

You know, I'm reminded of the famous quotation from Arthur C. Clarke , who said that if you ask a lot of experts in their field whether or not something is possible, and they say no, they're almost certainly wrong. And it's the same thing about the economists. They have been wrong repeatedly throughout history. So, going to a source like e conomists -

- I would look at the bare fac ts. We've got eight billion people in the world, fewer people starving than when I was growing up.

01:04:05 Vast ly more people employed. T hose are the bare facts. And furthermore --

furthermore , let me just say one more thing. We're just at the beginning of what we can do as a human species. We've got things like electric cars from Tesla that just 10 years ago, all the economists, and the technocrats, and the bureaucrats dismissed as impossible. Now people are saying, "Hey, soon the whole economy is going to be electric." If you are a person who loves life, you will never run out of cool things to do.



John Donvan:

I'll get Jaron in there.



Jaron Lanier:

Martine, I want you to be right. And I think you probably are, in the sense that there will be new things for people to do. All we have to admit is that people need to do them and that they could be paid for it, and then we can still have dignity. The only danger is not so much that people become obsolete. The danger is that we'll pretend they're obsolete.



Martine Rothblatt:

Absolutely. And already, the Europeans are leading the way to --



J aron Lanier :

So, you agree --Martine Rothblatt:



-- guaranteed annual income --



J aron Lanier :

You know, the [unintelligible] is screwed here.



Male Speaker:

We do. We totally a gree --

01:05:00 J aron Lanier :

You have like -- [laughter] --



Martine Rothblatt:

There will be social dividends that -- it's just like Bernie Sanders says. Medicare for all.

Social Security for all. So, there will be a basic social dividend that's paid for everybody.

We can afford it. Let's do it.



John Donvan:

James Hughes, one of the strongest statements your opponents made is that we are not ready to replicate ourselves. Take that question on.



James Hughes:

Well, I do think that there's a great deal of mystification about the nature of artificial intelligence in the future. Martine has a fairly embodied notion of what artificial intelligence would be. Mine i s much more diffuse. I think that the potential space of what artificial intelligence will be in the future is , it's very difficult for us to imagine. So far we have a lot of anthropomorphic projection onto that state. Those who expect that something is going to jump out of a box is going to take over the world, the world has 10

seconds to defeat it -- I mean, that's an anthropomorphic projection of an adolescent,

you know, male fantasy onto, "Well, if I w as king of the world." [laughs]

01:05:59 And I think, you kno w, it could be like -- a n artificial intelligence might just want to communicate with the -- with whales, or with the stock market, or with the stars. And grow like moss on the side of a mountain. I mean, we just have no idea what artificial minds are going to be like. So, yes. I think, if the project is to replicate a human mind and put human mind emulation into a machine, that's one kind of project. I f, on the other hand, self- awareness emerges out of the Internet or just through all the communications and information in the world, we have no idea what it's going to be like.John Donvan:



Let --



Martine Rothblatt:

I agree with James, because, you know, contrary to what Andrew said, the mass of people throughout the world are the ones who are creating our Internet. Websites were not created all by Google, and Amazon, and Apple. They we re created by literally millions of people creating the [unintelligible] --



J aron Lanier :

But then these things are appr opriated by --



Andrew Keen:

All right. Let's use -- okay. Martine, let's use --



John Donvan:

Okay. [unintelligible] Andrew Keen. Andrew Keen.



[speaking simultaneously]



Andrew Keen:

-- the example of Google. Google has created -- you know, J aron has defined artificial intelligence, I think, very intelligently, as the algorithm. Google owns the most valuable algorithm in the world.

01:07:00 There's no secret sauce to that a lgorithm. It's not as if Larry Page and Sergey Brin at Stanford suddenly figured out, "We're going to create this remarkably intelligent algorithm." And --



Martine Rothblatt:

[inaudible] --



Andrew Keen:

Well, let me finish. No, l et me -- let me --



John Donvan:

[unintelligible] --



[speaking simultaneously]



Martine Rothblatt:

You let him hold the most valuable --John Donvan:



Do let him finish.



Martine Rothblatt:

-- operating system in the world, and they [unintelligible] --



John Donvan:

Martine, let him -- Martine.



Martine Rothblatt:

-- that --



John Donvan:

Let him finish, please.



Martine Rothblatt:

-- Google is today.



John Donvan:

I'm going to give you your shot.



Andrew Keen:

So -- and this is just to reiterate what Jar on had said. That algorithm is a collection of our intelligence. Google essentially has aggregated the entire intelligence -- brilliant,

brilliant maneuver. Fantastic. I'm not saying it's immoral in any way. But Google now owns our collectiv e intelligence. It's a company now that's worth -- it probably w ill be the first company that's worth a trillion dollars. Where do they pay it back to us, our intelligence, that revenue? And that's Jaron's point. When you tear back the curtain,

it's n ot arti ficial intelligence.

01:08:00 It’s us. And we're not benefiting.



John Donvan:

Martine.



Martine Rothblatt:

Andrew, Google does not have --



[applause]Google does not have the intelligence of one 9

-year -old girl, okay? All they've got is a bunch of data that's been hoovered up, along with dozens of other companies who have hoovered up that same data. What Google thinks is magical and is special today -- 20 years from now, will be passé and unnecessary.



Andrew Keen:

So, why is Google such a valuable company?



Martine Rothblatt:

The same reason that IBM was 50 years ago. They had a point in time when they provided a valuable service to the market. IBM was more valuable relative to the economy 40 years ago than Go ogle is today. Now it's kind of irrelevant. Google will eventually become irrelevant.



James Hughes:

I just think it's fascinating that you reiterated my point --



John Donvan:

James Hughes --



James Hughes:

-- which is that artificial intelligence is, in fact, a crystallization, a formalization of collective human intelligence. And therefore, not to trust the promise of artificial intelligence is to dismiss the promise of collective --

01:09:00 John Donvan:

Okay.



James Hughes:

-- human intelligence.



John Donvan:

So --



James Hughes:

Yes, I don't think it should be in private hands. But that's a different question. So, you know --



John Donvan:

Jaron, hang on just one second.



[applause]Wait just one second.

Just one second. I'm going to let Jaron answer. But after his answer, I want to start going to audience questions. How this will work is if you will raise your hand, I will call on you. We would like it if you would stand, tell -- Jaron is going to s peak first. Stand, tell us your name. Wait for the mic so that you can be heard on the podcast. And then as k a question that is a question that takes less than 30 seconds. If you're a member of the media or a blogger, we would appreciate it if you wou ld identify yourselves. Let me let Jaron respond to that point.



Jaron Lanier:

So, we have the latitude to perceive agency in others or not. We have the latitude to perceive consciousness or not. We have the latitude to perceive God in the world or not. We are free to perceive different things. There's no consciousness meter that will tell you if somebody else is an automaton or not. Some of the people might feel -- seem to be so on occasions, and I -- you know, I don't know.

01:10:00 I mean, I ca n't tell what's inside somebody else's heart. We love each other on faith.

Just like we know God, we can't really know. We can only know our own consciousness.

And so, you absolutely have the latitude to perceive machines that way. The question is whe ther it's useful, smart -- I mean, my argument against it can never be to challenge your faith, or your idea, or your aesthetics, as I said before. I do have a pragmatic argument on a society level, though. And the thing is, tech companies -- which I'm t o tally in be d with. They're not like some other. That's me. But the t ech companies are so powerful that we're basically like de facto governments now of the world and so what we need to do is insist, as I said before, on a church -state separa tion.

01:11:10 What we have to do is insist that these massive powers don't adopt particular faiths about what's conscious or not because that totally screws you up. As soon as -- all of the sudden you're o wning women’s bodies because you believe that a fetus --



John Donvan:

I don't think your opponents disagree with that. Do you?



M artine Rothblatt :

No we don't.



James Hughes :

Bioethics, we have to determine what a person is. We have to have an ideology,

otherwise you treat a stone the same way you treat a human being. One thing hasconsciousness, the other doesn't.

I t is an empirical question that we will have to solve in the future.



Male Speaker:

It's --



James Hughes :

What kinds of artificial intelligence have moral standing and which kinds don't?



Jaron Lanier :

You know, we must give people latitude to have different beliefs about this. We must not impose beliefs on each other about this .



John Donvan:

But you all agree.



Jaron Lanier :

If you feel that they must be imposed then we have an extremely sharp disagreement that cuts to the core --



James Hughes :

You think there should be no law determining who's alive and who's dead?



Jaron Lanier :

No, no. Th ere prec isely can't --

01:12:00 Look, we have to give people the latitude to decide that a fetus can be aborted. We have to give people the latitude to decide that they treat their dog like a child. We have to give people the latitude to fall in love with their computer. We have to give people the latitude to think that their computer should be dumped in the East River.



John Donvan:

Okay.



[talking simultaneously]



J aron Lanier :

If we impose these beliefs on each other --



[talking simultaneously]



James Hughes :You also have to figure out legally whether your lab tech can be put in prison for not fixing your computer correctly, you know, whether that's medical malpractice or not.

So, we're going to have to figure these questions out. It will be a matter of law and public policy.



John Donvan:

All right. Let's go to some questions from the audience, please. Right down in the front. Third row. Thank you. Can you tell us your name, please?



Male Speaker:

Yeah, my name is Nick Hill. So we've been talking a lot about artificial intelligence as it could replicat e the human mind, and in the past when technology has been created we've been able to sort of adapt to it because humans have always been the smartest being on earth. The goal of artificial intelligence is to create a system that could actually be smarter than a human being.

01:13:00 So, I'm wondering when that happens how do we make sure that human beings are one,

on the good side of it, and two, what's really preventing it from --



John Donvan:

You're saying that time it would be different. That would be a different -- a game -

changing technological development.



Male Speaker:

That development when the computer is smarter than a human being and can replicate itself.



John Donvan:

Let's take it to Martine.



Martine Rothblatt:

So my view is that there's not a -- it's a false dichotomy between an artificial intelligence that's smarter than a human being and a human being. I think that for an artificial intelligence to be smart and intelligent it needs to be as human as we are and therefore there's a continuum of consciousness between humans and artificial intelligence. The dichotomy is false.



John Donvan:

Let's go to Andrew Keen.



Andrew Keen:

Jaron, would you like to respond?Jaron Lanier:



Yeah, I'd like to. I don't think this is new. It's happened before and the example I'd like to give you is Adam Smith's invisible hand, that markets can do things that people don't seem to be able to with planning and the thing about markets is I'm very much a Keynsian . We have to be able to use them as tools. We have to treat them as technology not as religion.

01:14:00 As soon as we treat the market principle as religion then we actually screw up markets even, so sorry Chicago school, but that just seems to happen. And so the wise thing to do is to recognize that we can build these things. We have and yet we build them in a way that we can use them as technology, not as religions, and we can use them well,

just like markets.



John Donvan:

Right down in front here.



[applause]



Male Speaker:

My name is Ahmad [spelled phonetically] . I'm a programmer. I'm not at the sa me level as the goat herder /AI researcher, but I'd like to ask one question to the panel because I feel that I agree with Jaron here. What do you think AI is? What do you think it's accomplishing? I mean, as a programmer what do you -- why would I want to build consciousness into a machine? Isn't it that technology throughout history has been automating structured t asks that need no intelligence, something that's repetitive that we can make it over and over again, and if we get away from the Phillip K. Dick s ci- fi -ish

thing there is really no need to have consciousness into a machine. So what --



John Donvan:

Okay. I think the question --

01:14:59 Male Speaker:

What is AI? [unintelligible] I'd love to, you know, yeah --



James Hughes :

Well, this goes back to the previous question. Is AI something outside of ourselves or is it an exten sion, an intelligence augmentation of us? And I don't think there's a market plan for toasters that don't want to give you toast because it has something better to do that morning. Or a bomb that doesn't want to bomb because it likes the person thatyou'

re trying to bomb. You know, there's no market plan for truly self -aware, self -

determined robotics, except for those who perhaps want to continue their own consciousness in an emulation of their brain. That's a particular market plan. Other than that I don't see a great market plan for that. So, for the rest of it it's an extension of human intelligence into the world and therefore I call that intelligence augmentation as opposed to artificial intelligence.



John Donvan:

Andrew Keen, do you want to tak e the question as well?



Andrew Keen:

Yeah. I mean, I'd rather -- I think Jaron should define artificial intelligence. He's probably the most suitable of everyone. He gets it more than anyone.



Jaron Lanier:

Well, I mean, AI is more than anything else a funding category for research.



[laughter]



No, I mean, it's like --



[applause]

01:16:00 -- and it incorporates a wide range of disciplines and pursuits that might or might not have been bundled together and they were bundled by historical accident in many cases. AI steps on its own foot periodically. What happens is there's a kind of a crazy wild -eyed -ness of like, "Oh, we're about to understand how to replicate a person," and then the funders are like, "Boy, you sure didn't deliver." And then there's this thing called an AI winter that keeps on happening. And then you watch your grad students having their careers ruined. And then it finally happens again. But -- like we step on our own feet constantly with this fantasy life. And if we could only just be good engineers and scientists, we would free ourselves from this burden of just constant self-

destruction.



John Donvan:

Okay, I want to -- Andrew, I think you might be regretting passing that.



[lau ghter]



Andrew Keen:

Is the -- are these AI winters getting shorter, and the summers are longer? And at some point is it going to become real? Because my experience in Silicon Valley as anentrepreneur and as a technologist, there --

it's always a questi on of timing. So Jaron's right about AI being a category for researchers. But now it's a category for entrepreneurs. You wander i nto a VR and you say, you know, "I've got AI for this, AI for that, and I'll write you a check."



20: 18:22



Jaron Lanier:

Lately the VP will say, "Oh, I spent all my money on VR startups."



John Donvan:

All right, let's go to another question.



Male Speaker:

But are we getting closer to the reality?



Jaron Lanier:

Well, it doesn't mean anything. There's no definition. I mean, that's the thing I'm trying to say is that we have a field that's lost its moorings to fantasy. So we don't know. I mean, I very strongly believe in research and coordinated systems and algorithms. I strongly believe t hat these things are improving the world and that they are essential. I strongly -- you know, I -- the actual work is great. But the fantasy life -- just --



John Donvan:

I want to -- I want to break in, Jaron, because Martine has direct experience in c reating this artificial intelligence cybord based on your wife. Tell us what AI is for you?



Martine Rothblatt:

So I think it was a great question that the gentleman from the audience just asked in terms of like, "What is the need or the desire for AI?" And my response is that there is nothing probably that people value more than other people. And we have families and children for that reason.

01:18:03 What we value most about other people is probably their mind, if you want to call it their soul, their spirit, their comradery. And so there is going to be an irresistible pressure for lots of people to try to create artificial people. That's what humans do. We like houses, we make artificial houses. We like to go fast, we make artificial horses. We l o ve minds, we make artificial minds. So this is the kind of irresistible pressure I think that will end up bringing artificial intelligence forward. And as James and Jaron and Andrew have said, create a touch point for solving the ethical issues about what we do with this AI, who controls the computing substrate and the software substrate for this AI, and where do we want it to go?

01:19:00 It's a societal issue , which is exactly why we need these type of debates and forum.



John Donvan:

Okay. Now, we've only had men from the audience so far and I only see -- oh, thank you. Down in the front row, please. It's coming from your left side.



Female Speaker:

So I think that point actually speaks directly to m y question. You brought up earlier that, as we go through this transition, I think every --



John Donvan:

Just so the podcast people know, you're addressing Martine.



Female Speaker:

Oh, yeah, to James.



John Donvan:

Oh, to James? I'm sorry. Okay.



Female Speaker:

So, to James, he brought up earlier as we go through this transition, I think actually everyone on the stage probably agrees that there is great eventual promise in the technology of AI. It's just, "How do we get to that promise?" And the team for the motion is saying there are some problems with that. And you brought up politics as --

and the political system as a way. And both of you, Martine and James, have spoken about human ingenuity.



John Donvan:

Okay, can --



Female Speaker:

So I just want --



John Donvan:

Thanks.

01:20:00 Female Speaker: ]

-- you to -- especially after the last two weeks, how do you get your optimism on that point?[laughter]





James Hughes:

Well.



Martine Rothblatt:

Yeah.



James Hughes:

Our ancestors did not par ticipate in the political process to the extent that we fantasize that they did. You know, there was a lot of backroom deals. It's fascinating to me that the Republicans are actually debating whether they should have a brokered convention and that there' s so much pushback from the leadership of the party because they realize that there would be a huge groundswell of opposition from the base. And that's because we have become a more democratic society. The smarter that we get, the more time and leisure t hat we have as a civilization, the more democracy becomes real. And what we're seeing is some of the unfortunate consequences of democracy becoming real. It's not -- it's not pretty, you know? It's not pretty in the Arab world. It's not pretty here. It's not pretty in a lot of places. But it is democracy maturing and going in the right direction.

01:21:09 John Donvan:

Another question? Right down [unintelligible] side.



Male Speaker:

[unintelligible] . So, the good life we have now is built on the destruction and enslavement of many, many people and many, many species . And the planet is in danger right now from our activity . Why do we think that the -- that we can handle this AI that's coming at a speed faster than anything we've ever done ? Why d o we really believe we have the moral and intellectual ability to do that, given our history ?



[applause]



John Donvan:

James Hughes.



Andrew Keen:

Because of the printing press, right?



James Hughes:Well, I think all technologies raise a similar questi

on. I mean, you're right that I am an accelerationist , as I do believe technologies are speeding up .

01:22:01 I don't think it raises a fundamentally different question than previous technologies did.

I mean, we could have said the same thing a bout the car, for instance. The consequences -- one of the distinctions we make in my field is between dual use technologies and single use technologies . Nuclear weapons, it's hard to defend as an ethical technology . But artificial intelligence is one w here we have to accept there are so many positive benefits that we have to take it on board and try to figure out how to mitigate the downsides.



John Donvan:

Okay . I want to remind you that we are in the Question and Answer section of this debate from In telligence Squared U.S., from Artificial Intelligence Squared U.S. I'm John Donvan, your moderator . We have four debaters -- two teams of two, debating this motion, Don't Trust the Promise of Artificial Intelligence . Let's go on to some more questions . Right there on the aisle . No, farther up on the aisle . Thanks.



Female Speaker:

Hi . This question is [unintelligible] --



John Donvan:

Can you tell us your name or at least first --



Female Speaker:

Oh, sure . Michelle Chevin [spelled phonetically] . I'm a senior analyst at Luminary Labs.

The question is primarily for Martine, but I'd love to hear everyone's take .

01:23:00 So, Bru no Latour talks about the central political questions of our time and of the near future as being around who is included in the society that we're building and the system that we're building . So, I'd like to problematize the idea that actually James brought up,

that maybe -- you know, we have to draw this distinction between conscious and unconscious beings . And is it maybe part of the problem that we have drawn so much distinction between levels of consciousness ? And I wonder -- you know, we could have said the same thing about the car, about agriculture . And maybe we should have. So,

isn't it ac tually irresponsible, as the gentleman just said, isn't it irresponsible to trust the promise that humans are actually capable of building a system that will deliver us from the --



John Donvan:Okay.





Female Speaker:

-- destruction that we've wrought on the planet?



John Donvan:

Let me -- let's bring that first to Jaron Lanier.



Jaron Lanier:

The great danger we face is within ourselves, we've gotten good enough at technology that most of our problems are brought by our own actions now .



20:25: 25

We are rarely attacked by giant asteroids, or dinosaurs, or anything . We are in our own troubles . And so, the solution to that has to be clarity of mind. We have to be able to see clearly what we're doing to ourselves in order to do something differently . So, the thing that I keep coming back to here is to do everything possible to not fall into fantasy .

And I -- it's odd, because it's a very technical sounding fantasy . And so, you think, "Well,

because it's got this technical flavor to it, it must be less fantastical than some of the other fantasies that we hear our political and social lives being drawn into all the time."

But it really isn't . I wish somehow there was a way for people to just see inside these algorithms . They're not all that fancy, honestly . I mean, they're cool, but they're not --

it's much more just the size of the data we can get these days, you know ?

01:25:03 Years and years ago, Marv in Minsky , who I mentioned, had helped some graduate students would make a language translator over a summer project . And it didn't work for decades and decades, until we could get enough samples from people in massive quantities . And that's where we still are . And it's fantasy that we've really uncorked the core of language and we understand how language works is not so . We have statistical correlations between what real people do . But we don't have any way beyond that yet.

We might , we might . But it's -- that's an unknown --



John Donvan:

Okay . Let me -- let me --



Jaron Lanier:

So we mustn't -- we mustn't hypnotize ourselves if we're to survive.



John Donvan:

Let me let Martine answer the original question.Martine Rothblatt:



Thanks . I agree with Jaron that we are at war with ourselves . And perha ps, as humans,

we always have been . But one of Eisenhower's best quotes was that no war was ever won by pessimists . And in order for us to win this war with ourselves, we too must be optimists .

01:26:03 And I believe that while we face towering probl ems -- global warming, starvation, and whatnot -- we've also made towering successes . We've raise more people out of poverty than ever before . We've been able to get literally billions of peoples through their elected representatives into a writte n agreement to reduce global warming. I mean , the a ct -- you talk about economics . They're getting billions of people to agree on anything . I mean, it's hard to get five people to agree on anything . So, I believe there is legitimate, pragmatic cause for optimism. And as James, and Andrew, and Jaron, and we have all said, the promise of artificial intelligence is not that it's going to be here tomorrow, in four years, or eight years .

01:27:00 This is a multi- decade process. And I do belie ve we have time to get it right .



John Donvan:

Andrew Keen, you have not -- your side has not reached for this sort of Hollywood version of the smart machines taking over, and making a direct enemy of humankind,

and killing humankind . But in fact, it's -- there are serious people outside of Hollywood scriptwriters who are coming up with those scenarios and saying they're real. Elon Musk, for example, said with “ artificial intelligence, we are summoning the demon ” --

meaning more than you 're talking about with jobs . Is that fantasy, Hollywood world off the table, as far as you're concerned? Is that so far out there as to be irrelevant?



Andrew Keen:

T here's a guy at Oxford University called Nick Bostrom, who has written an important book about this . And I think his book very much influenced Musk and a number of --

Cambridge University has a Center for Existential Study . It comes ba ck to the Rumsfeld

thing again . We don't know -- it's about knowing about unknowns .

01:28:01 At the moment, I don't buy -- it's a -- as you said, it's a Hollywood scenario . The idea that we're going to wake up in the next five or 10 years, and robots will not only be smart enough, but acquire their own consciousness, and have a -- w hat Mar x might call a species being -- I think, is absurd. But we don't know . And a lot of it depends on Mo ore's law . A lot of it depends on the -- sort of the runway of computational power .And the reality is it's not impossible

. But I think it's an unhealthy part of the debate,

because it's so speculative and so easy for Hollywood scriptwriters to take advantage of .

The real issu es are the ones that we're talking about today . The real issues are jobs,

what we're going to do, who owns all this, who owns all these algorithms, and how we,

as a species, benefit.



John Donvan:

James Hughes --



James Hughes:

Andrew, I just want to say --



John Donvan:

[unintelligible] --



James Hughes:

-- on this we actually agree . And I think that there is a path forward in how we prepare for the possibility that -- what I consider a currently remote possibility -- of catastrophic emergence of some kind of artificial intelligence .

01:29:08 And that would be precisely the path that we have explored around cyber theft, cyber security, having resilient information systems, being able to turn off pieces of the Internet -- that's if we need to . We have a path forward . And it's unfortunately the magical thinking, which I think you're correct about, in certain quarters, which leads people to say, "Oh, no, no . It's going to jump out of the box and in 10 seconds it's going to take over everything," and then send its nanobot minions to knock down your door .

That is fantasy . And that means that people don't pay attention to the political, and security, and criminal sanctions that we can put in place today to minimize that possibility.



John Donvan:

Jaron Lanier.



Jaron Lanier:

So, I feel as if I'm arguing against a great tide of Hollywood scripts, but I just have to repeat that this question of whether the technology that might destroy the planet would be alive, or conscious, or intelligent is actually irrelevant .

01:30:02 The only thing -- but such a technology could exist. Building a Skynet, as shown in the Terminator movies, wouldn't take a lot of code, honestly, at this point .[laughter]





Like, you know, I think we could do it in a week workshop or something . You know, you --



[laughter]



-- it's not that hard to see how you'd make a self- replicating drone at this point. I think we can print them . We can put guns on them . I mean, like, honestly, don't do it, you know? But --



[laughter]



-- the thing is that's the point. Like, the distinction between some smart machine that hurts you and some malfunctioning machine that you can't operate that hurts you is a nil distinction . They're just both ba dly engineered. As long as we are beholden to this mythology of the evil machine, we will not be able to see clear to design well-

functioning machines that we take responsibility for and use well . And I don't know how to shake you from this hypnosis . You've seen these movies and you keep on going back into these thought patterns . Please, please, consider that this isn't a real thing .

01:30:59 Please consider that the difference between one machine that works terribly and another machine that works terribly is not important . They just both work terribly. An important difference is between those and a machine that works well . That's what you should focus on. Can't you see that? I mean, I --



[laughter]



Martine Rothblatt:

And --



Jaron Lanier:

-- I love these Hollywood movies . I love them. They're great . But don't live in them.



James Hughes :

-- Jaron, I think you're right. I mean, it doesn't make any difference morally if a human being in human civilization or a robot does, but they're equal catastrophes.



John Donvan:But when we get to the point where the debaters are agreeing with each other, we're in trouble.

So, I'm going to say that concludes round two of this Intelligence Squared U.S.

debate.



[applause]



Where our motion is Don't Trust the Promise of Artificial Intelligence. Please remember how you voted before the arguing began, because we're going to have you vote again right after this brief closing round and again it's the numbers -- it's the team whose numbers have changed the most in percentage point terms who will be declared our winner. But first round three, closing statem ents. Each debater will speak uninterrupted in turn for two minutes each. The motion is Don't Trust the Promise of Artificial Intelligence. Here to summarize his position supporting the motion not to trust, Jaron Lanier, a computer scientist and author of the book, "Who Owns the Future?"

01:32:02 Jaron Lanier:

There haven't been too many times when you've been told by technical people that you must reduce your sense of autonomy in order to be hip and with it and intelligent and not left behind and not like, you know, in order to stay with the kids you must accept that you will have less autonomy in the future, that there will be these other things that you're not responsible for, even though they were made by people. Don't give up your sense of human responsibility. These machines are ultimately for the foreseeable future until unforeseen scientific breakthroughs rehashing the activities and the brilliance of people in new ways. If we recognize that properly we don't need to face technological unempl oyment, but if we fall into thi s mythology from all these beautiful screenplays and all these wonderful productions, if we accept the machines as being living things,

unfortunately there's a bit of a zero sum game. It means that we'll be saying oh, it's th e machine's fault. It was the machine that did this.

01:32:59 It was the machine that was good or evil. The moment you do that you're letting go of your responsibility as a person. The moment people let -- accept the notion that people have less resp onsibility, at that moment we start losing a bit of civilization. We lose a bit of society. We lose a bit of ourselves, and there's no reason for it. I mean, I love the digital technology. I've delivered my life to it, but honestly on the inside it's s till pretty crappy, honestly. It's not that impressive. Don't get snookered. Don't get snookered. Believe in yourselves as real mysterious living organisms who are not yet f ully understood by sci ence. Y ou have to unde rstand we live in a sea of mystery. We understand so little of our situation. Take joy in that mystery. In that mystery you alsofind the profound sense of responsibility that you're capable, that you're morally impelled to hold onto.

Science fiction is great in the theater. It stinks out on the street.



John Donvan:

Thank you, Jaron Lanier.



[applause]



The motion is Don't Trust the Promi se of Artificial Intelligence.

01:34:01 A nd here summarizing his position against the motion, James Hughes, executive director of the Institute for Ethics and Emerging Technologies, and author of "Citizen Cyborg."



James Hughes:

Two hundred twenty years ago the French aristocrat mathematician, Nicolas de Caritat,

marquis de Condorcet , was hiding from the terror of the guillotine. He had been the revolutionary. He had been a mathematician who had come up with the theorems that argued that the crowd could be wiser than the individual kind of the bedrock of democratic argument. And he had served in the revolutionary assembly, but he was a liberal and he fell on the wrong side of the Jacobins he was hiding from them. And he decided to write a history of human progress and he got through the first chapter basically, the first book of that history. The sketch for a historical picture of progress in the human spirit, and he argued that in the future, from his perspective, we would liberate women, we would eliminate slavery, we would e liminate toil. There would be no more work.

01:35:00 That we would eliminate disease and death. And that all of this would come about through the combined effort of the enlightenment project of democracy, equality,

solidarity, and the use of human re ason through the use of science and technology, new scientific languages and techniques. We face the same challenge today. He -- by the way, he finished that book, went outside, got caught by the J acobins and was executed. But for me he was an o ptimist, sitting there under threat of death looking forward to a future which we actually have achieved and we faced that challenge today. Remembering how far we've come and having the optimism that we can muddle through and use these tools that we have developed to create a better future, and therefore I urge you to vote against this resolution. Thank you.



John Donvan:

Thank you, James Hughes.[applause]





And that motion again, Don't Trust the Promise of Artificial Intelligence, and here to summariz e his closing statement against the motion, Andrew Keen. I'm sorry. Andrew Keen will be making his closing statement in support of the motion.

01:36:00 "The Internet is Not the Answer" is one of his books. He is also the host of "Keen On." Ladies and gentlemen, Andrew Keen.



[applause]



Andrew Keen:

Just ga ve me the subject to my next book title, "AI Is not the Answer." So we've -- Jaron I think in a very articulate way asked us not to fall into fantasy. This is a debate a

conversation about supposedly the reality or the fantasy of AI. It's -- when you make your votes at the end, we're not debating AI. We're not debating its potential. We're debating its promise. We're debating the ideology of the other team. We're debating the way in which they say AI can liberate us 