American library books ยป Other ยป The Coming Technological Singularity by Vernor Vinge (novels for teenagers .TXT) ๐Ÿ“•

Read book online ยซThe Coming Technological Singularity by Vernor Vinge (novels for teenagers .TXT) ๐Ÿ“•ยป.   Author   -   Vernor Vinge



1 2 3
Go to page:

The Coming Technological Singularity: How to Survive in the Post-Human Era

 

Vernor Vinge

Department of Mathematical Sciences

San Diego State University

 

(c) 1993 by Vernor Vinge (This article may be reproduced for noncommercial purposes if it is copied in its entirety, including this notice.)

 

The original version of this article was presented at the VISION-21 Symposium sponsored by NASA Lewis Research Center and the Ohio Aerospace Institute, March 30-31, 1993. A slightly changed version appeared in the Winter 1993 issue of Whole Earth Review.

 

Abstract

 

Within thirty years, we will have the technological means to create superhuman intelligence. Shortly after, the human era will be ended.

 

Is such progress avoidable? If not to be avoided, can events be guided so that we may survive? These questions are investigated. Some possible answers (and some further dangers) are presented.

 

What is The Singularity?

 

The acceleration of technological progress has been the central feature of this century. I argue in this paper that we are on the edge of change comparable to the rise of human life on Earth. The precise cause of this change is the imminent creation by technology of entities with greater than human intelligence. There are several means by which science may achieve this breakthrough (and this is another reason for having confidence that the event will occur):

o There may be developed computers that are โ€œawakeโ€ and

superhumanly intelligent. (To date, there has been much

controversy as to whether we can create human equivalence in a

machine. But if the answer is โ€œyes, we canโ€, then there is little

doubt that beings more intelligent can be constructed shortly

thereafter.)

o Large computer networks (and their associated users) may โ€œwake

upโ€ as a superhumanly intelligent entity.

o Computer/human interfaces may become so intimate that users

may reasonably be considered superhumanly intelligent.

o Biological science may provide means to improve natural

human intellect.

 

The first three possibilities depend in large part on improvements in computer hardware. Progress in computer hardware has followed an amazingly steady curve in the last few decades [17]. Based largely on this trend, I believe that the creation of greater than human intelligence will occur during the next thirty years. (Charles Platt [20] has pointed out that AI enthusiasts have been making claims like this for the last thirty years. Just so Iโ€™m not guilty of a relative-time ambiguity, let me more specific: Iโ€™ll be surprised if this event occurs before 2005 or after 2030.)

 

What are the consequences of this event? When greater-than-human intelligence drives progress, that progress will be much more rapid. In fact, there seems no reason why progress itself would not involve the creation of still more intelligent entities โ€” on a still-shorter time scale. The best analogy that I see is with the evolutionary past: Animals can adapt to problems and make inventions, but often no faster than natural selection can do its work โ€” the world acts as its own simulator in the case of natural selection. We humans have the ability to internalize the world and conduct โ€œwhat ifโ€™sโ€ in our heads; we can solve many problems thousands of times faster than natural selection. Now, by creating the means to execute those simulations at much higher speeds, we are entering a regime as radically different from our human past as we humans are from the lower animals.

 

From the human point of view this change will be a throwing away of all the previous rules, perhaps in the blink of an eye, an exponential runaway beyond any hope of control. Developments that before were thought might only happen in โ€œa million yearsโ€ (if ever) will likely happen in the next century. (In [5], Greg Bear paints a picture of the major changes happening in a matter of hours.)

 

I think itโ€™s fair to call this event a singularity (โ€œthe Singularityโ€ for the purposes of this paper). It is a point where our old models must be discarded and a new reality rules. As we move closer to this point, it will loom vaster and vaster over human affairs till the notion becomes a commonplace. Yet when it finally happens it may still be a great surprise and a greater unknown. In the 1950s there were very few who saw it: Stan Ulam [28] paraphrased John von Neumann as saying:

 

One conversation centered on the ever accelerating progress of

technology and changes in the mode of human life, which gives the

appearance of approaching some essential singularity in the

history of the race beyond which human affairs, as we know them,

could not continue.

 

Von Neumann even uses the term singularity, though it appears he is thinking of normal progress, not the creation of superhuman intellect. (For me, the superhumanity is the essence of the Singularity. Without that we would get a glut of technical riches, never properly absorbed (see [25]).)

 

In the 1960s there was recognition of some of the implications of superhuman intelligence. I. J. Good wrote [11]:

 

Let an ultraintelligent machine be defined as a machine

that can far surpass all the intellectual activities of any

any man however clever. Since the design of machines is one of

these intellectual activities, an ultraintelligent machine could

design even better machines; there would then unquestionably

be an โ€œintelligence explosion,โ€ and the intelligence of man

would be left far behind. Thus the first ultraintelligent

machine is the last invention that man need ever make,

provided that the machine is docile enough to tell us how to

keep it under control.

 

โ€ฆ

 

It is more probable than not that, within the twentieth century,

an ultraintelligent machine will be built and that it will be

the last invention that man need make.

 

Good has captured the essence of the runaway, but does not pursue its most disturbing consequences. Any intelligent machine of the sort he describes would not be humankindโ€™s โ€œtoolโ€ โ€” any more than humans are the tools of rabbits or robins or chimpanzees.

 

Through the โ€™60s and โ€™70s and โ€™80s, recognition of the cataclysm spread [29] [1] [31] [5]. Perhaps it was the science-fiction writers who felt the first concrete impact. After all, the โ€œhardโ€ science-fiction writers are the ones who try to write specific stories about all that technology may do for us. More and more, these writers felt an opaque wall across the future. Once, they could put such fantasies millions of years in the future [24]. Now they saw that their most diligent extrapolations resulted in the unknowable โ€ฆ soon. Once, galactic empires might have seemed a Post-Human domain. Now, sadly, even interplanetary ones are.

 

What about the โ€™90s and the โ€™00s and the โ€™10s, as we slide toward the edge? How will the approach of the Singularity spread across the human world view? For a while yet, the general critics of machine sapience will have good press. After all, till we have hardware as powerful as a human brain it is probably foolish to think weโ€™ll be able to create human equivalent (or greater) intelligence. (There is the far-fetched possibility that we could make a human equivalent out of less powerful hardware, if we were willing to give up speed, if we were willing to settle for an artificial being who was literally slow [30]. But itโ€™s much more likely that devising the software will be a tricky process, involving lots of false starts and experimentation. If so, then the arrival of selfaware machines will not happen till after the development of hardware that is substantially more powerful than humansโ€™ natural equipment.)

 

But as time passes, we should see more symptoms. The dilemma felt by science fiction writers will be perceived in other creative endeavors. (I have heard thoughtful comic book writers worry about how to have spectacular effects when everything visible can be produced by the technologically commonplace.) We will see automation replacing higher and higher level jobs. We have tools right now (symbolic math programs, cad/cam) that release us from most low-level drudgery. Or put another way: The work that is truly productive is the domain of a steadily smaller and more elite fraction of humanity. In the coming of the Singularity, we are seeing the predictions of true technological unemployment finally come true.

 

Another symptom of progress toward the Singularity: ideas themselves should spread ever faster, and even the most radical will quickly become commonplace. When I began writing science fiction in the middle โ€™60s, it seemed very easy to find ideas that took decades to percolate into the cultural consciousness; now the lead time seems more like eighteen months. (Of course, this could just be me losing my imagination as I get old, but I see the effect in others too.) Like the shock in a compressible flow, the Singularity moves closer as we accelerate through the critical speed.

 

And what of the arrival of the Singularity itself? What can be said of its actual appearance? Since it involves an intellectual runaway, it will probably occur faster than any technical revolution seen so far. The precipitating event will likely be unexpected โ€” perhaps even to the researchers involved. (โ€œBut all our previous models were catatonic! We were just tweaking some parametersโ€ฆ.โ€) If networking is widespread enough (into ubiquitous embedded systems), it may seem as if our artifacts as a whole had suddenly wakened.

 

And what happens a month or two (or a day or two) after that? I have only analogies to point to: The rise of humankind. We will be in the Post-Human era. And for all my rampant technological optimism, sometimes I think Iโ€™d be more comfortable if I were regarding these transcendental events from one thousand years remove โ€ฆ instead of twenty.

 

Can the Singularity be Avoided?

 

Well, maybe it wonโ€™t happen at all: Sometimes I try to imagine the symptoms that we should expect to see if the Singularity is not to develop. There are the widely respected arguments of Penrose [19] and Searle [22] against the practicality of machine sapience. In August of 1992, Thinking Machines Corporation held a workshop to investigate the question โ€œHow We Will Build a Machine that Thinksโ€ [27]. As you might guess from the workshopโ€™s title, the participants were not especially supportive of the arguments against machine intelligence. In fact, there was general agreement that minds can exist on nonbiological substrates and that algorithms are of central importance to the existence of minds. However, there was much debate about the raw hardware power that is present in organic brains. A minority felt that the largest 1992 computers were within three orders of magnitude of the power of the human brain. The majority of the participants agreed with Moravecโ€™s estimate [17] that we are ten to forty years away from hardware parity. And yet there was another minority who pointed to [7] [21], and conjectured that the computational competence of single neurons may be far higher than generally believed. If so, our present computer hardware might be as much as ten orders of magnitude short of the equipment we carry around in our heads. If this is true (or for that matter, if the Penrose or Searle critique is valid), we might never see a Singularity. Instead, in the early โ€™00s we would find our hardware performance curves beginning to level off โ€” this because of our inability to automate the design work needed

1 2 3
Go to page:

Free e-book: ยซThe Coming Technological Singularity by Vernor Vinge (novels for teenagers .TXT) ๐Ÿ“•ยป   -   read online now on website american library books (americanlibrarybooks.com)

Comments (0)

There are no comments yet. You can be the first!
Add a comment