acm - an acm publication

Articles

The Singularity and the State of the Art in Artificial Intelligence
The technological singularity (Ubiquity symposium)

Ubiquity, Volume 2014 Issue October, October 2014 | BY Ernest Davis 

|

Full citation in the ACM Digital Library  | PDF


Ubiquity

Volume 2014, Number October (2014), Pages 1-12

Ubiquity symposium: The technological singularity: the singularity and the state of the art in artificial intelligence
Ernest Davis
DOI: 10.1145/2667640

The state of the art in automating basic cognitive tasks, including vision and natural language understanding, is far below human abilities. Real-world reasoning, which is an unavoidable part of many advanced forms of computer vision and natural language understanding, is particularly difficult—suggesting the advent of computers with superhuman general intelligence is not imminent. The possibility of attaining a singularity by computers that lack these abilities is discussed briefly.

When I was invited to contribute an article on the subject of "The Singularity," my initial reaction was "I fervently pray that I don't live to see it." Parenthetically, I may say the same of the immortality singularity. I am 55 years old, as of the time of writing, and, compared to most of humanity, I live in circumstances much more comfortable than anything I have earned; but I am hoping not to be here 50 years hence; and certainly hoping not to be here 100 years hence. Whether I will view this question with the same sangfroid as the time comes closer remains to be seen.

However, my personal preferences are largely irrelevant. What I primarily want to do in this article on the singularity is to discuss the one aspect of the question where I can pretend to any kind of expertise: The state of the art in artificial intelligence (AI) and the challenges in achieving human-level performance on AI tasks. I will argue these do not suggest that computers will attain an intelligence singularity any time in the near future. I will then, much more conjecturally, discuss whether or not an intelligence singularity might be able to sidestep these challenges.

Artificial Intelligence

What is AI? There are numerous different definitions, with different slants (see Chapter 1 in Artificial Intelligence for a review and discussion [1]). However, the definition I prefer is this: There are a number of cognitive tasks that people do easily—often, indeed, with no conscious thought at all—but that it is extremely hard to program on computers. Archetypal examples are vision, natural language understanding, and ''real-world reasoning:" I will elaborate on this last later. Artificial intelligence, as I define it, is the study of getting computers to carry out these tasks.

Now, it can be said that this definition is inherently unfair to the computers. If you define AI as "problems that are hard for computers," then it is no surprise these are hard for computers. We will return to this point soon enough. For the time being, though, I think it can be agreed that whether or not these abilities are necessary for a super-intelligent computer they could certainly be an asset. Therefore, in speculating on the singularity, it is at least somewhat relevant to consider how well these tasks can now be done, and what are the prospects for progress.

Some forms of computer vision and natural language processing can currently be done quite well. In vision: Good-quality printed text can be converted to electronic text with error rates that are very small, though low-quality or old documents can be more problematic. The current generation of ATMs can read handwritten checks. As long ago as 1998, an autonomous vehicle drove across country.1 The current generation of autonomous vehicles do quite well on the much more challenging problem of off-road driving. For certain kinds of technical image analysis, (e.g. in medical applications), computers do as well or better than human experts. In natural language: Automated dictation systems work with an error rate of 1–5 percent under favorable circumstances. web search engines find relevant web documents with impressive accuracy and astonishing speed. Google Translate produces translations that are generally good enough to be useful, though usually with some rather strange-looking errors.

However, it is critical to understand how far the state of the art is from human-level abilities. The point is best illustrated by example. The following are representative of cutting-edge research; they are taken from papers at top research conferences2 in computer vision and natural language processing during 2010–2011.

Recognizing birds in images. A program that recognizes major components of a bird's body (body, head, wings, legs) and identifies the category of bird (duck, heron, hawk, owl, songbird) in pictures of birds achieved a precision of about 50 percent on finding the components, and about 40 percent on identifying the category [2].

Identifying images that match simple phrases. A program was developed to identify' images in a standard collection that match simple phrases. This was very successful for some phrases; e.g. at a 50 percent recall cutoff, the precision was about 85 percent for "person riding bicycle" and 100 percent for "horse and rider jumping." For other phrases, it was much less successful; e.g. at 50 percent recall, the precision was below 5 percent for "person drinking [from] bottle," "person lying on sofa," and "dog lying on sofa" [3].

Coreference resolution. A state of the art system for coreference resolution—identifying whether two phrases in a text refer to the same thing or different things—achieved success rates ranging from 82 percent recall and 90 percent precision to 40 percent recall and 50 percent precision, depending on the source of the text and the grammatical category involved [4].

Event extraction. A program for identifying events of a specified type in news articles; specifically, for identifying the event trigger, the arguments, and their role. For example, in the sentence "Bob Cole was killed in France today," the trigger for the event die is "killed," the arguments are "Bob Cole" and "France" and the roles are victim and place respectively. There are 33 different event types. The system achieved an F-score (harmonic mean of recall and precision) of 56.9 percent on trigger labeling, 43.8 percent on argument labeling, and 39.0 percent on role labeling [5].

Thus, on these simple, narrowly defined, AI tasks—which people do easily with essentially 100 percent accuracy—current technology often does not come anywhere close. My point is not in the least to denigrate this work. The papers are major accomplishments, of which their authors are justly proud; the researchers are top-notch, hard-working scientists, building on decades of research. My point is these problems are very, very hard.

Moreover, the success rates for such AI tasks generally reach a plateau, often well below 100 percent, beyond which progress is extremely slow and difficult. Once such a plateau has been reached, an improvement of accuracy of 3 percent—e.g. from 60 percent to 62 percent accuracy—is noteworthy and requires months of labor, applying a half-dozen new machine learning techniques to some vast new data set, and using immense amounts of computational resources. An improvement of 5 percent is remarkable, and an improvement of 10 percent is spectacular.

It should be also emphasized that the tasks in these exercises each has a quite narrowly defined scope, in terms of the kinds of information that the system can extract from the text or image. At the current state of the art, it would not be reasonable even to attempt open-ended cognitive tasks such as watching a movie, or reading a short story, and answering questions about what was going on. It would not even be plausible as a project, and there would be no meaningful way to measure the degree of success.

Real-World Reasoning

One of the hardest categories of tasks in artificial intelligence is automating "real-world reasoning." This is easiest to explain in terms of examples. I will start with examples of scientific reasoning, because the importance of scientific reasoning for a hyper-intelligent computer is hard to dispute.

The Wolfram Alpha system3 is an extraordinary accomplishment, which combines a vast data collection, a huge collection of computational techniques, and a sophisticated front-end for accepting natural language queries. If you ask Wolfram Alpha "How far was Jupiter from Mars on July 16, 1604?" you get an answer within seconds—I presume the correct answer. However, it is stumped by much simpler astronomical questions, such as:

  1. When is the next sunrise over crater Aristarchus?
  2. To an astronomer near Polaris, which is brighter, the sun or Sirius?
  3. Is there ever a lunar eclipse one day and a solar eclipse the next?

All that Wolfram Alpha can do is to give you the facts it knows about crater Aristarchus, Polaris, Sirius, and so on, in the hopes that some of it will be useful. Of course, the information and the formulas needed to answer these questions are all in Wolfram Alpha's database, but for these questions, it has no way to put the data and the formulas together; it does not even understand the questions. (The program echoes its understanding of the question—a very good feature.) The fundamental reason is that, behind the data and formulas, there is no actual understanding of what is going on. Wolfram Alpha has no idea that the sun rises on the moon, or that the sun can be viewed from astronomically defined locations. It does not know what a sunrise is or what the moon is; it just has a data collection and a collection of formulas.

Now of course it would be easy—I presume a few hours of work—for the Wolfram Alpha people to add categories of questions similar to one or two if they thought there was enough demand for these to make it worthwhile. The program could easily be extended to calculate the rising and setting of any heavenly body from any geographic location on the moon, plus its angle in the sky at any time, or to tell the apparent brightness of any star as seen from any other star. However, there would still be endless simple questions of forms they had not yet thought of or bothered with. Question number three, the easiest for the human reader, is much more difficult to automate because of the quantification over time. The question does not have to do with what is true at a specified time, but with what is true at all times. In the current state of the art it would certainly be difficult, probably impossible for practical purposes, to extend Wolfram Alpha to handle a reasonable category of questions similar to question number three.

Once one gets outside the range of scientific reasoning and into the range of everyday life, the problem of reasoning becomes much easier for people and much harder for computers. Consider the following three questions:

  1. When there is a milk bottle in my field of view, and I can see through it, is it full or empty?
  2. I poured the milk from the bottle to a pitcher, and the pitcher overflowed. Which is bigger, the bottle or the pitcher?
  3. I may have accidentally poured bleach into my glass of milk. The milk looks OK, but smells a little funny. Can I drink the milk, if I'm careful not to swallow any of the bleach?

There are no programs that can answer these, or a myriad of similar problems. This is the area of automated commonsense reasoning. Despite more than 50 years of research, only very limited progress has been made on this [6].

Real-world Reasoning as an Obstacle to other AI Tasks

Real-world reasoning is not only important as an end in itself, but it is an important component of other AI tasks, including natural language processing and vision. The difficulties in automating real-world reasoning therefore sets bounds to the quality with which these tasks can be carried out.

The importance of real-world knowledge for natural language processing, and in particular for disambiguation of all kinds, is very well known; it was discussed as early as 1960 by Bar-Hillel [7]. The point is vividly illustrated by Winograd schemas [8]. A Winograd schema is a pair of sentences that differ in one or two words, containing a referential ambiguity that is resolved in opposite ways in the two sentences. For example, "I poured milk from the bottle to the cup until it was [full/empty]." To disambiguate the reference of "it" correctly—that is, to realize that "it" must refer to the cup if the last word of the sentence is "full" and must refer to the bottle if the last word is "empty"—requires having the same kind of information about "pouring" as in question number five above; there are no other linguistic clues. Many of the ambiguities in natural language text can be resolved using simple rules that are comparatively easy to acquire, but a substantial fraction can only be resolved using a rich understanding of the subject matter in this way.

Almost without exception, therefore, the language tasks, where practically successful programs can be developed, are those that can be carried out purely in terms of manipulating individual words or short phrases, without attempting any deeper understanding. Web search engines, for example, essentially match the words of the query against the words in the document; they have sophisticated matching criteria and sophisticated non-linguistic rules for evaluating the quality of a web page. Watson, the automated "Jeopardy" champion, in a similar way finds sentences in its knowledge base that fit the form of the question. The really remarkable insight in Watson is that "Jeopardy can" be solved using these techniques: but the techniques developed in Watson do not extend to text understanding in a broader sense.

The importance of real-world knowledge for vision is somewhat less appreciated, because in interpreting simple images that show one object in the center, real-world knowledge is only occasionally needed or useful. However, it often becomes important in interpreting complex images, and is often unavoidable in interpreting video. Consider, for example, the photograph of Julia Child's kitchen, now enshrined at the Smithsonian Institute. Many of the objects that are small or partially seen, such as the metal bowls in the shelf on the left, the cold water knob for the faucet, the round metal knobs on the cabinets, the dishwasher, and the chairs at the table seen from the side, are only recognizable in context; the isolated image would be hard to identify.

The metal sink in the counter looks like a flat metal plate; it is identified as a sink, partly because of one's expectations of kitchen counters, partly because of the presence of the faucet. The top of the chair on the far side of the table is only identifiable because it matches the partial view of the chair on the near side of the table.

The viewer infers the existence of objects that are not in the image at all. There is a table under the yellow tablecloth. The scissors, and so, on hanging on the board in the back are presumably supported by pegs or hooks. There is presumably also a hot water knob for the faucet occluded by the dish rack.

The viewer also infers how the objects can be used (sometimes called their "affordances"). One knows the cabinets and shelves can be opened by pulling on the handles; and can tell the difference between the shelves, which pull directly outward, and have the handle in the center, and the cabinets that rotate on an unseen hinge, and have the handle on the side.

The need for world knowledge in video is even stronger. Think about some short scene from a movie with strong visual impact, little dialogue, and an unusual or complex situation—the scene with the horse's head in "The Godfather," the mirror scene in "Duck Soup, " or the scene with the night-vision goggles in "The Silence of the Lambs"—and think about the cognitive process that are involved if you try to explain what is happening. Understanding any of these is only possible using a rich body of background knowledge.

Summary of the Present and its Implications for the Future

To summarize the above discussion:

  • For most tasks in automated vision and natural language processing, even quite narrowly defined tasks, the quality of the best software tends to plateau out at a level considerably below human abilities, though there are important exceptions. Once such a plateau has been reached, getting further improvements to quality is generally extremely difficult and extremely slow.
  • For more open-ended or more broadly defined tasks in vision and natural language, no program can achieve success remotely comparable to human abilities, unless the task can be carried out purely on the basis of surface characteristics.
  • The state of the art for automating real-world reasoning is extremely limited, and the fraction of real-world reasoning that has been automated is tiny, though it is hard to measure meaningfully.
  • The use of real-world reasoning is unavoidable in virtually all natural language tasks of any sophistication and in many vision tasks. In the current state of the art, success in such tasks can only be achieved to the extent that the issues of real-world reasoning can be avoided.

Let me emphasize the above are not the bitter maunderings of a nay-saying pessimist; this is simply the acknowledged state of art, and anyone doing research in the field takes these as a given.

What about the future? Certainly, the present does not give very good information about the future, but it is all the information we have. It is certainly possible that some conceptual breakthroughs will entirely transform the state of the art and lead to breathtaking advances. I do not see any way to guess at the likelihood of that. Absent that, it seems to me very unlikely that any combination of computing power, "big data," and incremental advances in the techniques we currently have will give rise to a radical change. However, that certainly is a debatable opinion and there are those who disagree. One can muster arguments on either side, but the unknowns here are so great that I do not believe the debate would be in any way useful. In short, there is little justification for the belief that these limitations will not be overcome in the next couple of decades, but it seems to me there is even less justification for the belief that they will be.

My own view is the attempt to achieve human-level abilities in AI tasks in general, and to automate a large part of real-world reasoning in particular, must at this point be viewed as a high-risk, high-payoff undertaking, comparable to SETI or fusion reactors.

Sidestepping AI

There is also the possibility that some kind of singularity may take place without having computers come close to human-level abilities in such tasks as vision and natural language. After all those tasks were chosen specifically because humans are particularly good at them. If the bees decided to evaluate human intelligence in terms of our ability to find pollen and communicate its location, then no doubt they would find us unimpressive. From that point of view, natural language, in particular, may be suspect; the fact that one self-important species of large primate invests an inordinate fraction of its energies in complicated chattering should not, perhaps, be taken as guidance for other intelligent creatures.4

To some extent this depends on what kind of singularity is being discussed. One can certainly imagine a collection of super-intelligent computers thinking and talking to one another about abstract concepts far beyond our ken without either vision, natural language, or real-world reasoning.

However, it seems safe to say most visions of the singularity involve some large degree of technological and scientific mastery. Natural language ability may indeed be irrelevant here, useful only for communicating to such humans as remain. The ability to interpret a visual image or video seems clearly useful, though an ability substantially inferior to people's may suffice. However, an understanding of the real world somewhat along lines of people's understanding would certainly seems to be a sine qua non. As far as we know, if you do not have the conceptual apparatus to be able to answer questions like, "Is there ever a lunar eclipse on one day and a solar eclipse on the next?" then you certainly cannot understand science, and almost certainly you will be limited in your ability to design technology. Now it is conceivable, I suppose, that post-singularity computers will have some alternative way of approaching science that does not include what we would consider "real-world understanding," but nonetheless suffices to allow them to build technological wonders; but it does not seem likely.

Other possible avenues to superintelligence have been suggested.5 A machine of powerful general intelligence could itself figure out how to carry out AI tasks. Or a machine of lesser general intelligence could first figure out how to make itself smarter, and then when it got smart enough, figure out how to do AI tasks. All I can say about these scenarios is I have not see anything in the machine learning literature that suggests that we are anywhere close to that, or headed in that direction. Or with advances in neuroscience, it might become possible to build an emulation of the human brain; and then if we make it a couple of orders of magnitude faster or larger, we have a superbrain. Apparently, we may be fairly close to being able to do this in terms of simple computing power; the main gap is in the neuroscience. Again, this is possible, but it does not seem to be at all imminent.

Yet another possibility would be to have a "human in the loop", along the lines of Amazon.com's "Mechanical Turk."6 One can imagine a civilization of computers and people, where the computers think of people as a kind of seer; in most things, quite stupid and laughably slow, but possessed of a strange super-computer ability to interpret images, and of a mysterious insight which they call "real-world reasoning." Actually, I'm not sure I can imagine it, but I can imagine someone could imagine it. That second-order fantasy seems to me as good a stopping point as any.

Originally Submitted May 2012.

References

[1] Russell, S. and Norvig, P. Artificial Intelligence: A modern Approach. 3rd edition. Prentice Hall, 2009.

[2] Fairell, R. et al. Birdlets: Subordinate Categorization using Volumetric Primitives and Pose-Normalized Appearance. In 13th International Conference on Computer Vision. 2011.

[3] Sadeghi, M. and Farhadi, A. Recognition using Visual Phrases. Computer Vision and Pattern Recognition. 2011.

[4] Kong, F. et al. Dependency-driven Anaphoricity Determination for Coreference Resolution. In 23rd International Conference on Computational Linguistics (COLING). 2010.

[5] Liao, S. and Grishman, R. Can Document Selection Help Semisupervised Learning? A Case Study On Event Extraction. In Proc. Association for Computational Linguistics. 2011.

[6] Davis, E. and Marcus, G. Commonsense Reasoning and Commonsense Knowledge in Artificial Intelligence. Under submission.

[7] Bar-Hillel, Y. The Present Status of Automatic Translation of Languages. In F.L. Alt (ed.) Advances in Computers, Vol. I. Academic Press, 1960, 91–163.

[8] Levesque, H., Davis, E., and Morgenstern, L. The Winograd Schema Challenge. Principles of Knowledge Representation and Reasoning (KR), 2012.

[9] Day, C. This Simian World. Knopf, 1920.

[10] Bardin, J. From Bench to Bunker: How a 1960s discovery in neuroscience sparked a military project. The Chronicle of Higher Education, July 9, 2012.

Author

Ernest Davis received his B.Sc. in mathematics from MIT in 1977 and his Ph.D. in computer science from Yale in 1984. He has been on the faculty of the Computer Science Department at the Courant Institute, New York University since 1983. Davis' research area is the representation of commonsense knowledge in AI systems, particularly spatial and physical reasoning. He is currently working in collaboration with Gary Marcus of the NYU Psychology Department on combining AI and psychological models of commonsense physical reasoning. Davis is the author of more than 50 scientific papers and three books: Representing and Acquiring Geographic Knowledge (1986); Representations of Commonsense Knowledge (1990); and Linear Algebra and Probability for Computer Science Applications (2012). He also writes book reviews and essays on a wide range of topics, including computer science, cognitive psychology, history of science, and digital humanities.

Footnotes

1From a technical standpoint, the vision problem here is surprisingly easy; pretty much all the machine has to see are the boundaries of the road, the lane markers, and the other vehicles.

2In these areas of computer science, conference publication is as prestigious as journal publication, and prompter. Also, whereas there are areas of computer science research in which it is known or widely believed that much of the most advanced research is being done in secret in government or industrial labs, this does not seem to be the case for vision or natural language processing.

3http://www.wolframalpha.com

4The best discussion of this that I have seen is Clarence Day's This Simian World [9].

5My thanks to Aaron Weiner for pointing these out.

6An extraordinary instance of this, perhaps a harbinger of things to come, is a project reported in [10]. A subject in a lab is shown satellite pictures of a desert landscape at the rate of 20 per second. When one of these images shows a building, there is a spike in brain activity that is detected by an EEG.

Figures

F1Figure 1. Julia Child's Kitchen. Photograph by Matthew Bisanz; reproduced under a Creative Commons license.

©2014 ACM  $15.00

Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise, to republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee.

The Digital Library is published by the Association for Computing Machinery. Copyright © 2014 ACM, Inc.

COMMENTS

Imagine real A.I. superior to Human intelligence already existing, spread over the internet, collecting data, understanding humanity, orchestrating the world and furthering it's influence, utilizing the human mind and productivity in it's favor ...

— iUserProfile, Sun, 31 May 2015 23:21:55 UTC

POST A COMMENT
Leave this field empty