AGI skepticism

From Lesswrongwiki
Revision as of 19:21, 7 November 2012 by Kaj Sotala (talk | contribs)
Jump to: navigation, search

AGI skepticism involves objections to the possibility of Artificial General Intelligence being developed in the near future. Skeptics include various technology and science luminaries such as Douglas Hofstadter, Gordon Bell, Steven Pinker, and Gordon Moore:

"It might happen someday, but I think life and intelligence are far more complex than the current singularitarians seem to believe, so I doubt it will happen in the next couple of centuries." -- Douglas Hofstadter

A typical argument is that we currently only have narrow AI, and that there is no sign of progress towards general intelligence. Some critics have gone as far as to argue that predictions of near-term AGI belong to the realm of religion, not science or engineering.

Some skeptics go even more far, saying that discussion about AGI risk is a dangerous of waste time that diverts attention from more important issues. Daniel Dennett considers AGI risk an "imprudent pastime" because it distracts our attention from more immediate threats, and the philosopher Alfred Nordmann holds the view that ethical concern is a scarce resource, not to be wasted on unlikely future scenarios (1, 2).

There are also skeptics who think that the prospect of near-term AGI seems remote, but don't go so far as to dismiss the issue entirely. An (AAAI presidential panel on long-term AI futures) concluded that

There was overall skepticism about the prospect of an intelligence explosion as well as of a “coming singularity,” and also about the large-scale loss of control of intelligent systems. Nevertheless, there was a shared sense that additional research would be valuable on methods for understanding and verifying the range of behaviors of complex computational systems to minimize unexpected outcomes. Some panelists recommended that more research needs to be done to better define “intelligence explosion,” and also to better formulate different classes of such accelerating intelligences. Technical work would likely lead to enhanced understanding of the likelihood of such phenomena, and the nature, risks, and overall outcomes associated with different conceived variants.

External Links

See Also