Difference between revisions of "Singularity"

From Lesswrongwiki
Jump to: navigation, search
Line 1: Line 1:
The '''Singularity''' or '''Technological Singularity''' refers to a hypothetical future event where Artificial Intelligence vastly outperforms the abilities of the human mind. Due to the fact these Intelligences are, by definition, beyond Human comprehension it becomes difficult for us to imagine how such beings would behave. Various commentators have provided varying dates for when this will occur and the implications it would have for humanity.
+
The '''Singularity''' or '''Technological Singularity''' refers to a hypothetical future event where Artificial Intelligence vastly outperforms the abilities of the human mind. Due to the fact these Intelligences are, by definition, beyond Human comprehension it becomes difficult for us to imagine how such beings would behave or how the future may unfold beyond that point. Various commentators have provided varying dates for when this will occur and the implications it would have for humanity.
  
These predictions are based on the mathematical projections of Moore’s Law which has been accurately predicting the exponential growth of computers for over 50 years. These projections allow computer scientists to estimate the dates when certain computing projects (such as Brain Emulation) will be feasible, even if they are beyond the capabilities of today’s computers.   
+
These predictions are largely based on the mathematical projections of Moore’s Law which has been accurately predicting the exponential growth of computers for over 50 years. These projections allow computer scientists to estimate the dates when certain computing projects (such as Brain Emulation) will be feasible, even if they are beyond the capabilities of today’s computers.   
  
A number of prominent computer scientists have also speculated on the Singularity happening in the near term future. These include Hans Moravec, Eliezer Yudkowsky,  Bill Joy and most notoriously Ray Kurzweil in his book “The Singularity is Near”.
+
Eliezer Yudkowsky has observed that the varying perspectives on the Singularity can be broadly split into three different “major schools”  -  Accelerating Change (Ray Kurzweil), the Event Horizon (Vernor Vinge), and the Intelligence Explosion (I.J. Good).
  
The consequences of such an event range from Kurzweil’s largely positive predictions to Bill Joy’s existential pessimism outlined in his essay “Why the future doesn’t need us.”
+
The consequences of such an event range from Kurzweil’s largely positive predictions to Bill Joy’s existential pessimism outlined in his essay “Why the future doesn’t need us.”.
  
 
==Blog Post==
 
==Blog Post==
  
*[http://yudkowsky.net/singularity/schools Blog post by Eliezer S. Yudkowky on the Singularity.
+
*[http://yudkowsky.net/singularity/schools Blog post] by Eliezer S. Yudkowky on the Singularity.
*[http://facingthesingularity.com/ Singularity Blog] by The Singularity Institute's Executive Director Luke Muehlhauser.
 
  
 
==External Links==
 
==External Links==
  
 
*[http://www.youtube.com/watch?v=IfbOyw3CT6A Ray Kurzweil] Singularity Ted Talk by Ray Kurzweil
 
*[http://www.youtube.com/watch?v=IfbOyw3CT6A Ray Kurzweil] Singularity Ted Talk by Ray Kurzweil
* [http://www.wired.com/wired/archive/8.04/joy.html Why the future doesn’t need us] Bill Joy’s artcle for Wired magazine.
+
*[http://www.wired.com/wired/archive/8.04/joy.html Why the future doesn’t need us] Bill Joy’s artcle for Wired magazine.  
 
*[http://www.youtube.com/watch?v=mDhdt58ySJA Eliezer Yudkowsky] The Singularity Three Major Schools of Thought from Singularity Summit YouTube
 
*[http://www.youtube.com/watch?v=mDhdt58ySJA Eliezer Yudkowsky] The Singularity Three Major Schools of Thought from Singularity Summit YouTube
 
*[http://www.youtube.com/watch?v=LN2shXeJNz8 Bill Joy] Ted Talk “What I’m Worried about, what I’m excited about” by Bill Joy
 
*[http://www.youtube.com/watch?v=LN2shXeJNz8 Bill Joy] Ted Talk “What I’m Worried about, what I’m excited about” by Bill Joy

Revision as of 22:14, 7 June 2012

The Singularity or Technological Singularity refers to a hypothetical future event where Artificial Intelligence vastly outperforms the abilities of the human mind. Due to the fact these Intelligences are, by definition, beyond Human comprehension it becomes difficult for us to imagine how such beings would behave or how the future may unfold beyond that point. Various commentators have provided varying dates for when this will occur and the implications it would have for humanity.

These predictions are largely based on the mathematical projections of Moore’s Law which has been accurately predicting the exponential growth of computers for over 50 years. These projections allow computer scientists to estimate the dates when certain computing projects (such as Brain Emulation) will be feasible, even if they are beyond the capabilities of today’s computers.

Eliezer Yudkowsky has observed that the varying perspectives on the Singularity can be broadly split into three different “major schools” - Accelerating Change (Ray Kurzweil), the Event Horizon (Vernor Vinge), and the Intelligence Explosion (I.J. Good).

The consequences of such an event range from Kurzweil’s largely positive predictions to Bill Joy’s existential pessimism outlined in his essay “Why the future doesn’t need us.”.

Blog Post

  • Blog post by Eliezer S. Yudkowky on the Singularity.

External Links