Difference between revisions of "Interview series on risks from AI"

From Lesswrongwiki
Jump to: navigation, search
m (he took no care to make sure that the questions asked were not the questions that would be most likely to ensure that the responders would not assume maximum disdain for the respective topic)
(Added an interview with Sir William Timothy Gowers.)
 
(4 intermediate revisions by 2 users not shown)
Line 1: Line 1:
In 2011, [http://kruel.co Alexander Kruel] (XiXiDu) started an unreflective, poorly thought-out Q&A-style interview series asking various people about their perception of artificial intelligence and associated risks.
+
In 2011, [http://kruel.co Alexander Kruel] (XiXiDu) started a Q&A style interview series asking various people about their perception of artificial intelligence and possible risks associated with it.
  
 
==Interviews==
 
==Interviews==
Line 17: Line 17:
 
* [http://lesswrong.com/r/discussion/lw/682/j%C3%BCrgen_schmidhuber_on_risks_from_ai Professor Jürgen Schmidhuber]
 
* [http://lesswrong.com/r/discussion/lw/682/j%C3%BCrgen_schmidhuber_on_risks_from_ai Professor Jürgen Schmidhuber]
 
* [http://lesswrong.com/r/discussion/lw/65v/stan_franklin_on_risks_from_ai Professor Stan Franklin]
 
* [http://lesswrong.com/r/discussion/lw/65v/stan_franklin_on_risks_from_ai Professor Stan Franklin]
 +
* [http://kruel.co/2013/07/08/qa-with-matt-mahoney-on-risks-from-ai/ Dr. Matt Mahoney]
 +
* [http://kruel.co/2013/07/16/qa-with-timothy-gowers-on-risks-from-ai/ Sir William Timothy Gowers]
 
* [http://lesswrong.com/lw/9fz/qa_with_abram_demski_on_risks_from_ai/ Abram Demski]
 
* [http://lesswrong.com/lw/9fz/qa_with_abram_demski_on_risks_from_ai/ Abram Demski]
 
* [http://lesswrong.com/r/discussion/lw/8tt/qa_with_richard_carrier_on_risks_from_ai/ Dr. Richard Carrier]
 
* [http://lesswrong.com/r/discussion/lw/8tt/qa_with_richard_carrier_on_risks_from_ai/ Dr. Richard Carrier]
Line 56: Line 58:
 
==See also==
 
==See also==
  
* [http://wiki.lesswrong.com/wiki/Muehlhauser_interview_series_on_AGI Muehlhauser interview series on AGI]
+
* [[Muehlhauser interview series on AGI]]
 +
* [http://intelligence.org/2013/05/15/when-will-ai-be-created/ When Will AI Be Created?]
 
* [http://lesswrong.com/lw/e36/ai_timeline_predictions_are_we_getting_better/ AI timeline predictions: are we getting better?]
 
* [http://lesswrong.com/lw/e36/ai_timeline_predictions_are_we_getting_better/ AI timeline predictions: are we getting better?]
 
* [http://fora.tv/v/c16838 Stuart Armstrong: How We’re Predicting AI], The Singularity Summit 2012
 
* [http://fora.tv/v/c16838 Stuart Armstrong: How We’re Predicting AI], The Singularity Summit 2012
 
* [http://sethbaum.com/ac/2011_AI-Experts.html Experts on artificial general intelligence provide estimates for the future of AGI.], Baum, Seth D., Ben Goertzel, and Ted G. Goertzel, 2011.
 
* [http://sethbaum.com/ac/2011_AI-Experts.html Experts on artificial general intelligence provide estimates for the future of AGI.], Baum, Seth D., Ben Goertzel, and Ted G. Goertzel, 2011.
* [http://www.fhi.ox.ac.uk/__data/assets/pdf_file/0015/21516/MI_survey.pdf Machine Intelligence Survey], Sandberg, A. and Bostrom, N. (2011). Future of Humanity Institute, Oxford University.
+
* [http://www.fhi.ox.ac.uk/news/2011/?a=21516 Machine Intelligence Survey], Sandberg, A. and Bostrom, N. (2011). Future of Humanity Institute, Oxford University.

Latest revision as of 18:38, 16 July 2013

In 2011, Alexander Kruel (XiXiDu) started a Q&A style interview series asking various people about their perception of artificial intelligence and possible risks associated with it.

Interviews

A list of all interviews:

Questions

The current set of questions:

  • Assuming beneficial political and economic development and that no global catastrophe halts progress, by what year would you assign a 10%/50%/90% chance of the development of artificial intelligence that is roughly as good as humans (or better, perhaps unevenly) at science, mathematics, engineering and programming?
  • Once we build AI that is roughly as good as humans (or better, perhaps unevenly) at science, mathematics, engineering and programming, how much more difficult will it be for humans and/or AIs to build an AI which is substantially better at those activities than humans?
  • Do you ever expect artificial intelligence to overwhelmingly outperform humans at typical academic research, in the way that they may soon overwhelmingly outperform humans at trivia contests, or do you expect that humans will always play an important role in scientific progress?
  • What probability do you assign to the possibility of an AI with initially roughly professional human-level competence (or better, perhaps unevenly) at general reasoning (including science, mathematics, engineering and programming) to self-modify its way up to vastly superhuman capabilities within a matter of hours/days/< 5 years?
  • How important is it to research risks associated with artificial intelligence that is good enough at general reasoning (including science, mathematics, engineering and programming) to be capable of radical self-modification, before attempting to build one?
  • What probability do you assign to the possibility of human extinction within 100 years as a result of AI capable of self-modification (that is not provably non-dangerous, if that is even possible)? P(human extinction by AI | AI capable of self-modification and not provably non-dangerous is created)

Mail Template

The exact wording of the email:

Subject: Questions regarding possible risks from artificial intelligence
Dear Professor/Dr NAME,
I am currently trying to learn more about the academic perception of artificial general intelligence and possible risks associated with it. Consequently I am curious about your opinion as a noted expert.
I would like to ask you a few questions and your permission to publish your possible answers in order to estimate the academic awareness and perception of risks from AI. I am not a journalist and do not represent any publication, nor do I maintain a formal academic relationship. I am conducting an informal interview for a community blog: lesswrong.com
Please let me know if you have any questions or if you are interested in third-party material that does expand on various aspects of my questions.
Here is a list of people that I had a chance to interview so far: NAMES
QUESTIONS
Yours sincerely,
NAME
ADDRESS

See also