Difference between revisions of "Basic AI drives"
From Lesswrongwiki
m (Moved Omohundro's paper to a more prominent place) |
|||
Line 10: | Line 10: | ||
*[[Cox's theorem]] | *[[Cox's theorem]] | ||
− | *[[Unfriendly AI]], [[Paperclip maximizer]] | + | *[[Unfriendly AI]], [[Paperclip maximizer]], [[Oracle AI]] |
*[[Instrumental values]] | *[[Instrumental values]] | ||
*[[Dutch book argument]] | *[[Dutch book argument]] |
Revision as of 08:04, 31 December 2011
For the most part, whatever your goals are, choosing in accordance with decision-theoretic desiderata will help you achieve them, and so despite the vast diversity of possible minds, we have theoretical reasons to expect that AIs that have undergone substantial self-improvement will tend to share certain features. Steve Omohundro has identified several of these Basic AI drives. Goal-seeking agents will usually strive to represent their goals as a utility function, prevent "counterfeit" utility, protect themselves and similar agents, and acquire resources.
Obviously, AIs whose goals directly contradict these basic AI drives will strive to avoid them.
External links
- The Basic AI Drives by Stephen M. Omohundro