Back to LessWrong

Difference between revisions of "Wireheading"

From Lesswrongwiki

Jump to: navigation, search
(Created page with 'Wireheading is the direct stimulation of an individual's brain's reward center with electrical current. In both thought experiments and [laboratory experiments](http://www.mindha...')
 
 
(18 intermediate revisions by 8 users not shown)
Line 1: Line 1:
Wireheading is the direct stimulation of an individual's brain's reward center with electrical current. In both thought experiments and [laboratory experiments](http://www.mindhacks.com/blog/2008/09/erotic_selfstimulat.html), this makes the individual feel happy. In theory, wireheading with a powerful enough current would be the most pleasurable experience imaginable.
+
{{hpp}}
 +
'''Wireheading''' is the artificial stimulation of the brain to experience pleasure, usually through the direct stimulation of an individual's brain's reward or pleasure center with electrical current. It can also be used in a more expanded sense, to refer to any kind of method that produces a form of ''counterfeit utility'' by directly maximizing a good feeling, but that fails to realize what we value.
  
This raises difficult ethical questions for those who believe that morality is based on human happiness. A civilization of wireheads "blissing out" all day while being fed and maintained by robots would be a state of maximum happiness, but such a civilization would have no art, love, scientific discovery, or any of the other things humans find valuable. Anyone who would not value such a future for humanity, must consider a [[complexity of values]] stretching beyond simple happiness.
+
In both thought experiments and [http://www.mindhacks.com/blog/2008/09/erotic_selfstimulat.html laboratory experiments] direct stimulation of the brain’s reward center makes the individual feel happy. In theory, wireheading with a powerful enough current would be the most pleasurable experience imaginable. There is some evidence that [http://lesswrong.com/lw/1lb/are_wireheads_happy/ reward is distinct from pleasure], and that most currently hypothesized forms of wireheading just motivate a person to continue the wirehead experience, not to feel happy. However, there seems to be no reason to believe that a different form of wireheading which does create subjective pleasure could not be found. The possibility of wireheading raises difficult ethical questions for [[Hedonism|those who believe that morality is based on human happiness]]. A civilization of wireheads "blissing out" all day while being fed and maintained by robots would be a state of maximum happiness, but such a civilization would have no art, love, scientific discovery, or any of the other things humans find valuable.
  
Wireheading is also an illustration of the complexities of creating a Friendly AI. Any AI naively programmed to increase human happiness could devote its energies to wireheading people, possibly without their consent, in preference to any other goals. Equivalent problems arise for any simple attempt to create AIs who care directly about human feelings ("love", "compassion", "excitement", etc).
+
If we take wireheading as a more general form of producing counterfeit utility, there are many examples of ways of directly stimulating of the reward and pleasure centers of the brain, without actually engaging in valuable experiences. Cocaine, heroin, cigarettes and gambling are all examples of current methods of directly achieving pleasure or reward, but can be seen by many as lacking much of what we value and are potentially extremely detrimental. [[Wikipedia:Steve Omohundro|Steve Omohundro]] argues[http://selfawaresystems.files.wordpress.com/2008/01/ai_drives_final.pdf] that: “An important class of vulnerabilities arises when the subsystems for measuring utility become corrupted. Human pleasure may be thought of as the experiential correlate of an assessment of high utility. But pleasure is mediated by neurochemicals and these are subject to manipulation.
  
There is some evidence that [reward is distinct from pleasure](http://lesswrong.com/lw/1lb/are_wireheads_happy/), and that most currently hypothesized forms of wireheading just motivate a person to continue the wirehead experience, not to feel happy. However, there seems no reason to believe that a different form of wireheading which does create subjective pleasure could not be found.
+
Wireheading is also an illustration of the complexities of creating a [[Friendly AI]]. Any AGI naively programmed to increase human happiness could devote its energies to wireheading people, possibly without their consent, in preference to any other goals. Equivalent problems arise for any simple attempt to create AGIs who care directly about human feelings ("love", "compassion", "excitement", etc). An AGI could wirehead people to feel in love all the time, but this wouldn’t correctly realize what we value when we say love is a virtue. For Omohundro, because exploiting those vulnerabilities in our subsystems for measuring utility is much easier then truly realizing our values, a wrongly designed AGI would most certainly prefer to wirehead humanity instead of pursuing human values. In addition, an AGI itself could be vulnerable to wirehead and would need to implement “police forces” or “immune systems” to ensure its measuring system doesn’t become corrupted by trying to produce counterfeit utility.
 +
 
 +
==Blog posts==
 +
*[http://lesswrong.com/lw/1lb/are_wireheads_happy/ Are wireheads happy?]
 +
*[http://lesswrong.com/lw/15c/would_your_real_preferences_please_stand_up/ Would Your Real Preferences Please Stand Up?]
 +
*[http://lesswrong.com/lw/1oc/you_cannot_be_mistaken_about_not_wanting_to/ You cannot be mistaken about (not) wanting to wirehead]
 +
 
 +
==See also==
 +
*[http://selfawaresystems.files.wordpress.com/2008/01/ai_drives_final.pdf Steve Omohundro's paper, section 4 deals with vulnerabilities to counterfeit utility and wireheading]
 +
*[[Hedonism]]
 +
*[[Complexity of value]]
 +
*[[Wanting and liking]]
 +
*[[Near/far thinking]]
 +
*[[Hedonium]]
 +
*[[Abolitionism]]

Latest revision as of 21:56, 19 June 2017

H+LogoHR.png This article might be suitable to merge into H+Pedia.
Discuss

Wireheading is the artificial stimulation of the brain to experience pleasure, usually through the direct stimulation of an individual's brain's reward or pleasure center with electrical current. It can also be used in a more expanded sense, to refer to any kind of method that produces a form of counterfeit utility by directly maximizing a good feeling, but that fails to realize what we value.

In both thought experiments and laboratory experiments direct stimulation of the brain’s reward center makes the individual feel happy. In theory, wireheading with a powerful enough current would be the most pleasurable experience imaginable. There is some evidence that reward is distinct from pleasure, and that most currently hypothesized forms of wireheading just motivate a person to continue the wirehead experience, not to feel happy. However, there seems to be no reason to believe that a different form of wireheading which does create subjective pleasure could not be found. The possibility of wireheading raises difficult ethical questions for those who believe that morality is based on human happiness. A civilization of wireheads "blissing out" all day while being fed and maintained by robots would be a state of maximum happiness, but such a civilization would have no art, love, scientific discovery, or any of the other things humans find valuable.

If we take wireheading as a more general form of producing counterfeit utility, there are many examples of ways of directly stimulating of the reward and pleasure centers of the brain, without actually engaging in valuable experiences. Cocaine, heroin, cigarettes and gambling are all examples of current methods of directly achieving pleasure or reward, but can be seen by many as lacking much of what we value and are potentially extremely detrimental. Steve Omohundro argues[1] that: “An important class of vulnerabilities arises when the subsystems for measuring utility become corrupted. Human pleasure may be thought of as the experiential correlate of an assessment of high utility. But pleasure is mediated by neurochemicals and these are subject to manipulation.”

Wireheading is also an illustration of the complexities of creating a Friendly AI. Any AGI naively programmed to increase human happiness could devote its energies to wireheading people, possibly without their consent, in preference to any other goals. Equivalent problems arise for any simple attempt to create AGIs who care directly about human feelings ("love", "compassion", "excitement", etc). An AGI could wirehead people to feel in love all the time, but this wouldn’t correctly realize what we value when we say love is a virtue. For Omohundro, because exploiting those vulnerabilities in our subsystems for measuring utility is much easier then truly realizing our values, a wrongly designed AGI would most certainly prefer to wirehead humanity instead of pursuing human values. In addition, an AGI itself could be vulnerable to wirehead and would need to implement “police forces” or “immune systems” to ensure its measuring system doesn’t become corrupted by trying to produce counterfeit utility.

Blog posts

See also