In acausal trade, agents P and Q cooperatre as follows: P simulates or otherwise analyzes agent Q and learns that Q does something that P wants if P does what Q wants; and Q symmetrically learns the same about P; and they both know can prove this about the other.
We have agents P and Q, possibly space/time separated so that no interaction is possible. They might be in separate Everett Branches; or they might be outside each other's light cones, e.g., they might be very distant in an expanding universe, or Q might be a billion years in A's future on Earth. Or on the other hand they might be able to communicate, as in ordinary trade.
P and Q each can do something the other wants, and value that thing less than the other does. This is the usual condition for trade. This can happen even if P and Q are completely separated. One example: An FAI wants the well-being of all humans, even those it cannot be in causal contact with.
In acausal trade, P and Q cannot count on the usual enforcement mechanisms to ensure cooperation, e.g., an expectation of future interactions or an outside enforcer.
P and Q will cooperate because each knows that the other can somehow predict its behavior very well, like Newcomb's Omega.
Each knows that if it defects (respectively: cooperates), the other will know this and defect (respectively: cooperate), and so the best choice is to cooperate, since, as usual in trade, Cooperate/Cooperate is better for both sides than Defect/Defect.
This can also be described as P and Q provably precommitting to cooperate.
This scenario does not specify how P and Q can predict each other.
- They might have each other's source code.
- Or they might be able to simulate each other, or to predict the other's behavior analytically. The simulation may be approximate and make probabilistic predictions to avoid the problem of simulating something of the same complexity of oneself in full.
- In particular, P and Q might have identical or similar mental architectures, so that the each one knows that its own mental processes approximately simulate the other's. See Gary Drescher's acausal subjunctive cooperation.
At its most abstract, agents P and Q in this model are simply optimization algorithms, and acausal trade is simply one subalgorithm that is often improves optimization power. Let P be an algorithm for which time is most valuable in optimizing its utility function; while for Q, space is most valuable.
P and Q must both deal with subjective uncertainty as to whether they will be instantiated in a universe where time is in abundance (a universe which will not end for a long time), or a space-rich universe. (This is sometimes expressed in terms of multiverse theory, but only requires that we specify subjective uncertainly, i.e., that the algorithms be able to optimize in either kind of universe.)
P can prove about Q that it will "do a favor" for P, i.e., that it will run P if it is instantiated in a time-rich universe. P can also prove that Q will do so because Q can prove that P will "reciprocate"-- i.e., P will run Q in a space-rich universe, in acausal trade for own proven Q's commitment.
P and Q also estimate that the other is likely to exist. This is because super-powerful intelligences (optimizers) are an attractor in any universe that has a general intelligence at all, since any general intelligence [Intelligence Explosion] will want to self-improve in order to better optimize towards its goals. Given that superintelligences exist in other possible universes, P and Q will realize that one efficient technique in a super-optimizer is acausal trade, thus implying the existence of the superintelligent acausal trading partner.
Algorithms P and Q are so structured because, as we have specified, they are good optimizers, and acausal trade is a useful technique in certain conditions for optimizing towards each algorithms own goals: Acausal trade is one of a trick a good optimizer can use to optimize utility functions.
Relevance to Friendly AI
Acausal trade may be a way to get the cooperation of a future AI. If we know that the AI will want us to behave a certain way, and we can prove that it will be Friendlier, once it arises, if we do what it wants now; and it can symmetrically prove that we do what it wants, if we've proven this behavior about it, then we can trade with it even though it does not yet exist.
This approach rests on being able to prove certain facts about human behavior; or similarly, for humans to be able to provably commit to behavior.
Alexander Kruel, http://kruel.co/2012/07/27/the-acausal-trade-argument/ "The Acausal Trade Argument"].