**The
Resolution of Newcomb's Paradox**

**by Chris Langan**

We begin by defining Newcomb's problem. Enter Newcomb's Demon, a rather close
relative of Maxwell's Demon (the similarity goes well beyond a shared affinity
for black boxes). Newcomb's, Demon - call him ND for short - is a paragon of
intelligence, a genius's genius whose acuity transcends the temporal boundaries
of merely ephemeral human beings. Like many geniuses, he displays a certain
playfulness: his "hobby" is testing his own infallibility in the prediction
of human behaviour by making a generous offer to random human playthings. This
offer is couched as a "choice"; though ND's human subjects are in
fact forbidden to predicate their choices on external events, they are tacitly
allowed to use deterministic or nondeterministic internal strategies computed
by corresponding classes of neural events (you may already see the
superficiality of the distinction between inward and outward events; I include
it only as a customary ingredient of the formulation).

Suppose that ND has chosen you as an experimental subject, and that you know that he has never failed to achieve the result he predicted for any subject throughout the extensive history of his experimental run. Nothing appears rigged; in particular, the data are plausibly distributed with respect to the critical behavior of experimental subjects (if Nozick's data are valid, at a ratio of approximately 2.5/1). The experimental set-up is minimal. You are led to a table on which are two boxes. One is transparent, and in it you can clearly see $1000. The other is dead black and totally opaque. ND tells you that he has placed $1,000,000 in the black box if and only if he has predicted that you will take only it and not the other (transparent) box. Otherwise - if he has predicted that you will try to "outsmart" him and take both boxes - the black box has been left empty. ND earnestly assures you that this is no joke; he does not intend to fail. Just in case there are any doubts about this, he offers to take you through a few trial runs using monopoly money and a couple of spare boxes (naturally, you will have to leave the room between these "rehearsals"). You take him up on it. He wins no matter what you think or do. Finally, he tells you that play time is over; next time, the deal is for real. He adds that, lest you suppose he has an infinite amount of time to waste on you, he is giving you a reasonable but finite amount of time to make your move.

Having been around the block, you have already perceived that this is all a kind of game in which ND is the "house", and you his opponent. Of course, since ND is playing for the pure satisfaction of being right while you are playing for big money, a minor conceptual adjustment is in order. But, all things considered, you reason that you should nonetheless apply the standard theory of games in order to maximize your gain. Your reasoning then becomes a little more involved.

You know of certain

But there is a similar rule, called the "dominance

But this seems to constitute a dilemma. These two

As if on cue, your neurons begin again to flicker. It occurs to you that you
could make no decisions whatever, and would know nothing at all, without
relying on various scientific axioms. But axioms are mutually independent, or
virtually random with respect to each other. None of them can be deduced from
other axioms. Doesn't accepting these axioms - including the dominance

You are also roughly as well-informed on what you should do as any professional scientist or logician who has ever published an analysis of Newcomb's paradox. Fortunately, I am no professional, and need grind the favorite axe of no publisher, professor, or employer. The trail that unwinds below is freshly blazed; though a bit steep in spots, it can be negotiated purely on the strength of an open and reasonably sharp mind.

The first step towards resolving the paradox is to provide a logical scaffolding from which Co construct the mathematical model necessary for sound inference. Past arguments involving the problem have used either the standard decision-theoretic model of people playing a game, or the linear "arrow of time" fundamental to classical physics. These two models have been taken to imply opposite solutions, and this suggests that they be somehow unified in an extended "meta-model" which adequately relates the concepts to each other. It has been less obvious just what this higher model should be.

The solution of problems, and the resolution of paradoxes, are inherently computational activities. What, then, could be a better setting for this resolution than a computative one? And what could possibly be a more fitting preface than a brief introduction to the abstract theory of computation?

Consider an acceptor F = (Q, S,d,q_{o},A). Q is a finite nonempty
set of internal states, S is an
alphabet whose symbols are concatenated as strings, q_{o} c Q is the
initial state, and A __c__ Q is the set of accepting states triggered only
by input recognizable to F. The transition mapping d, which governs the way in which F changes states, is
deterministic if d:Q x S —> Q, but nondeterministic if d:Q x S
—> 2^{Q} (where 2^{Q} is the set of all subsets of Q). In
the nondeterministic case, d will be
written d_{n}, for clarity. In
terms of human psychology, we might regard the 5-tuple of F as its
"categorical imperative", or accepting syntax, and say that F
projects this syntax onto its universe. Nothing in the universe of F is
recognizable to it but the particular input strings (sense data, facts) which
cause it to pass through some q e A;
they are its phenomenal "reality", a subset of the noumenal
metareality of the wider universe in which strings are representationally generated
and entered by programmers. The restriction to finite Q is pragmatic and
amenable to conditional relaxation.

If F is deterministic, it accepts (recognizes) a string s e S*
if and only if d(q_{o}, s) e A. Since we have defined d only for the individual symbols s e S, we must define an extended transition
function d': where l is the null string, d'(q,l) = q; and for all
q e Q, s
e S,
and s e S*
(where S* is the set of
varirepetitional permutations of the d e S)
, d'(q,ss)
= d' (d'
(q, s) ,s). Thus, the accepting
behavior of F is defined inductively for s-quantized
string extensions; the way in which we recognize and assimilate new bits of
information within our reality is specified in d'.
Were we to widen the discussion to imagination, conceptualization,
theorization, or other intrinsic computations, we would need to consider
"ideas"; we would have to generalize from recognition to abstraction
by means of a nonlocal or self-iterating, input-free extension of d. If the reference to "strings"
seems to imply a dimensional limitation on input, this too can be generalized.

Where F is nondeterministic, it accepts a string s e S* if and only if d_{n}'(q_{o},s) n A = ~Ø. The
nondeterministic extension d_{n}'
of d_{n} is defined by
induction: d_{n}'(q_{o},
l) = {q} , and d_{n}'(q,ss) = U_{q'c}_{dn'(q,s)} d_{n}(q',s). I.e., q' is one of the possible
successors of q under d_{n}'
given s; the unextended mapping d_{n}
on singletons of S* then determines the
image under d_{n}' of s plus an
adjoint symbol s, given q'. This is a
classical recursive definition. It describes a stepwise probabilistic
ramification of computative potential whose complexity depends on d_{n}.

Nondeterminism is not always restricted to Q x S; under certain conditions, either Q or S can be extended, or A shifted within Q. This. of course,
entails a modification of F, unless F is defined to allow for parametric
extension and adjustment. To this

Having thus formalized the logical abstraction of recognition - i.e., the
passive phase of organo-mechanical cognition - we now proceed to the output
behavior of computative automata, or to the active phase of cognition. Consider
a transducer M = (S,Q,T,d,m),
where S is a finite nonempty input
alphabet, Q a finite nonempty state set, T a finite nonempty output alphabet, d:Q x S
—> Q the state-transition function, and m:
Q x S —> T the output function. A
computation of M has internal and external phases; through m, the output that M delivers back to its
outward universe depends on strings of d-iterated
transitional internal states. Thus, m
is a functional of the function d of input.
Together, m and d totally determine the behavior of M. They can be extended from
S and T to S* and T* as for the acceptor F: d'(q,l) = q, m'(q,l) = l;
and d'(q, ss) = d( d'(q,s),s),
m'(q,ss)
= m'(q,s) m(d'(q,s), s). Where appropriate, we can add to M an
initial state ("reset control") q_{o}: M = (S,Q,T,d',m',q_{o}), to be regarded as a
separator of locally independent computations, and put it at the disposal of a
function r c M which determines computational relevancy.

Considered as a robotic brain, M T-behaves according to m, but Q-reasons towards its decisions along paths generated by d. Where d is deterministic, m may be related to it as a "timing function" according to which any computation can be arrested (input aborted, regression terminated) and converted to output on passing certain tests. Where the duration of the computation is determined with d, input becomes the only variable. Where input as well is fixed in content and scheduling, the entire system is tightly determined. As Laplace might have observed, predicting the behavior of determinate mechanisms requires only data, the means to acquire it, and a valid scientific methodology to organize and interpret it. While the situation is actually more complex, the fact remains that were one to play a deterministic game with a deterministic transducer like M, one would need only a detailed knowledge of its input and programming to predict the outcome, given analytic tools adequate for that purpose and consistent with one's own constraints (e.g., the amount of time available for analysis and strategy). If one's object in the game were merely the validation of one's prediction, so much the easier to win.

Suppose instead that M has nondetenninistic output capping m_{n}, where state-transition may or
may not be deterministic. Then the prediction of output entails control of m_{n} by the predictor. To win a game
of prediction, one must now control m_{n}
as well d_{n} Sn to the extent
chat it is output-critical; one must take over where the probabilistic m_{n} leaves off. Since whichever
control the transducer has over itself resides in m_{n} and d_{n},
one must in

Computation is purposive. The purpose of an acceptor is pure recognition; no action is explicitly predicated on its internal transitions. The purpose of a transducer is conversion of input to output; yet, such a conversion is aimless unless algorithmic. Like yin and yang, acceptors and transducers are complementary; only together can they begin to resemble functional systems of organic complexity. In order to model organic systems, transducers must be endowed with goals and algorithms comparable to the ends and means of living beings. Algorithms are themselves purposive procedures which model both acceptance and transduction. The problems which comprise their input are scanned by preliminary steps for certain kinds of information, which must in turn be accepted as parameters by subsequent steps, and so on to the output stage (at which point the algorithm delivers its answer). The mechanistic representation of an algorithm must allow for the innate structure of a device, considered apart from the algorithm itself; this structure may have variant and invariant aspects. The algorithm simply conforms variables to purpose given the invariants. As the definitions of F and M might lead one to expect, this generally involves importing to M the set A c Q defined for F.

Human beings, it is said, are self-programming. Their thought is polyalgorithmic; useful algorithms are either meta-algorithmically constructed, or selected from a learned store, to deal with input. If learning, construction, and selection are deterministic, then they characterize a deterministic meta-algorithm not fundamentally different from any other deterministic algorithm we might study. If they are nondeterministic, then they are characteristic of a nondeterministic meta-algorithm, and likewise. It follows that the formal transductive model of human nature withstands any objection from the relative complexity of human mentality or behavior.

Newcomb's object-transducer M_{N} naturally includes an acceptor: M_{N}
(S,Q,q_{o},d,A,m,T).
Recognition is phasic; a string must often be "pre-accepted" for M to
tell whether to accept or reject it. Ordinarily, this tentative phase of
recognition is easily computed by the physical entities whose behavior is
predicted by ND. To be "real", an input-quantum s must simply possess a certain first-order
predicate, "reality", which - this being a self-validating tautology
- induces a type-theoretic predicate stratification like that involving the old
Cretan, Epimenides. To this sine qua non of recognition corresponds a primary
element q_{1} of A; no input-quantum failing the q_{1}-test is
reified, whereas all those passing are relayed to Q - q_{1}.
Higher-order recognition of "passed" quanta then proceeds at a rate
determined by the respective computational demands of the
stratified-algorithmic phases of d_{M}.
Corresponding to the structure of A are various ordered states analogous to q_{1}
within their respective levels of acceptance.

Let us narrow the definition of M_{N} in a way consistent with
Newcomb's problem. Suppose that associated with m_{M}
is a threshhold value b > 0 below
which output is nil, but above which a decision will be finalized and
implemented. With each q_{i} c Q we associate a pair of strength
coefficients a_{i}, and a_{i2}, to be incremented and
decremented according to a strategic d_{M},
appropriate to the Newcomb decision-theoretic context; these represent the
current tendency, given the present amount of input, for M_{N} to
output either possible behavior (taking one or both boxes, respectively). The a_{i}
divide Q into three classes X^{Q}, Y^{Q}, and Z^{Q},
with membership conditions a, > a_{2}, a, < a_{2}, and a,
= a_{2} respectively. To each q_{i} is attached a total weight a_{i} = |a_{i}, - a_{i2}|.
As soon as the output condition [a = b] is met, a decision and behavior result
which correspond to the Q-class of the current state (note that d_{M}, being strategic, precludes q c
Z^{Q}, or "indecision", at the output deadline). Thus, the
states of Q c M_{N} are preferential and impetal, the graded
fore-images of the outputs they favor. This is just a convenient way to view
the internal configurations to which they correspond, and does not violate the
general definition of transducers. Nor, for that matter, does it violate the
way human beings perceive their own decision-making processes.

The question posed by Newcomb's problem involves the computative analysis,
by a predictive agency with computative characteristics, of the computative
analysis undertaken by a transducer on a given input. That input is the problem
itself, presented in the manner prescribed by the formulation. This situation,
which defines a computative regression, is recursive and inductively extensible.
The regression in turn defines the only soluble context for the higher-level
"paradox" generated by the problem. This context translates as
mechanism. The mechanism is a stratified automaton G containing both the predictor and its object-transducer as
sub-automata. Whether "free will" is defined determlnistically as
mere outside non-interference in m and d, or nondeterministically as the ability of
M_{N} to override any exogenous restriction of m_{n} or d_{n},
its mechanism is contained in that of G.

Logical diagonalization of the formal computational language generated by
the accepting syntax of M_{N} directly implies that certain structural
aspects of G may be unrecognizable to M_{N}.
In particular, those aspects involving M_{N}-relativized
nondeterminacy, as well as those involving certain higher-order predicates of
the nondistributive, nonlocal organizations involving m_{M} and d_{M},
are formally undecidable to it and need not be recognized directly by it with
any degree of specificity. To understand why, consider the extent to which a
common computer "recognizes" the extended system including its cpu,
its other components, its programmers, and the environment it inhabits. In
fact, it can recognize nothing that does not conform to its input-to-output
transformational grammar. Even if it were self-analytic, such analysis could be
limited to a schematic electronic syntax which overlooks the material elements
of which it is constructed. In any case, it can make sense of nothing but
strings of input translated and rearranged according to the internal
stratification of its hard and soft programming.

You, your purposes, and your dependencies are undecidable to it, and so are
the mechanisms by which you can predict and control its behavior. It matters
not who formulates this undecidability; if the machine's internal logic is
inadequate to do so, yours surely is not (currently, most mechanical acceptors
are nongeneralistic, treating complementation as negation and negation as
rejection; this bars the tools of diagonalization from their computations).
Should it ignore your higher prerogatives, you could "diagonalize" it
- if nothing extrinsic to the machine were to stop you - with a sledgehammer
whose _{N}
cannot preclude G on grounds of
"insensibility". Nor, for chat matter, can we.

There are many self-styled experts on undecidability who have expressed the opinion that all attempts to reify Godel's theorem along paranormal lines reflect a misunderstanding of its "real nature". Such experts are quite correct in that a misunderstanding exists, but the misunderstanding is all theirs. What the theorem forces by juxtaposing truth and derivability (or consistency and completeness) is a hierarchical stratification of classes of truth functions and the inferential syntaxes which parametrize them. This stratification follows that of G, fractionating computative reality along with the "truth" to which it corresponds.

The stratification of G induces
stratum-relativizations of computative time and space. Thus, the timetype in
which M_{N} computes recognition and output is a mere subtype of that
in which it is programmed. Dynamical "arrows of determinacy" which
are inviolable to M_{N}, being programmed into its accepting syntax,
have no force whatsoever to the programmatic agencies themselves. This applies
just as well to "metrical" restrictions embodied in the M_{N}-syntax;
these may allow M_{N} to recognize nothing but an artificial submetric
of the metric in which these agencies define their own existence. M_{N}
and its reality might consist of quanta with higher-dimensional interpretations
as the termini of channels for the transmission of information between strata.
Metatemporal predicates may exist with respect to which those of M_{N},
are definite only in a mutual sense; predicates which M_{N} accepts as
"before" and "after" could be the programmatic projections
of "in front of" and "in back of", or any other G-consistent higher-prepositional
relationships.

There can thus exist a mechanism x c G
through which a predictor like ND could measure and/or control the mappings d, m
c M_{N} in ways directly insensible to M_{N}. Where in G relative to M_{N} would such a
predictor have to be located? Precisely where access is available.
Sirnplistically, we might characterize the predictor-M relationship as one of
proper inclusion, where it is understood that prediction is direct rather than
second-hand, and programmatic in the passive and active senses. That is, a
programmer mentally internalizes the structure of that which he programs, and
this internalization amounts to computative inclusion. The fine structure of G, while to a degree analytic, is a natter of
some complexity. For now, it will suffice to have demonstrated the possibility
of x and its utility to well-situated G-subautomata.
Because G is structured to allow for
relativized deteririnacy and nondeterminacy, the solution is invariant with
respect to argumentation involving mind-brain dichotomy. That is, such
dichotomies reduce to distinctions of determinacy and nondeterminacy, and may
be treated in kind.

Restricted dominance, which relies on probabilistic independence derived
from the lower-order, local istic dynamical timetype of M_{N}'s
artificially restricted "reality", is revealed under G-extension to be 'itself dominated by
utility. That is, the subjective utility of M_{N} forces the
assimilation by d_{M} of this
entire demonstration, which disables restricted dominance and thus frees the
strategic component to recognize higher patterns among observed data. The _{N} remains unbreached, loses
all force in the presence of exodynamic influence.

Let's sum it up. You can be modeled as a deterministic or nondeterministic transducer with an accepting syntax that can be diagonalized, or complemented by logical self-negation. ND can be modeled as a metalogical, metamechanistic programmatic agency, some insensible part of whom surrounds you in a computational space Including physical reality, but not limited thereto. This space is the mechanistic equivalent of the computative regression around which Newcomb's problem is essentially formulated. The existence of this space cannot be precluded by you on the grounds that you cannot directly observe it, nor can it be said by you to deny ND a mechanism of control and prediction of your thought and behavior. Additionally, you have an open-ended run of data which lowers to 1/ 8 the probability that NO is "just lucky". This implies that mechanism does indeed exist, and warrants the adjunction to the axioms of physics an independent, empirical physical axiom affirming that mechanism. This then implies that ND can predict or control human thought and behavior (a somewhat weaker implication, you will notice, than "omniscience"). ND possesses means, motive, opportunity...and you. You are "possessed" by Newcomb's Demon, and whatever self-interest remains to you will make you take the black box only. (Q.E.D.)

Do not let incredulity cheat you of understanding. It may be hard to accept the idea of a mathematical proof of the possibility of demons; if so, you may take solace from the fact that this does not of itself imply their actual existence. Remember that Newcomb's problem includes a hatful of hypothetical empirical data that may have no correspondents in your sector of reality. The proof just given merely prevents you from generally precluding the paranormal experiences reported by others on logical grounds. That is, it restrains you from any blanket distribution of too narrow a brand of "rationalism" over reality. Because I have long been exploring the ramifications of the logic this proof employs, I already know how easy it would be to deal with whomever might dispute it on "scientific" or other grounds, regardless of their prestige or reputed intelligence. There is a certainly a strong correlation between rationalistic dogma and the persecution of its foes. Fortunately, it is not too late for "rationalism" to catch up with the advances of twentieth-century logic.

Regarding free will, consider that demons are generally reputed to offer a
material reward for the surrender of one's soul. Where we define
"soul" as the autoprogrammatic G-extension
M' of one's characteristic transductive representation M, this means only that
one is rewarded for voluntarily moving aside and letting the demon take over as
captain of one's fate. He does this by adjoining to his subject any functional
extensions he requires; e.g., d(d_{n}),
or m(m_{n}). Or, if restricted
determinism prevails but he is not in an especially predictive mood, he might
create a parametric extension F_{ND}' (analogous to F') enabling a
hyperdeterninistic override m ? m of
one's deterministic output function. Actually, volition may be irrelevant; one
might be shoved aside rather than nudged. Where freedom is defined negatively
on restricted determinacy, it is duly stratified.

It has by now occurred to most readers that "demon" is a darkly connoted term that could just as well be replaced with "guardian angel" or something equally comforting. After all, ND does not ask his subjects to commit evil, but only to let him do them a rather large favor with no apparent strings attached. Insisting on "free will" here is reminiscent of the child who stares defiantly at its mother while touching the hot stove she has just finished telling it to avoid. If the child's act has any meaning at all, it is just that poorly timed independence can lead to more painful dependency down the line (surprisingly, Isaac Asimov advocates just this kind of defiance in the Newcomb context, and on what he seems to feel are ethical grounds; but we will deal with ethics later). One cannot be willingly "enslaved" by ND before being enslaved by the money he offers. Who among us always behaves in willful defiance of economic reality? My heart goes out to him, for he is either "enslaved" by an institution, or he is at the absolute mercy of the spontaneous, unsolicited charity of those whose own demons generally'predispose them to pass him by.

Having quenched the fire, we will now "mop up". Professor Nozick himself has remarked on the possible nonrecursivity of scientific laws involved in the determination of behavior. This, he surmises, prevents the inference of predictability from determinism. But this conclusion entails the confinement of these predicates within a restricted syntax, whereas the extensionality of G renders such a constraint powerless. In other words, restricted recursivity is beside the point for agencies defined to possess sufficient access to the channels through which such laws project as restricted reality, and particularly to their source. All ND has to do is run a G-presimulation of interstratum transmission. Note that the computat ive requirements for tills, which correspond to assumptions on the nature of G, are satisfied given ND's efficacy.

Nozick also remarks on the regression of determinism to self-validating premises. Self-validating languages comprise their own inferential schemata. Validation, being inferential to the extent of quantification, follows the G-stratification of inference. So self-validation is also stratified, and such languages correspond to appropriate levels of computative reality. For sufficiently expressive languages, self-validation becomes diagonalistic self-vitiation. Determinism is thus stratified, and so is its counterpart, freedom. The seemingly infinite regression can be brought to closure at an ultimate form of undecidability, which in turn can be locally resolved with a relativized degree of confirmation.

Regarding various remarks that have been submitted to Nozick by members of
the physics community, I should explain that the above resolution has a
metaphysical generalization of which certain implications may broadly be termed
physical. This generalization, as developed by the author, is called the CTMU,
or Computation-Theoretic Model of the Universe (the acronym has a mnemonic
pronunciation: '*cat-mew*). We will later examine chosen aspects of this
model, which transcends characterization as a "theory" by virtue of
its categoricality with respect to computative entities such as human beings.
That is, the CTMU is as close to "absolute truth" as we will ever be
privileged to get along computative or intellectual pathways. If I might be
indulged a bit of testimony: beyond the CTMU ultimate syntax, the lattice which
gives shape to the multilayered layered veil of maya, there is but the light
that shines forever.

Christopher Michael Chappelle-Langan

(copyright 1990 by C. M. Langan. No portion of this article is to be reproduced in any form without the author's written consent.)

Bibliography:

__Newcomb's Problem and Two Principles
of Choice__

Robert Nozick in

Maya Bar-Hillel and Avishai Margalit in

Martin Gardner and Robert Nozick in the Mathematical Games Department of

Copyright © 1989 by the Mega Society. All rights reserved. Copyright for each individual contribution is retained by the author unless otherwise indicated.