Edit: Please note that this write-up is poor quality, having the style of a hastily written personal note.

It has been mentioned that there should be a better write-up of open problems in FAI, and as I understand it there is an ongoing effort to explain such open problems. My feeling has been that the recent effort has tended too hold off proposing solutions for too long. I prefer the approach in the Tiling Agents paper, which explained problems through example systems which fail in various respects. What follows is an outline of what I'd write if I spent significant time; I think it is enough to be of some use. This list very much reflects my personal interests & beliefs.

 

- Tarski's Undefinability Theorem

  - We would like a system to be able to reason about itself (in a few critical ways), and Tarski's Theorem is one of the important obstacles. Kripke provided the first hope of progress in this area by showing that we can embed a partial truth predicate in a language if we accept a "gap" (statements which cannot be assessed as true or false within the system's self-theory). Work over the decades has "reduced the gap" (capturing an increasing number of the self-judgements we want, while always leaving a gap). There are also "glut" theories (which must assess some things as both true and false), which typically mirror gap theories. Paul Christiano provided a theory of probabilistic self-reference which intuitively reduces the "gap" to infinitesimal size: the system's knowledge about its own probabilities can be wrong, but only by an infinitesimal. (For example, if it believes X, then it may fail to believe P(X)=1, but it will still believe P(X)>c for all c<1.) (Note, this feels a bit like a "glut" theory since the system solves the problem by saying too much rather than remaining silent.)

- First Incompleteness Theorem

  - Logical Uncertainty:

    - First-Order ("easy"): assigning probabilities to eventual results of (halting) computations we don't have time to make. I claim this is mostly solved by the FOL prior: we can prefer simpler hypotheses about the behavior of systems, treating computations as black boxes which we use universal induction to predict, while allowing us to incorporate logical reasoning about the function's behavior via bayesian updates. It also solves somewhat more; I claim it will have better properties than Solomonoff if the environment contains objects like halting oracles. (Some deficiencies with respect to reasoning about halting will be mentioned in the next section, however.)

    - Second-Order ("impossible"): If we want to assign probabilities to programs halting, facts of number theory, or facts of set theory, we're in serious trouble. Using the FOL prior admits nonstandard models. It's not yet clear what qualities such a probability distribution should have. It seems reasonable to want universal statements to approach probability 1 as the set of positive examples approaches all the examples; this turns out to be as difficult to compute as all the bits in the arithmetic hierarchy. I thought it would be reasonable to restrict this to just the universal statements about computable predicates; ie, halting facts & equivalent. Will Sawin proved that it is not possible for our beliefs about halting to approach arbitrarily close to the correct values without some false Sigma_2 statements approaching arbitrarily close to 1. It remains an open problem to construct such a prior. My proposal is to (focusing on sentences in the regular forms Pi_n or Sigma_n) require that we only introduce a quantified statement into a theory if a statement of the same form already exists; so sentences at a given level in the arithmetic hierarchy must wait for a sentence at the next lowest level to be introduced. This does not block any true sentences from being produced, but causes halting facts to converge as we eliminate possible halting times. It is an open problem whether this proposed distribution converges. If this distribution exists, call it the bad arithmetic prior (BAP).

- Second Incompleteness Theorem

  - Lobian Obstacle:

    For a machine to plan its actions in the future, it needs to trust itself. The second incompleteness theorem (and, more generally, Lob's theorem) makes this difficult. (All this is insufficiently formal, but the tiling agents paper gives a good explanation.) Several partial solutions have been proposed in a deterministic setting. It is an open problem whether one of Dan Willard's several self-verifying systems solves this problem. (He had multiple proposals...) In case there is no purely logical solution, it seems intuitively promising to look for probabilistic self-trust. Difficulties are already presented by the previous section. If the BAP converges, then we can show that it has self-knowledge of that convergence of the form Paul Christiano described! This makes the false Sigma_2 beliefs feel more acceptable, because it is a necessary feature of the system's self-reference. However, I think it's the case that BAP ends up converging to 1 for *all* sigma_2 statements, which is really terrible.

  - Anti-Lobian Obstacle:

    In case the Lobian obstacle is solved, the anti-lobian obstacle may be a concern.

 

New to LessWrong?

New Comment
12 comments, sorted by Click to highlight new comments since: Today at 8:56 PM

For my own records, I'll keep a list of open problems in technical fields here:

Mathematics

Computer science

Is there a short writeup of what you mean by "FOL prior"? Is it the same construction that Benja and Paul came up with, or the one defined in section 5 of Hutter's paper?

Also, do you think that the "5 and 10" family of problems is subsumed by "logical uncertainty"?

I refer to my own construction: ict.usc.edu/pubs/Logical%20Prior%20Probability.pdf

I think the 5 and 10 problem is a feature of a solution to logical uncertainty combined with a technique of making decisions with it.

[-][anonymous]10y00

I see. It looks like many people got the "FOL prior" idea at the same time :-)

[This comment is no longer endorsed by its author]Reply

You begin with mathematics as it currently is. Real numbers are granted, as well as some high level theorems.

I guess, at least something has to give here. Most likely "the reals" and everything based upon them.

Indeed, it seems very difficult to recover natural numbers or real numbers. I take the position that we want to recover as much as we can.

We should preserve as much as possible, but no more. Giving up infinity means, you have no trouble with Godel's incompleteness anymore.

Gödel's incompleteness theorems apply to any system that contains peano arithmetic. Infinity doesn't come into it.

Peano arithmetic implies infinity. The infinite set of all natural numbers.

Only if you have some additional theory that says sets exist and have cardinalities and the set of all natural numbers is a valid set. None of which is assumed by the incompleteness theorem.

In any case, do you have some alternative theory of arithmetic in mind? Being able to form questions that it can't prove an answer to seems like much less of a problem (and a problem that a non omniscient AI needs to solve anyway) than not being able to do arithmetic.

Only if you have some additional theory that says sets exist and have cardinalities and the set of all natural numbers is a valid set. None of which is assumed by the incompleteness theorem.

Peano system demands a successor for every natural. And that 0 (or 1, depends of a variant), is no natural's successor. So the Peano system is only applicable to infinite sets.

do you have some alternative theory of arithmetic in mind?

Sure. The finite number of possible numbers. As in the computer science already is.

Bounded arithmetic is in some sense trivial; it's just a big lookup table. I don't think you're going to get a useful AI that way; if you're going to avoid Godel by limiting the complexity of statements the system can consider, it has to be so simple that it can't form a statement of the form "this statement is not provable" (and not because there's an explicit rule against such statements, the system has to naturally be that simple).