Comment author: lukeprog 21 June 2013 09:18:59PM 9 points [-]

General note to everyone: I recommend composing posts in something that converts easily to clean HTML. I always write my posts in markdown, then convert to HTML with the Markdown Dingus, then paste it into the HTML editor in the LW post editor.

Comment author: loup-vaillant 25 June 2013 03:40:11PM 3 points [-]

By the way, why posts aren't written like comments, in Markdown format? Could we consider adding markdown formatting as an option?

Comment author: solipsist 19 June 2013 12:38:07PM *  3 points [-]

It still seems to me that you can't have a BestDecisionAgent. Suppose agents are black boxes -- Omegas can simulate agents at will, but not view their source code. An Omega goes around offering agents a choice between:

  • $1, or
  • $100 if the Omega thinks the agent acts differently than BestDecisionAgent in a simulated rationality test, otherwise $2 if the agent acts like BestDecisionAgent in the rationality test.

Does this test meet your criteria for a fair test? If not, why not?

Comment author: loup-vaillant 20 June 2013 10:37:23AM *  1 point [-]

I think I have left a loophole. In your example, Omega is analysing the agent by analysing its outputs in unrelated, and most of all, unspecified problems. I think the end result should only depend on the output of the agent on the problem at hand.

Here's a possibly real life variation. Instead of simulating the agent, you throw a number of problems at it beforehand, without telling it it will be related to a future problem. Like, throw an exam at a human student (with a real stake at the end, such as grades). Then, later you submit the student to the following problem:

Welcome to my dungeon. Sorry for the headache, but I figured you wouldn't have followed someone like me in a place like this. Anyway. I was studying Decision Theory, and wanted to perform an experiment. So, I will give you a choice:

Option 1 : you die a most painful death. See those sharp, shimmering tools? Lots of fun.

Option 2 : if I think you're not the kind of person who makes good life decisions, I'll let you go unharmed. Hopefully you will harm yourself later. On the other hand, if I think you are the kind of person who makes good life decisions, well, too bad for you: I'll let you most of you go, but you'll have to give me your hand.

Option 2? Well that doesn't surprise me, though it does disappoint me a little. I would have hoped, after 17 times already… well, no matter. So, do you make good decisions? Sorry, I'm afraid "no" isn't enough. Let's see… oh, you're you're applying for College, if I recall correctly. Yes, I did my homework. I'm studying, remember? So, let's see your SAT scores. Oh, impressive. That should explain why you never left home those past three weeks. Looks like you know how to trade off short term well being for long term projects. Looks like a good life decision.

So. I'm not exactly omniscient, but this should be enough. I'll let you go. But first, I believe you'll have to put up with a little surgery job.

Sounds like something like that could "reasonably" happen in real life. But I don't think it's "fair" either, if only because being discriminated for being capable of taking good decisions is so unexpected.

Comment author: solipsist 19 June 2013 05:27:09AM 0 points [-]

Newcomb's problem seemed like a contrived trick to punish CDT, and it seemed that any other decision theory was just as likely to run into some other strange scenario to punish it, until I started thinking about AIs that could simulate you accurately

Are you implying there exist decision theories that are are less likely to run into strange scenarios that punish it? I would think that Omegas could choose to give any agent prejudicial treatment.

Comment author: loup-vaillant 19 June 2013 08:44:06AM *  3 points [-]

We have to determine what counts as "unfair". Newcomb's problem looks unfair because your decision seems to change the past. I have seen another Newcomb-like problem that was (I believe) genuinely unfair, because depending on their decision theory, the agents were not in the same epistemic state.

Here what I think is a "fair" problem. It's when

  1. the initial epistemic state of the agent is independent of its source code;
  2. given the decisions of the agent, the end result is independent of its source code;
  3. if there are intermediary steps, then given the decisions of the agent up to any given point, its epistemic state and any intermediate result accessible to the agent at that point are independent of its source code.

If we think of the agent as a program, I think we can equate "decision" with the agent's output. It's harder however to equate "epistemic state" with its input: recall Omega saying "Here is the 2 usual boxes. I have submitted this very problem in a simulation to TDT. If it one boxed, box B has the million. If it two boxed, box B is empty". So, if you're TDT, this problem is equivalent to the old Newcomb problem, where oneBox <=> $$$. But any other agent could 2 box, and get the million and the bonus. (Also, "TDT" could be replaced by a source code listing that the agent would recognize as its own.)


Anyway, I believe there's a good chance a decision theory exists such that it gets the best results out of any "fair" problem.Though now that I think of it, condition 2 may be a sufficient criterion for "fairness", for the problem above violates it: if TDT two-boxes, it does not get the million. Well except it does not two box, so my counter-factual doesn't really mean anything…

Comment author: bogdanb 15 June 2013 07:48:12PM *  1 point [-]

Oh, OK. In that case, what you are trying to achieve is (theoretically) boxing a (potential) AGI, without a gatekeeper. Which is kind of overkill in this case, and wouldn’t be solved with a choice of language anyway :)

Comment author: loup-vaillant 16 June 2013 09:37:19PM 4 points [-]

I think it is possible to prove that a given boxing works, if it's sufficiently simple. Choosing the language isn't enough, but choosing the interpreter should be.

Take Brainfuck for instance: replace the dot ('.'), which prints a character, by two other statements: one that prints "yes" and exits, and one that prints "no" and exits. If the interpreter has no bug, a program can only:

  • Print "yes" and kill itself.
  • Print "no" and kill itself.
  • Do nothing until we kill it, or otherwise fail.

Assuming the AI doesn't control the external world by heating the host Intel processor in a smart way, we should be able to prove that we're otherwise safe.

Comment author: bogdanb 07 June 2013 11:08:11PM 0 points [-]

I’m sorry, how is that relevant to my no-network-adapter comment? (I mean this literally, not rhetorically. I don’t see the connection. Did you mean to reply to a different comment?)

Comment author: loup-vaillant 08 June 2013 08:33:42AM *  0 points [-]

It's the whole thread. I was not sure where to place my comment. The connection is, the network may not be the only source of "cheating". My solutions plug them all in one fell swoop.

Comment author: bogdanb 07 June 2013 07:16:06PM *  0 points [-]

Yes, but then you’re in a very bad position if the test is run without network access. (I.e., you’re allowed to use the network, but there’s no network adapter.)

Comment author: loup-vaillant 07 June 2013 07:57:18PM 0 points [-]

Well, I just though about it for 2 seconds. I tend to be a purist: if it were me, I would start from pure call-by-need λ-calculus, and limit the number of β-reductions, instead of the number of seconds. Cooperation and defection would be represented by Church booleans. From there, I could extend the language (explicit bindings, fast arithmetic…), provide a standard library, including some functions specific to this contest.

Or, I would start from the smallest possible subset of Scheme that can implement a meta-circular evaluator. It may be easier to examine an S-expression in Scheme than a λ-expression in λ-calculus.

Or, I would start from lambda calculus with de-Brujin indices, so we don't have to worry about α-conversions. In this case, I would provide a compiler from regular λ-calculus. (I suspect however that this one doesn't change a thing.)

Or, I would start from a Forth dialect (probably with an implicit return stack).

Or, I would start from BrainFuck (only half joking: the language is really dead simple, and fast interpreters already exist).

But it's not me, and I don't have the courage right now. If I ever implement the necessary tools (which I may: I'm studying programming languages in my spare time), then I will submit them here, and possibly run a contest myself.

Comment author: AlexMennen 05 June 2013 09:38:22PM 1 point [-]

More generally, the set of legal programs doesn't seem clearly defined.

I haven't forbade use of any library functions except for file IO. I'm not confident this is optimal, but how is it underspecified?

Comment author: loup-vaillant 06 June 2013 08:34:43AM *  1 point [-]

Okay, it's not. But I'm sure there's a way to circumvent the spirit of your rule, while still abiding the letter. What about network I/O, for instance? As in, download some code from some remote location, and execute that? Or even worse, run your code in the remote location, where you can enjoy superior computing power?

Comment author: Lumifer 05 June 2013 08:00:45PM 4 points [-]

Will these lambda functions have access to external resources, notably a random-number generator (or a source of entropy to implement it internally)?

Comment author: loup-vaillant 05 June 2013 09:05:19PM *  3 points [-]

More generally, the set of legal programs doesn't seem clearly defined. If it were me, I would be tempted to only accept externally pure functions, and to precisely define what parts of the standard library are allowed. Then I would enforce this rule by modifying the global environment such that any disallowed behaviour would result in an exception being thrown, resulting in an "other" result.

But it's not me. So, what exactly will be allowed?

Comment author: NancyLebovitz 02 June 2013 09:18:40AM 0 points [-]

Scared of?

Comment author: loup-vaillant 02 June 2013 10:17:23AM *  3 points [-]

Hmm, leaving everything and everyone behind, and a general feeling of uncertainty: what live will be like? Will I find a job? Will I enjoy my job (super-important)? How will this affect my relationship with my SO? Less critically, should I bring my Cello, or should I buy another one? What about the rest of my stuff?

We're not talking moving a couple hundred miles here. I've done it for a year and, I could see my family every 3 week-ends, and my SO twice as much. Living in Toulouse, France, I could even push to England if I had a good opportunity. But to go to the US, I have to Cross the Ocean. If I leave this summer and find a job by September, I likely won't make a single trip back before the next summer.

Also, I don't think I value money all that much. I mainly care about the sense of security it provides. If I were guaranteed half of what I currently make to work at home on the computer science research that I want to do, I would take it.

So, If I were to move to the US, it couldn't be just about the money. The job matters. And I'd better get closer to the LW-MIRI-CFAR community. And even then, I'm still not sure. Indefinitely postponing such a big decision is so easy.

Comment author: John_Maxwell_IV 01 June 2013 03:13:07AM *  9 points [-]

Given that your name looks familiar from Hacker News and your website suggests you like programming for its own sake, you should consider coming to Silicon Valley after the US congress finishes loosening up immigration restrictions for foreign STEM workers (which seems like it will probably happen). In the San Francisco area, $100K + stock is typical for entry-level people and good programmers in general are famously difficult to hire. Also, lots of LW peeps live here. My housemates and I ought to have a couch you can crash on while you look for a job. In the worst case it'll just be a vacation for you to visit the US and hang out with the LW people that live around here. PM me if you want more info and stuff.

(This also goes for other people who are good programmers who don't live in Silicon Valley and don't have a good reason not to move here.)

Comment author: loup-vaillant 01 June 2013 10:21:40AM 1 point [-]

(Yep, I'm loup-vaillant on HN too)

Thank you, I'll think about it. Though for now, seriously considering moving to the US tends to trigger my Ugh shields. I'm quite scared.

View more: Prev | Next