whpearson comments on My problems with Formal Friendly Artificial Intelligence work - Less Wrong
You are viewing a comment permalink. View the original post to see all comments and the full post content.
You are viewing a comment permalink. View the original post to see all comments and the full post content.
Comments (26)
2+2=4 pays rent.
Ugh can use it right away for counting days. Source code based decision theory not so much. There aren't the societies based on agents that can read each others source code, so I can't try and predict them with source code based decision theories. It seems like it is mathematically interesting thing though, so it is still interesting. I just don't want it to be a core part of our sole pathway to try and solve the AI problem.
Perhaps it would lead people to avoid trying to find optimal decision theories and accept the answer that the best decision theory depends on the circumstances. Then we can figure out what our circumstances are and find good decisions theories for those. And create designs that can do similarly.
Like the best search algorithm is context dependent, where even algorithms of a worst complexity class can be better due to memory locality and small size.
Supposition> If answers to how decisions are made (and a whole host of other problems) are contextual and complex then it is worth trading information about what answers they have found within their context.
This pathway makes no sense from people that expect there to be winner take all optimal AI designs as any one embryonic system might find the keys to the future and take it over. But if that is not the way the world works....
There are mathematical concepts that didn't pay rent immediately: imaginary numbers, quaternions, non-Euclidean geometry, Turing machines...
But thanks for the specific plan, it sounds like it could work.
If you have any interest in my work into a hypothesis for 1 let me know.
Human beings can probabilistically read each others' source code. That's why we use primitive versions of noncausal decision theory like getting angry, wanting to take revenge, etc.
This seems like a weird way of say, humans can make/refine hypotheses about other agents. What does talking about source code give you?
Tit for tat (which seems like revenge) works in normal game theories for IPD (of infinite length) which is a closer to what we experience in everyday life. I thought Non-causal decision theories are needed for winning on one-shots?
In the case of humans, "talking about source code" is perhaps not that useful, though we do have source code, it's written in quaternary and has a rather complex probabilistic compiler. And that source code was optimized by a purely causal process, demonstrating the fact that causal decision theory agents self modify into acausal decision theory agents in many circumstances.
Revenge and anger work for one shot problems, for example if some stranger comes and sexually assaults your wife, they cannot escape your wrath by "saying oh it's only one shot, I'm a stranger in a huge city you'll never see me again so there's no point taking revenge". You want to punch the in the face as an end in itself now, this is a simple way of our brains being a bit acausal, decision theory wise.
I thought anger and revenge (used in one shot situations) might be generalising from what to do in the iterated version which is what we had for more of our evolutionary history.
I kinda like a-causal decision theory for choosing to vote at all. I will choose to vote so that other people like me choose to vote.