I'm writing this to get information about the lesswrong community and whether it worth engaging. I'm a bit out of the loop in terms of what the LW community is like and whether it can maintain multiple view points (and how known the criticisms are).
The TL;DR is I have problems with treating computation in an overly formal fashion. The more pragmatic philosophy suggested implies (but doesn't prove) that AI will not be as powerful as expected as the physicality of computation is important and instantiating computing in a physical fashion is expensive.
I think all the things I will talk about are interesting, but I don't see the sufficiency of them when considering AI running in the real world in real computers.
1. Source code based decision theory
I don't understand why:
- other agents trust that your source code is what you say it is
- other agents trust that your implementation of your interpreter matches their understanding of the interpreter. I don't see how they get round trustless trust (inserting code/behaviour via malicious compiler/interpreter) issues when they don't have the ability to do diverse compilation.
2. General Functionalism
The idea that it doesn't matter how you compute something just whether the inputs and outputs are the same.
- The battery life of my phone says that the way of computation is very important, is it done on the cloud and I have to power up my antennae to transmit the result.
- Timing attacks say that speed of the computation is important, that faster is not always better.
- Rowhammer says that how you layout your memory is important. Can I flip a bit of your utility calculation?
- Memory usage, overheating, van Eck phreaking etc etc....
Human beings can probabilistically read each others' source code. That's why we use primitive versions of noncausal decision theory like getting angry, wanting to take revenge, etc.
This seems like a weird way of say, humans can make/refine hypotheses about other agents. What does talking about source code give you?
Tit for tat (which seems like revenge) works in normal game theories for IPD (of infinite length) which is a closer to what we experience in everyday life. I thought Non-causal decision theories are needed for winning on one-shots?