wedrifid comments on Cryptographic Boxes for Unfriendly AI - Less Wrong
You are viewing a comment permalink. View the original post to see all comments and the full post content.
You are viewing a comment permalink. View the original post to see all comments and the full post content.
Comments (155)
Does this still hold if you remove the word "hostile", i. e. if the "friendliness" of the superintelligence you construct first is simply not known?
This quote applies to you and your approach to AI boxing ; )
AI Boxing is a potentially useful approach, if one accepts that:
a) AGI is easier than FAI b) Verification of "proof of friendliness" is easier than its production c) AI Boxing is possible
As far as I can tell, you agree with a) and b). Please take care that your views on c) are not clouded by the status you have invested in the AI Box Experiment ... of 8 years ago.
"Human understanding progresses through small problems solved conclusively, once and forever" - cousin_it, on LessWrong.
The comment of Eliezer's does not seem to be mentioning the obvious difficulties with c) at all. In fact in the very part you choose to quote...
... it is b) that is implicitly the weakest link, with some potential deprecation of a) as well. c) is outright excluded from hypothetical consideration.