Humanity basically already knows how to build AGI. That doesn't really matter, though. FAI and AGI are incredibly different problems. They share basically none of the same architectural features. I think that understanding what an AGI is or what it would do is incredibly important for FAI; is that what you mean? But knowing how to build AGI doesn't do that for you; you still have to actually turn it on.
Agree that my parenthetical was straight-up wrong, but it was indeed a parenthetical. I still think the analogy is too loose to be useful.
I'd like evidence of the claim that humanity, or some subset thereof, "basically already knows how to build AGI". I would stipulate that we know how to do whole-brain emulation, giving adequate computational resources; is that what you had in mind?
One of the reasons that I am skeptical of contributing money to the SIAI is that I simply don't know what they would do with more money. The SIAI currently seems to be viable. Another reason is that I believe that an empirical approach is required, that we need to learn more about the nature of intelligence before we can even attempt to solve something like friendly AI.
I bring this up because I just came across an old post (2007) on the SIAI blog:
Some questions:
I also have some questions regarding the hiring of experts. Is there a way to figure out what exactly the current crew is working on in terms of friendly AI research? Peter de Blanc seems to be the only person who has done some actual work related to artificial intelligence.
I am aware that preparatory groundwork has to be done and capital has to be raised. But why is there no timeline? Why is there no progress report? What is missing for the SIAI to actually start working on friendly AI? The Singularity Institute is 10 years old, what is planned for the decade ahead?