I really wish he had used the time to put this into a coherent, complete paper, with references, open peer review, and definitions, citable, attackable.
Just tried going through the transcript looking for novel arguments, and... gah. Because it is audio, you do not even have correct grammar.
Another option is to hire someone else to convert ideas into properly written papers. Pay them with money, co-authorship, or both.
I really like the "when you don't have a good detailed model you need to figure out what space you should have the maximum entropy distribution over" framing
If I understood correctly, he mentions augmenting humans as a way out of the existential risk. At least I understood he has more faith in it than in making AI do our alignment homework. What does he mean by that? Increasing productivity? New drug development? Helping us get insights into new technology to develop? All of the above? I'd love to understand the ideas around that possible way out.
Really extremely happy with this podcast - but I feel like it also contributed to a major concern I have about how this PR campaign is being conducted.
Another podcast appearance. Great to see Eliezer on a real media blitz tour.
Youtube:
Full transcript: https://www.dwarkeshpatel.com/p/eliezer-yudkowsky#details