If one thinks the chance of an existential disaster is close to 100%, one might tend to worry less about the potential of a plan to counter it to backfire. It's not clear if that is a correct approach even if one thinks the chances of an existential disaster are...
mishka So, to create context, this is a continuation of our remarks in the comments to Zvi's "AI #40: A Vision from Vitalik" mishka There I was asking about 1. the boundary between humans merging with AI and digital humans (can these approaches be reliably differentiated from each other? or...
I have been trying to find out what is known about invariants in self-modifying systems. This might become a rather acute topic if we end up moving towards self-modifying AIs or self-modifying ecosystems of AIs. But it seems that not much has been done. For example, I have found a...
MadHatter Hello! Welcome to the dialogue. MadHatter I'll just wait a few minutes for you to see the notification, I guess. mishka Hi, yes, I see this. Great! mishka So, we have this conversation in the comments in your post [https://www.lesswrong.com/posts/DkkfPEwTnPQyvrgK8/ethicophysics-i](https://www.lesswrong.com/posts/DkkfPEwTnPQyvrgK8/ethicophysics-i) as a starting point. MadHatter Yes, I think that's...
I wonder if some people here had a chance to play with base-GPT-4 (the access is given very selectively for research purposes) and would not mind sharing some of their impressions? I know that some people have been playing with it, but I've never seen a discussion of impressions and...
There has been a 25 min interview with Ilya conducted by Sven Strohband and released on July 17: https://www.youtube.com/watch?v=xym5f0XYlSc This interview has a section dedicated to AI safety (7 min starting from 14:56). Ilya is now the co-lead of the OpenAI "superalignment" effort, and his thinking will likely be particularly...