What's Going on With OpenAI's Messaging?
This is a quickly-written opinion piece, of what I understand about OpenAI. I first posted it to Facebook, where it had some discussion. Some arguments that OpenAI is making, simultaneously: 1. OpenAI will likely reach and own transformative AI (useful for attracting talent to work there). 2. OpenAI cares a lot about safety (good for public PR and government regulations). 3. OpenAI isn’t making anything dangerous and is unlikely to do so in the future (good for public PR and government regulations). 4. OpenAI doesn’t need to spend many resources on safety, and implementing safe AI won’t put it at any competitive disadvantage (important for investors who own most of the company). 5. Transformative AI will be incredibly valuable for all of humanity in the long term (for public PR and developers). 6. People at OpenAI have thought long and hard about what will happen, and it will be fine. 7. We can’t predict concretely what transformative AI will look like or what will happen after (Note: Any specific scenario they propose would upset a lot of people. Vague hand-waving upsets fewer people). 8. OpenAI can be held accountable to the public because it has a capable board of advisors overseeing Sam Altman (he said this explicitly in an interview). 9. The previous board scuffle was a one-time random event that was a very minor deal. 10. OpenAI has a nonprofit structure that provides an unusual focus on public welfare. 11. The nonprofit structure of OpenAI won’t inconvenience its business prospects or shareholders in any way. 12. The name “OpenAI,” which clearly comes from the early days when the mission was actually to make open-source AI, is an equally good name for where the company is now.* (I don’t actually care about this, but find it telling that the company doubles down on arguing the name still is applicable). So they need to simultaneously say: “We’re making something that will dominate the global economy and outperform humans at all cap
I imagine we'd both agree that there can and should be a lot of evals and attempts at robustness / reliability for small / low-level systems?
It seems like the disagreement is in how useful such work will be for critical and broader alignment challenges.