Kim Suozzi was a neuroscience student with brain cancer who wanted to be cryonically preserved but lacked the funds. She appealed to reddit and a foundation was set up, called the Society for Venturism. Enough money was raised, and when she died on the January 17th, she was preserved by Alcor.
I wasn't sure if I should post about this, but I was glad to see that enough money was raised and it was discussed on LessWrong here, here, and here.
Edit: It looks like Alcor actually worked with her to lower the costs, and waived some of the fees.
Edit 2: The Society for Venturism has been around for a while, and wasn't set up just for her.
My median timeline estimate for loosely human-level AI (i.e. it is technically feasible to build AI that can do most anything a human can do, although AI performance would be superhuman in many areas, as it already is), conditional on no catastrophes stopping forward progress would be near the end of the century. This is not a very stable or solid estimate, and I would update a lot on seeing the views of folks who had studied the issues and accumulated strong track records in prediction exercises like DAGGRE focused on technological forecasting and other relevant areas, among many other things.
Median doom time toward the end of the century? That seems enormously optimistic. If I believed this I'd breathe a huge sigh of relief, upgrade my cryonics coverage, spend almost all current time and funding trying to launch CFAR, and write a whole lot more about the importance of avoiding biocatastrophes and moderating global warming and so on. I might still work on FAI due to comparative advantage, but I'd be writing mostly with an eye to my successors. But it just doesn't seem like ninety more years out is a reasonable median estimate. I'd expect bloody uploads before 2100.
Carl, ???