100 years? So what will be valuable/interesting to a future AI or other transhuman intelligence? Maybe well-packed genetic material from highly endangered species. Most of them will probably have been preserved elsewhere, but if you have even one hit on a truly lost species, that would probably be of some value/interest to our future children.
Of course. Just let me know if I can be of help.
The target for researchers to "be able to think unusually clearly" personally pushes towards the Bellingham location. That sort of semi-isolation in, for me, one of the most beautiful regions in the US is highly conducive to focused thought.
Although, I think it trades off with other potential goals, for example: community expansion or access to power. Those of you in miri know better where you are on timeline but research institute in the woods feels like it optimizes for a very particular move and may leave you less flexibility if the game isn't in the state you imagine. As noted, I lack the information to know whether that's a good or bad bet.
If you prefer a more flexible approach, I'd consider one of the new or old tech hubs: Austin being an example of the former, Boston the latter. Both seem in some ways to be more future oriented than the bay which sadly feels like it's being consumed by a hustler/MBA ethos rather than a creative one. Also, perhaps consider hubs focused around next-gen industries such as biotech (Boston again or maybe San Diego) as there's just a difference in cultural dynamism as opposed to locations very much in exploit mode.
Finally, if you're still considering moving the community en masse, as discussed in the earlier posts, the only communities I can think of that have successfully done that have not left by choice. While in two minutes of thought I haven't come up with a way to drive the rationalists from the bay without physical risk or legal jeopardy, that does not mean it can't be done. Short of that, schisms can do wonders for motivation.
Given growth in both AI research and alignment research over the past 5 years, how do the rates of progress compare? Maybe separating absolute change, first and second derivatives.
Or, imagine if this were a service available to restaurants such that they could have an option on menu items: +$1 for ethically sourced eggs. Now, the service is transparent for them (maybe integrate with a payment provider willing to facilitate the network for free marketing) and they don't have to deal with buying two sets of eggs or taking supply risks.
Hm, starting to think there's a version of this that's viable.
I really like this idea of moral fungibility.
By cutting out the need for a separate "ethical" packaging, marketing, and distribution system, you vastly lower the costs for new entrants into the market. More, there would be additional benefits since you could cut other costs like the impact of long-range transportation (how do I weigh the environmental costs of shipping ethically sourced eggs from Maine to California vs buying local factory?).
I worry that consumers derive as much value from the act of buying the ethically sourced product as the actual reduction in harm, but maybe there's ways to market around that. Not sure, but it seems worth finding out :).
This is a huge problem area in NLP. Quite a few issues you raised, but just to pick 2: