Two facts:
Here's the paper which mentions both of these facts. (The actual paper is not important, I expect these facts to be well-known to coronavirus researchers, if the paper itself is not terribly mistaken and if I haven't misread anything.)
Even if cross-immunity is mild, won't it make sense to intentionally infect people with HCoV-OC43? Downside seems quite small compared to the number of deaths, and intuitively it seems that "mild cross-immunity" = "less severe SARS-CoV-2 cases", which is extremely valuable.
I notice I'm confused, since these facts should be well-known to pretty much everyone who's working on the vaccine. What's the explanation for why it's not a good idea?
Possible explanations, but I'm probably missing something:
So, what am I missing here?
Yes! We have an English club each Saturday at 5 PM.
Whoa, for some reason I thought that LTF fund is not relevant to us, but looks like I was wrong. Thank you!
For context: in the last few months I applied for two CEA grants.
We could work more on improving our reputation on LW and EA forum (I have a few long posts in mind, e.g. our community building strategy which we've developed recently and which we're very hopeful about), but that's a costly strategy and there's a lot of uncertainty on whether that would be useful (for us or for the international community).
Thanks! I wonder if there'd be legal issues because Kocherga is not a non-profit (non-profits in Russia can be politically complicated, as I've heard). But it's defnitely worth trying.
One more thing: unlike the other stuff, I feel like developing EA movement in Russia is more talent-constrained: it could be much more active if we had one enthusiastic person with managerial skills and ~10 hours/week on their hands. I'm not sure we have such a person in our community - maybe we do, maybe we don't.
(Sometimes I consider taking on this role myself, but right now that's impossible, since I'm juggling 3 or 4 different roles already.)
OTOH, I'm also not sure how much better things would be if we had more funding and could hire such people directly. I might significantly underestimate this course of action because I don't have much experience yet with extending organizational capacity through hiring.
We've tried to start a local EA movement early on and had a few meetups in 2016. Introductory talks got stale quite quickly, so we put together a core EA team, with Trello board and everything.
It wasn't very clear what we were supposed to do, though:
So, we had ~5-6 people on the team and were doing fine for a while, but eventually it all fizzled out due to the lack of time, shared vision and organizational capacity.
We tested several approaches to reboot it a few times since then. Haven't succeeded yet, but we'll try again.
---
Currently, EA movement in Russia is mostly promoted by Alexey Ivanov from Saint-Petersburg. He takes care of online resources and organizes introductory EA talks and AI Safety meetups. He's doing a great work.
Another guy is working on a cool project to promote EA/rationality among the talented students, but that project is still in its early stages and I feel like it's not my story to tell.
Thank you!
I've applied to CFAR's workshop in Prague myself (and asked for financial aid, of course); they haven't contacted me yet.
I'll explain about EA in reply to this comment.
Thanks! I'm planning to write a separate post with more details on our community, activities and accumulated experiences; there's much more stuff I'd like to share which didn't fit in this one. It might take a few weeks, though, since my English writing is quite sluggish.
Okay, SARS-CoV-2 is pretty different from SARS-2003 ("~76% amino acid identity in the spike protein"), this might be the reason it won't work. OTOH, I don't know how different HCoV-OC43 is from both SARS strains.