This is my first post on lesswrong. I'll merely be linkposting content on epistemics and alignment here while getting more familiar with the culture.
tl;dr:
We attempt to automatically infer one’s beliefs from their writing in three different ways. Initial results based on Twitter data hint at embeddings and language models being particularly promising approaches.
What do you think the results would be like if you try to use a language model to automatically filter for direct-opinion tweets and do automatic negation?
Cool to hear you tried it!