If you get an email from aisafetyresearch@gmail.com , that is most likely me. I also read it weekly, so you can pass a message into my mind that way.
Other ~personal contacts: https://linktr.ee/uhuge
I've heard that hypothesis in a review of that blog post of Anthropic, likely by
AI Explained
maybe by
bycloud
.
They've called it "Chekov's gun".
What's your view on sceptic claims about RL on transformer LMs like https://arxiv.org/abs/2504.13837v2 or one that CoT instruction yields better results than <thinking> training?
Not the content I expect labeled AIb Capabilities,
although I see how that'd be vindicated.
By the way, if I write an article about LMs generating SVG, that's a plaintext and if I put an SVG illustration up, that's an image, not a plaintext?
I'd update my take from a very pessimist/gloom one to an (additional) excited one: Those more intelligent models building a clear view of the person they/it interacts with is a sign of emerging empathy, which is a hopeful property for alignment/respect.
False Trichotomy?
Your model assumes that one cannot be all three, however, some roles demand it, and in reality people do navigate all three traits, my top example would be empathic project managers.
Hello @habryka, could you please adjust the text on the page to include the year when applications closed, so that it confuses people( like me) less and they won't spend reading it all wasting their time stupidly?
THANKS!
You mean the chevrons like this is non-standard, but also sub-standard, although it has the neat property to represent >Speaker one< and >>Speaker two<<? I can see the typography of those here is meh at best.-\
Link to Induction section on https://www.lesswrong.com/lw/dhg/an_intuitive_explanation_of_solomonoff_induction/#induction seems broken on mobile Chrome, @habryka