Salty
Salty has not written any posts yet.

Is Lindsey using a nuanced definition of ""concept injection?
I am a non-specialist, just trying to follow and understand. I have to look up many [most] definitions and terms. This may ne a trivial matter, but for me to understand, definitions matter.
When I look up a meaning of "application steering" I find something more permanent. Has any discussion focused on Lindsey' use of the term concept injection as an application of activation steering: "We refer to this technique as concept injection—an application of activation steering"
To me the term suggests something in training that persists in the model, when in reality it's a per-query intervention. [Claude tells me that] another term would be: "inference-time modifications"... (read more)
Actually interested in your post and spaced repetition (SR) techniques, although I am not a specialist. I had no idea what means: "Anki," but it was on my LessWrong feed, so I gave it a look.
Consider adding a brief line to the intro - for those who, clueless like me, find their way to your post.
Something to identify what it is and most important acknowledge the developer - like: Anki is a free, open-source flashcard program that utilizes spaced repetition and active recall to help users memorize information effectively.
Elmes, D. (2024). Anki (Version 2.1.66) [Software]. https://apps.ankiweb.net/
https://en.wikipedia.org/wiki/Anki_(software)
Apologies if this is too basic or repeated elsewhere.
Thank you for the post.
Leo
Minor Point which may be mentioned in comments, but is the numbering in the subheads 'off' or 'deliberate?'
If you revise / reprint this post here or in your "Don't Worry About the Vase" substack, perhaps a note or correction?
Principles #1 / #4 / #8 / #10 / #13
Great post - I only noticed the numbering when I was making notes.
Cheers,
First, thank you for your work and this post. I am not a specialist, just interested, but confused. I don’t get the significance of the results, but appreciate the thought and effort you put into this project.
I am pushing back on the ’romantic framing’ that that LLMs are "blind models" that somehow develop an some degree of internal spatial understanding of Earth through pure reasoning or emergent intelligence.
In this case didn’t the author in effect say to the model "given this list of numbers - which happen to be latitude and longitude pairs - access your core intelligence (learned parameters / weights / internal representations) and decide if it would represent... (read more)