Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by

Edit suggestions for this article:

1) Typo in this article:

I suggest replacing

Scaffolded Large Language Models (S-LLMs) are rapidly evolving domain of AI. Their naturally modular structure and natural language data usage gives alignment researchers appear to give researchers a head start on interpretability.

with

Scaffolded Large Language Models (S-LLMs) are rapidly evolving domain of AI. Their naturally modular structure and natural language data usage appear to give alignment researchers a head start on interpretability.

(Text in bold found in only one version)

2) Also, I feel the last sentence could be clarified:

Also the modular nature of these systems enables novel advances in their design to be shared between human developers.

I think you are trying to communicate that alignment problems could spread more quickly from developer to developer because the system is modular, encouraging software reuse. (That is, a backward reference to 'The key intuition here is that the actual selection pressure is "having humans fork your code" '.) But as written, it feels like a positive, not a negative. Isn't the sharing of novel advances between developers a good thing? I don't think that's what you are trying to communicate here, though I'm not sure how to suggest an edit on this point. In part, because I'm not 100% sure I know what you are trying to communicate with this sentence.