This is a companion post to Keeping content out of LLM training datasets, which discusses the various techniques we could use and their tradeoffs. My intention is primarily to start a discussion, I am not myself very opinionated on this.

As AIs become more capable, we may at least want the option of discussing them out of their earshot.

Places to consider (at time of writing, none of the below robots.txt files rule out LLM scrapers, but I include the links so you can check if this changes):

Options to consider:

  • Blocking everything in robots.txt, or by User-Agent
    • it would perhaps seem a shame for AIs to know none of the content here
  • (for EAF / LW / AF) Writing code for ForumMagnum to:
    • enable blocking (probably by User-Agent) to be configured per-post (I haven't investigated this for feasibility, but naively it doesn't seem so hard).
    • enable posts to be configured to be visible only to logged-in users (maybe this is a feature we might want for other reasons?)
  • Get clarity on how to ensure that canaries are effective, and then start using them more widely.

Feel free to suggest additions to either category.

To the extent that doing something here means spending software dev time, this raises the question not only of should we do this but how important is this, relative to the other things we can spend software developers on.

Link preview image by Jonny Gios on Unsplash

New Answer
New Comment

1 Answers sorted by

Yonatan Cale

32

As AIs become more capable, we may at least want the option of discussing them out of their earshot.

If I'd want to discuss something outside of an AI's earshot, I'd use something like Signal, or something that would keep out a human too.

AIs sometimes have internet access, and robots.txt won't keep them out.

I don't think having this info in their training set is a big difference (but maybe I don't see the problem you're pointing out, so this isn't confident).