(Epistemic status: I do not understand GPT deeply, so this is just a random idea.)
If I understand it correctly, GPT learns from existing texts. Lots of texts.
Would it be possible to make GPT smarter by simply giving it smarter text? Of course, writing tons of smarter text would be a lot of work, but what about annotating the existing text, like "take this more seriously" and "take this less seriously"? (From technical perspective, maybe the GPT should read the text marked as serious five times?) Assuming that the annonation is roughly correct, would this improve the results?
*
If yes, the problem is how to select smarter texts, especially if we want lots of them? But I think some good guesses can be made:
- High-school textbooks. They should be educational and relatively uncontroversial (settled science). There probably are some reviews of textbooks, so only annotate the ones with good reviews.
- Some parts of Reddit are probably much better than average, in the sense that smart content gets upvoted. So include the upvoted comments.
Tagging all text with the date it was produced is very important for other reasons as well. (And as far as I know, it isn't done at present.)
With such date tags, one could ask GPT to produce text for a specific date. For example, one could ask it to continue the prompt "I think people of different races should" with dates of 1880, 1920, 1960, and 2000 to see how racial attitudes have changed over the years, which is very interesting historical information. It also would allow one to force a 2023 date for normal responses, to avoid unwanted regurgitation of attitudes from 1880. That in turn would allow one to use training text from 1880 without worrying about such attitudes polluting normal interactions.