(Epistemic status: I do not understand GPT deeply, so this is just a random idea.)
If I understand it correctly, GPT learns from existing texts. Lots of texts.
Would it be possible to make GPT smarter by simply giving it smarter text? Of course, writing tons of smarter text would be a lot of work, but what about annotating the existing text, like "take this more seriously" and "take this less seriously"? (From technical perspective, maybe the GPT should read the text marked as serious five times?) Assuming that the annonation is roughly correct, would this improve the results?
*
If yes, the problem is how to select smarter texts, especially if we want lots of them? But I think some good guesses can be made:
- High-school textbooks. They should be educational and relatively uncontroversial (settled science). There probably are some reviews of textbooks, so only annotate the ones with good reviews.
- Some parts of Reddit are probably much better than average, in the sense that smart content gets upvoted. So include the upvoted comments.
Tagging is conditioning (see also). If instead of having SSL model learn
text
, you have it learn(summary, text)
, then it learns to predicttext
fromsummary
. A summary can be any measurement of text such that some values of that measurement can be a useful query. For example, iftext
is a chess game, asummary
could be a statement of who wins. Then, starting a prompt with a claim of winning will tend to lead to a winning game. Similarly, ifsummary
says iftext
is a valid argument, you gain the ability to query for valid arguments. Finally,summary
can well be written by a language model, using a prompt that includestext
and instructions to write an appropriate summary. Ifsummary
is a free-form text description, queries can become arbitrary texts as well, including those not appearing in the training dataset.