(Epistemic status: I do not understand GPT deeply, so this is just a random idea.)
If I understand it correctly, GPT learns from existing texts. Lots of texts.
Would it be possible to make GPT smarter by simply giving it smarter text? Of course, writing tons of smarter text would be a lot of work, but what about annotating the existing text, like "take this more seriously" and "take this less seriously"? (From technical perspective, maybe the GPT should read the text marked as serious five times?) Assuming that the annonation is roughly correct, would this improve the results?
*
If yes, the problem is how to select smarter texts, especially if we want lots of them? But I think some good guesses can be made:
- High-school textbooks. They should be educational and relatively uncontroversial (settled science). There probably are some reviews of textbooks, so only annotate the ones with good reviews.
- Some parts of Reddit are probably much better than average, in the sense that smart content gets upvoted. So include the upvoted comments.
See the Galatica model (https://arxiv.org/abs/2211.09085) from Meta. It's trained on a curated dataset of scientific papers, reference materials and scientific knowledge bases (with only a very small % of random internet text). IIRC the benefits of this seem limited (better to train on a bigger dataset and use other techniques to make the model access the sciencey parts of the training set).