I am getting worried that people are having so much fun doing interesting stuff with GPT-3 and AI Dungeon that they're forgetting how easy it is to fool yourself. Maybe we should think about how many different cognitive biases are in play here? Here are some features that make it particularly easy during casual exploration.
First, it works much like autocomplete, which makes it the most natural thing in the world to "correct" the transcript to be more interesting. You can undo and retry, or trim off extra text if it generates more than you want.
Randomness is turned on by default, so if you try multiple times then you will get multiple replies and keep going until you get a good one. It would be better science but less fun to keep the entire distribution rather than stopping at a good one. Randomness also makes a lot of gamblers' fallacies more likely.
Suppose you don't do that. Then you have to decide whether to share the transcript. You will probably share the interesting transcripts and not the boring failures, resulting in a "file drawer" bias.
And even if you don't do that, "interesting" transcripts will be linked to and upvoted and reshared, for another kind of survivor bias.
What other biases do you think will be a problem?
The current docs do seem to be behind the login wall. (They're integrated with your API token to make copy-paste easier, so that's not too surprising.) It's also true that people have been using different algorithms, but regular API users are typically clear if they're not using davinci and confusion is mostly the fault of AI Dungeon users: we don't know what AID does, and AID users sometimes don't even pick the right model option and still say they are using "GPT-3".