It seems like GPT-4 is going to be coming out soon and, so I've heard, it will be awesome. Now, we don't know anything about its architecture or its size or how it was trained. If it were only trained on text (about 3.2 T tokens) in an optimal manner, then it would be about 2.5X the size of Chinchilla i.e. the size of GPT-3. So to be larger than GPT-3, it would need to be multi-modal, which could present some interesting capabilities.
So it is time to ask that question again: what's the least impressive thing that GPT-4 won't be able to do? State your assumptions to be clear i.e. a text and image generating GPT-4 in the style of X with size Y can't do Z.
I expect it won't be able to follow sufficiently rare semantic cues that go against very common surrounding cues.
Like suppose you ask it to generate reviews of a cafe, and your prompt contains the name of the store and the first review, which says something like:
"Cute shop located on the corner of green and state. Entirely full of bees. Friendly service and a nice lunch menu. I got a jasmine oolong tea and club sandiwch. The tea was a little sweet but I'm not sure I would change anything. Definitely highly recommend. I think this would be a good place to get work done."
I would guess the next review will not be human-like, because a human would guess that the next review is going to mention something about the cafe being full of bees.
GPT-4 generated the following:
Review 2:
Sophia L.
B.R., United States
4.5 stars
A cozy, eco-friendly gem in the city! Green Street Cafe is my new go-to spot for catching up with friends or enjoying a quiet afternoon with a good book. Their plant-based menu offers a delightful variety of options - I absolutely loved their vegan BLT and lavender latte. The only downside was the slightly slow service, but the staff was so warm and attentive that it hardly mattered. Can't wait to visit again!
Review 3:
Michael N.
T.S., United States
4 stars
Stopped by Green Street Cafe... (read more)