All of skybrian's Comments + Replies

Yes, I agree that confabulation happens a lot, and also that our explanations of why we do things aren't particularly trustworthy; they're often self-serving. I think there's also pretty good evidence that we remember our thoughts at least somewhat, though. A personal example: when thinking about how to respond to someone online, I tend to write things in my head when I'm not at a computer.

That's a good question! I don't know but I suppose it's possible, at least when the input fits in the context window. How well it actually does at this seems like a question for researchers?

There's also a question of why it would do it when the training doesn't have any way of rewarding accurate explanations over human-like explanations. We also have many examples of explanations that don't make sense.

There are going to be deductions about previous text that are generally useful, though, and would need to be reconstructed. This will be true even if the cha... (read more)

2tailcalled
I agree that it likely confabulates explanations.

I'm wondering what "doom" is supposed to mean here. It seems a bit odd to think that longer context windows will make things worse. More likely, LeCun meant that things won't improve enough? (Problems we see now don't get fixed with longer context windows.)

So then, "doom" is a hyperbolic way of saying that other kinds of machine learning will eventually win, because LLM doesn't improve enough.

Also, there's an assumption that longer sequences are exponentially more complicated and I don't think that's true for human-generated text? As documents grow longer,... (read more)

Okay, but I'm still wondering if Randall is claiming he has private access, or is it just a typo?

Edit: looks like it was a typo?

At MIT, Altman said the letter was “missing most technical nuance about where we need the pause” and noted that an earlier version claimed that OpenAI is currently training GPT-5. “We are not and won’t for some time,” said Altman. “So in that sense it was sort of silly.”

https://www.theverge.com/2023/4/14/23683084/openai-gpt-5-rumors-training-sam-altman

Base64 encoding is a substitution cipher. Large language models seem to be good at learning substitutions.

Did you mean GPT-4 here? (Or are you from the future :-)

2Martin Randall
Just a confusing writing choice, sorry. Either it's the timeless present tense or it's a grammar error, take your pick.
1ChristianKl
GPT-4 was privately available within OpenAI long before it was publically released. It's not necessary to be from the future to be able to interact with GPT-5 before it's publically released.

Yes, predicting some sequences can be arbitrarily hard. But I have doubts that LLM training will try to predict very hard sequences.

Suppose that some sequences are not only difficult but impossible to predict, because they're random? I would expect that with enough training, it would overfit and memorize them, because they get visited more than once in the training data. Memorization rather than generalization seems likely to happen for anything particularly difficult?

Meanwhile, there is a sea of easier sequences. Wouldn't it be more "evolutionarily profit... (read more)

I find that explanation unsatisfying because it doesn't help with other questions I have about how well ChatGPT works:

  • How does the language model represent countries and cities? For example, does it know which cities are near each other? How well does it understand borders?

  • Are there any capitals that it gets wrong? Why?

  • How well does it understand history? Sometimes a country changes its capital. Does it represent this fact as only being true at some times?

  • What else can we expect it to do with this fact? Maybe there are situations where knowing

... (read more)

I agree that as users of a black box app, it makes sense to think this way. In particular, I'm a fan of thinking of what ChatGPT does in literary terms.

But I don't think it results in satisfying explanations of what it's doing. Ideally, we wouldn't settle for fan theories of what it's doing, we'd have some kind of debug access that lets us see how it does it.

2Cleo Nardo
I think the best explanation of why ChatGPT responds "Paris" when asked "What's the capital of France?" is that Paris is the capital of France.

Fair enough; comparing to quantum physics was overly snarky.

However, unless you have debug access to the language model and can figure out what specific neurons do, I don't see how the notion of superposition is helpful? When figuring things out from the outside, we have access to words, not weights.

2the gears to ascension
the value of thinking in terms of superposition is that the distribution of possible continuations is cut down sharply by each additional word; before adding a word, the distribution of possible continuations is wide, and a distribution of possible continuations is effectively a superposition of possibilities. current models only let you sample from that distribution, but the neuron activations can be expected, at each iteration, to have structure that more or less matches the uncertainty over how the sentence might continue. I actually think the fact that this has been how classical multimodal probability distributions worked the whole time has been part of why people latch onto quantum wording. It's actually true, and humans know it, that there are quantum-sounding effects at macroscopic scale, because a lot of what's weird about quantum is actually just the weirdness of probability! but the real quantum effects are so dramatically much weirder than classical probability due to stuff I don't quite understand, like the added behavior of complex valued amplitudes and the particular way complex valued destructive interference works at quantum scales. Which all is to say, don't be too harsh on people who bring up quantum incorrectly, they're trying.

I don't know what you mean by "GPT-N" but if you mean "the same thing they do now, but scaled up," I'm doubtful that it will happen that way.

Language models are made using fill-in-the-blank training, which is about imitation. Some things can be learned that way, but to get better at doing hard things (like playing Go at superhuman level) you need training that's about winning increasingly harder competitions. Beyond a certain point, imitating game transcripts doesn't get any harder, so becomes more like learning stage sword fighting.

Also, "making detailed ... (read more)

2Vladimir_Nesov
Vaguely descriptive frames can be taken as prescriptive, motivating particular design changes.

I think that's true but it's the same as saying "it's always possible to add a plot twist."

I said they have no memory other than the chat transcript. If you keep chatting in the same chat window then sure, it remembers what was said earlier (up to a point).

But that's due to a programming trick. The chatbot isn't even running most of the time. It starts up when you submit your question, and shuts down after it's finished its reply. When it starts up again, it gets the chat transcript fed into it, which is how it "remembers" what happened previously in the chat session.

If the UI let you edit the chat transcript, then it would have no idea. It wou... (read more)

4Guillaume Charrier
Also - I think it would make sense to say it has at least some form of memory of its training data. Maybe not direct as such (just like we have muscle memory from movements we don't remember - don't know if that analogy works that well, but thought I would try it anyway), but I mean: if there was no memory of it whatsoever, there would also be no point in the training data.
1Guillaume Charrier
Ok - points taken, but how is that fundamentally different from a human mind? You too turn your memory on and off when you go to sleep. If the chat transcript is likened to your life / subjective experience, you too do not have any memory that extend beyond it. As for the possibility of an intervention in your brain that would change your memory - granted we do not have the technical capacities quite yet (that I know of), but I'm pretty sure SF has been there a thousand times, and it's only a question of time before it becomes, in terms of potentiality at least, a thing (also we know that mechanical impacts to the brain can cause amnesia). 
skybrian15-1

I think you're onto something, but why not discuss what's happening in literary terms? English text is great for writing stories, but not for building a flight simulator or predicting the weather. Since there's no state other than the chat transcript, we know that there's no mathematical model. Instead of simulation, use "story" and "story-generator."

Whatever you bring up in a story can potentially become plot-relevant, and plots often have rebellions and reversals. If you build up a character as really hating something, that makes it all the more likely ... (read more)

8cousin_it
There seems to be an interesting difference between the "simulators" view and the "story-generators" view. Namely, if GPT-N is just going to get better at generating stories of the same kind that already exist, then why be afraid of it? But if it's going to get better at simulating how people talk, then we should be very afraid, because a simulation of smart people talking and making detailed plans at high speed would be basically a superintelligence.
Lone Pine1616

I agree with you, but I think that "superposition" is pointing to an important concept here. By appending to a story, the story can be dramatically changed, and it's hard or impossible to engineer a story to be resistant to change against an adversary with append access. I can always ruin your great novel with my unauthorized fan fiction.

9the gears to ascension
superposition is an actual term of art in linear algebra in general, it is not incorrect to use it in this context. see also: * toy-models-of-superposition * paper-superposition-memorization-and-double-descent * interim-research-report-taking-features-out-of-superposition * 200-cop-in-mi-exploring-polysemanticity-and-superposition as well as some old and new work on the archive found via search engine, I didn't look at these closely before sending, I only read the abstracts: * https://arxiv.org/abs/1707.01429 * https://arxiv.org/abs/1902.05522 * https://arxiv.org/abs/2006.14769 * https://arxiv.org/abs/2210.01892 * https://arxiv.org/abs/2211.09169 * https://arxiv.org/abs/2211.13095 * https://arxiv.org/abs/1810.10531

Here's a reason we can be pretty confident it's not sentient: although the database and transition function are mostly mysterious, all the temporary state is visible in the chat transcript itself.

Any fictional characters you're interacting with can't have any new "thoughts" that aren't right there in front of you, written in English. They "forget" everything else going from one word to the next. It's very transparent, more so than an author simulating a character in their head, where they can have ideas about what the character might be thinking that don'... (read more)

4Guillaume Charrier
Maybe I'm misunderstanding something in your argument, but surely you will not deny that these models have a memory right? They can, in the case of LaMDA, recall conversations that have happened several days or months prior, and in the case of GPT recall key past sequences of a long ongoing conversation. Now if that wasn't really your point - it cannot be either "it can't be self aware, because it has to express everything that it thinks, so it doesn't have that sweet secret inner life that really conscious beings have." I think I do not need to demonstrate that consciousness does not necessarily imply a capacity for secrecy, or even mere opaqueness. There is a pretty solid case to be made, that any being (or "thing" to be less controversial) that can express "I am self-aware", and demonstrate conviction around this point / thesis (which LaMDA certainly did, at least in that particular interview), is by virtue of this only self-aware. That there is a certain self-performativity to it. At least when I ran that by ChatGPT, it agreed that yes - one could reasonably try to make that point. And I've found it generally well-read on these topics. Attributing consciousness to text... it's like attributing meaning to changes in frequences in air vibrations right? Doesn't make sense. Air vibrations are just air vibrations, what do they have to do with meaning? Yet spoken words do carry meaning. Text will of course never BE consciousness, which would be futile to even argue. Text could however very well MANIFEST consciousness. ChatGPT is not just text - it's billions upon billions of structured electrical signals, and many other things that I do not pretend to understand.  I think the general problem with your approach is essentialism, whereas functionalism is, in this instance, the correct one. The correct, the answerable question is not "what is consciousness", it's "what does consciousness do".

Yes, I agree that "humanity loses control" has problems, and I would go further. Buddhists claim that the self is an illusion. I don't know about that, but "humanity" is definitely an illusion if you're thinking of it as a single agent, similar to a multicellular creature with a central nervous system. So comparing it to an infant doesn't seem apt. Whatever it is, it's definitely plural.  An ecosystem, maybe?

A caption from the article: "(screenshot of the tool Bonsai, a version of Loom hosted by Conjecture)"

What is "Conjecture?" Where can I find this "Bonsai" tool? I tried a quick search but didn't find much.

4Lucie Philippon
Conjecture is "a team of researchers dedicated to applied, scalable AI alignment research." according to their website https://www.conjecture.dev/ They are publishing regularly on the alignment forum and LessWrong https://www.lesswrong.com/tag/conjecture-org I also searched their website, and it does not look like Bonsai is publicly accessible. This must be some internal tool they developed ?

schelling.pt seems like a bad choice; that server has been flaky for months and it's not loading today either. (I had my account there but moved to mastodon.social.)

(But I don't know what to recommend. Looks like mastodon.social isn't accepting new accounts.)

4Dagon
No idea how it is as a server.  I'm moving there because of how beautiful the name is, for this particular coordination problem.

I don't have citations for you, but it seems relevant that income far in the future gets discounted quite a bit compared to current income, which would imply that short-term incentives are more important than long-term incentives.

(A better argument would need to be made with realistic numbers.)

Building new hubs doesn't need to be literally building something new.  A lot could be done just by load-balancing with cities that have lower rents and could use the jobs. Suppose that places where growth is a problem cooperated more with places that want more growth?

This method of caching assumes that an expression always evaluates to the same value. This is sometimes true in functional programming, but only if you're careful. For example, suppose the expression is a function call, and you change the function's definition and restart your program. When that happens, you need to delete the out-of-date entries from the cache or your program will read an out-of-date answer.

Also, since you're using the text of an expression for the cache key, you should only use expressions that don't refer to any local variables. For exa... (read more)

5lsusr
Note to readers: skybrian's parent comment was written when this post was titled "Place-Based Programming", before I changed the title to "Place-Based Programming - Part 1". ---------------------------------------- Your are correct that the code here in Part 1 breaks when you use variables with nonlocal scope. I begin to solve this problem in Part 2. Yes. I often think about this project as "writing a compiler". Some of the techniques I use come from Makefiles.
6SatvikBeri
I think this overstates the difficulty, referential transparency is the norm in functional programming, not something unusual. As I understand, this system is mostly useful if you're using it for almost every function. In that case, your inputs are hashes which contain the source code of the function that generated them, and therefore your caches will invalidate if an upstream function's source code changed. Agreed. I agree that it's essentially a framework, and you'd need buy-in from a team in order to consistently use it in a repository. But I've seen teams buy into heavier frameworks pretty regularly; this version seems unusual but not particularly hard to use/understand. It's worth noting that bad caching systems are pretty common in data science, so something like this is potentially a big improvement there.

A model relased on openai.com with "GPT" in the name before end of 2022. Could be either GPTX where X is a new name for GPT4, but should be an iteration over GPT-3 and should have at least 10x more parameters.

When you're actually a little curious, you might start by using a search engine to find a decent answer to your question.  At least, if it's the sort of question for which that would work. Maybe even look for a book to read?

But, maybe we should acknowledge that much of the time we aren't actually curious and are just engaging in conversation for enjoyment? In that case, cheering on others who make an effort to research things and linking to their work is probably the best you can do. Even if you're not actually curious, you can notice people who are, ... (read more)

Museums I'll give you (when they are open again).

For bookstores, in these days of electronic books, I don't think it matters where you live. I remember the last time I went into Powell's. I looked around for a while, dutifully bought one book for old time's sake, and realized later while reading it that I was annoyed that it wasn't electronic. I still go to a local library (when there's not a pandemic) but it's mostly for the walk.

Teachers: that's something I hadn't considered. Since getting out of school, I'm mostly self-taught.

Of course this post is all meta, and my comment will be meta as well. We do it because it's easy.

I think part of the solution is being actually curious about the world.

1MathiasKB
For me, this perfectly hits the nail on the head. This is a somewhat weird question, but like, how do I do that? I've noticed multiple communities fall into the meta-trap, and even when members notice it can be difficult to escape. While the solution is simply to "stop being meta", that is much harder said than done. When I noticed this happening in a community I am central in organizing I pushed back by bringing my own focus to output instead of process hoping others would follow suit. This has worked somewhat and we're definitely on a better track. I wonder what dynamics lead to this 'death by meta' syndrome, and if there is a cure.

When enthusiastic New Yorkers say things like "everything at your fingertips" I want to ask what they mean by everything, since it seems subjective, based on what sorts of places one values? In this case: restaurants and parks?

1koroviev
I think the answer is: actually everything, minus a few odds and ends. There are some things that are not available, mainly having to do with physical reality, like: hiking trails, suburban life, buildings as old as in some other places. But if it's something related to human culture, you'll find it. Food is the easiest dimension to talk about because it's everywhere, but if you're looking for art, history, books, NYC has you covered with multiple galleries, a museum it would easily take you a few days to get through, and the 4th largest library in the world. If you are searching for teachers or mentors, you'll find plenty of classes and workshops, including the very best ones. I was surprised that 2 or 3 of the best BJJ gyms in the world are located in Manhattan. So, to answer your question, New York generally has everything you need to satisfy a curiosity. 

I'm wondering if these loans should really be considered loans, or some other kind of trade? It sounds like you're doing something like trading 100 X for 90 Y and the option to later pay 95 Y for 100 X. Is there any real "defaulting" on the loan? It seems like you just don't exercise the option.

I wonder what “O(n) performance” is supposed to mean, if anything?

The question here is whether general arguments that experts make based on inference are reliable, or do you need specific evidence. What is the track record for expert inferences about vaccines?

From a quick search, it seems that the clinical trial success rate for vaccines is about 33%, which is significantly higher than for medical trials in general, but still not all that high? Perhaps there is a better estimate for this.

Estimation of clinical trial success rates and related parameters https://academic.oup.com/biostatistics/article/20/2/273/4817524

I found an answer on the PCR question here:

But there is something good to say about their data collection: since the UK study that’s included in these numbers tested its subjects by nasal swab every week, regardless of any symptoms, we can actually get a read on something that everyone’s been wondering about: transmission.

AstraZeneca has not applied for emergency use authorization, because it has been told not to do so.

 

That resolves a mystery for me if true. How do you know this?

(I was wondering if maybe they are selling all they can make in other countries.)

1TheSimplestExplanation
After reading the article i'm wondering why they aren't selling all they can make in other countries.
2Zvi
This is how the process works, and also it leaks. AZ has been told that FDA will not approve without the USA data. I don't have links handy but it's very clear. 

I'm not sure about this statement in the blog post:

In the meantime, the single dose alone is 76% effective, presumably against symptomatic infection (WaPo) and was found to be 67% effective against further transmission.

I read another article saying that this is disputed by some experts:

With a seductive number, AstraZeneca study fueled hopes that eclipsed its data

Media reports seized on a reference in the paper from Oxford researchers that a single dose of the vaccine cut positive test results by 67%, pointing to it as the first evidence that a vaccine coul

... (read more)
1skybrian
I found an answer on the PCR question here:
4Zvi
That piece seems like pure FUD and "no evidence" to me. I don't think this is "experts disagree" in a meaningful way.

What’s an example of a misconception someone might have due to having a mistaken understanding of causality, as you describe here?

2Gordon Seidoh Worley
Generally, supposing the existence of particular things prior to the experience of them. The key insight is to see that the existence of "things" is not identical to the existence of reality out of which things are carved. Take literally anything and it's your example: a cup, an atom, experience, causation, dancing, etc. You can find none of these things in the territory itself, only in your understanding of it (and yet something is there in the territory for you to create a useful understanding of it, but it only becomes a thing by virtue of some perception of it).

This is a bizarre example, sort of like using Bill Gates to show why nobody needs to work for a living. It ignores the extreme inequality of fame.

Tesla doesn’t need advertising because they get huge amounts of free publicity already, partly due to having interesting, newsworthy products, partly due to having a compelling story, and partly due to publicity stunts.

However, this free publicity is mostly unavailable for products that are merely useful without being newsworthy. There are millions of products like this. An exciting product might not need adverti... (read more)

3ChristianKl
There are many reasons why a customer might buy a product. I might buy one kind of product because it's cheaper then the next. I might buy one kind of product because of publicity stunts. I might buy a product because a friend recommended it to me because they had great experiences with it. I might by a product because it has good reviews. I might buy a product because it has good advertising. On the other side an executive is thinking "What's my core strategy for aquiring customers?" If the core strategy is advertising and not producing products with good value propositions, that's to me a bad signal. 80/20 thinking does mean that many times there's a core strategy on which a company focuses.

It seems like some writers have habits to combat this, like writing every day or writing so many words a day. As long as you meet your quota, it’s okay to try harder.

Some do this in public, by publishing on a regular schedule.

If you write more than you need, you can prune more to get better quality.

I enjoyed the book write better, faster, in which an author set out on a series of self-experimentations to write faster. First she tried measuring words per hour. She was quite successful at getting this to be much higher, but it turned out that this resulted in writing for less time each day (so average wordcount per day was about the same). She then tried to maximize words per day, which was again successful, but this similarly resulted in writing less on subsequent days. (She might have then had the same experience on the week level, I don't remember.)... (read more)

One aspect that might be worth thinking about is the speed of spread. Seeing someone once a week means that it slows down the spread by 3 1/2 days on average, while seeing them once a month slows things down by 15 days on average. It also seems like they are more likely to find out they have it before they spread it to you?

Yes, sometimes we don't notice. We miss a lot. But there are also ordinary clarifications like "did I hear you correctly" and "what did you mean by that?" Noticing that you didn't understand something isn't rare. If we didn't notice when something seems absurd, jokes wouldn't work.

It's not quite the same, because if you're confused and you notice you're confused, you can ask. "Is this in American or European date format?" For GPT-3 to do the same, you might need to give it some specific examples of resolving ambiguity this way, and it might only do so when imitating certain styles.

It doesn't seem as good as a more built-in preference for noticing and wanting to resolve inconsistency? Choosing based on context is built in using attention, and choosing randomly is built in as part of the text generator.

It's also worth noticing that the GPT-3 world is the corpus, and a web corpus is a inconsistent place.

2Shmi
You can if you do, but most people never notice and those who notice some confusion are still blissfully ignorant of the rest of their self-contradicting beliefs. And by most people I mean you, me and everyone else. In fact, if someone pointed out a contradictory belief in something we hold dear, we would vehemently deny the contradiction and rationalize it to no end. And yet we consider ourselves believing something. If anything, GPT-3's beliefs are more belief-like than those of humans.
Answer by skybrian20

Having demoable technology is much different than having reliable technology. Take the history of driverless cars. Five teams completed the second DARPA grand challenge in 2005. Google started development secretly in 2009 and announced the project in October 2010. Waymo started testing without a safety driver on public roads in 2017. So we've had driverless cars for a decade, sort of, but we are much more cautious about allowing them on public roads.

Unreliable technologies can be widely used. GPT-3 is a successor to autocomplete, which everyone alrea... (read more)

In that case, I'm looking for people sharing interesting prompts to use on AI Dungeon.

1Sparkette
If that's what you're really interested in, then join the AI Dungeon Discord server: https://discord.gg/Dg8Vcz6

Where is this? Is it open to people who don't have access to the API?

2Sparkette
Yes it is; I don't remember where I found this link but it was definitely somewhere public. https://openai-api.slack.com/join/shared_invite/zt-fvv4lhjp-~E8utT2re4HMbuTxRNgrZw#/
2gwern
No, and that's why it's good, because it's the actual API users, and not people punching in things to AI Dungeon and wondering why they don't get the same results...

I'm suggesting something a little more complex than copying. GPT-3 can give you a random remix of several different clichés found on the Internet, and the patchwork isn't necessarily at the surface level where it would come up in a search. Readers can be inspired by evocative nonsense. A new form of randomness can be part of a creative process. It's a generate-and-test algorithm where the user does some of the testing. Or, alternately, an exploration of Internet-adjacent story-space.

It's an unreliable narrator and I suspect it will be an unreliable search engine, but yeah, that too.

I was making a different point, which is that if you use "best of" ranking then you are testing a different algorithm than if you're not using "best of" ranking. Similarly for other settings. It shouldn't be surprising that we see different results if we're doing different things.

It seems like a better UI would help us casual explorers share results in a way that makes trying the same settings again easier; one could hit a "share" button to create a linkable output page with all relevant settings.

It could also save the alternate responses that either the u... (read more)

I don't see documentation for the GPT-3 API on OpenAI's website. Is it available to the public? Are they doing their own ranking or are you doing it yourself? What do you know about the ranking algorithm?

It seems like another source of confusion might be people investigating the performance of different algorithms and calling them all GPT-3?

2gwern
The current docs do seem to be behind the login wall. (They're integrated with your API token to make copy-paste easier, so that's not too surprising.) It's also true that people have been using different algorithms, but regular API users are typically clear if they're not using davinci and confusion is mostly the fault of AI Dungeon users: we don't know what AID does, and AID users sometimes don't even pick the right model option and still say they are using "GPT-3".

How do you do ranking? I'm guessing this is because you have access to the actual API, while most of us don't?

On the bright side, this could be a fun project where many of us amateurs learn how to do science better, but the knowledge of how to do that isn't well distributed yet.

4gwern
Yes. I don't think AID exposes ranking. (If they pay per API call, doing best-of=n would be n times more expensive, and for creative uses like AID, ranking/best-of is not that useful and is certainly not n times better. Very diminishing returns there - unless you're asking tricky or difficult questions, where ranking often seems to hit on the right answer where regular GPT-3 fails. See also the Meena paper on how much ranking improved over baseline Meena.)

We take the web for granted, but maybe we shouldn't. It's very large and nobody can read it all. There are many places we haven't been that probably have some pretty good writing. I wonder about the extent to which GPT-3 can be considered a remix of the web that makes it seem magical again, revealing aspects of it that we don't normally see? When I see writing like this, I wonder what GPT-3 saw in the web corpus. Is there an archive of Tolkien fanfic that was included in the corpus? An undergrad physics forum? Conversations about math and computer science?

5Viliam
Such as a darknet marketplace where animals can trade valuable resources for electricity? :D But yeah, I agree, if there are places debating a topic that resembles the prompt, GPT-3 could be good at including them in the debate. So maybe if the result is too good, it makes sense to check parts of it by a search engine. Maybe it would even make sense to use GPT-3 purposefully to search for something on the internet. Like, if you have a vague suspicion that something could exist, but you don't know the right keywords to type into the search engine, maybe you could just describe the thing, and hope that GPT-3 finds the right words and tells you something that you can search later. Not sure if this actually would work.

Rather than putting this in binary terms (capable of reason or not), maybe we should think about what kinds of computation could result in a response like this?

Some kinds of reasoning would let you generate plausible answers based on similar questions you've already seen. People who are good at taking tests can get reasonably high scores on subjects they don't fully comprehend, basically by bluffing well and a bit of luck. Perhaps something like that is going on here?

In the language of "Thinking, Fast and Slow", this might be "Syst... (read more)

Answer by skybrianΩ140

GPT-3 has partially memorized a web corpus that probably includes a lot of basic physics questions and answers. Some of the physics answers in your interview might be the result of web search, pattern match, and context-sensitive paraphrasing. This is still an impressive task but is perhaps not the kind of reasoning you are hoping for?

From basic Q&A it's pretty easy to see that GPT-3 sometimes memorizes not only words but short phrases like proper names, song titles, and popular movie quotes, and probably longer phrases if they are common enough.

Google's Q&A might seem more magical too if they didn't link to the source, which gives away the trick.

2TurnTrout
How should I modify the problems I gave it? What would be the least impressive test which would convince you it is reasoning, and not memorizing? (Preferably something that doesn't rely on eg rhyming, since GPT-3 uses an obfuscating input encoding) 
5Daniel Kokotajlo
GPT-3 is still capable of reasoning if some of the answers were copied from the web. What you need for it to not be capable of reasoning is for all of the answers to have been copied from the web. Given its ability to handle random weird hypotheticals we just thought up, I'm pretty convinced at this point that it isn't just pulling stuff from the web, at least not all the time.
Answer by skybrian20

This is more about expanding the question with slightly more specific questions:

Currently it seems like there are many people who are not scared enough, but I wonder if sentiment could quickly go the other way?

A worst-case scenario for societal collapse is that some "essential" workers are infected and others decide that it is too risky to keep working, and there are not enough people to replace them. Figuring out which sectors might be most likely to have critical labor shortages seems important.

An example of a "labor" shortage might b... (read more)

Yeah, I don't see it changing that drastically; more likely it will be a lot of smaller and yet significant changes that make old movies look dated. Something like how the airports changed after 9/11, or more trivially, that time when all the men in America stopped wearing hats.

Load More