All of Veedrac's Comments + Replies

Veedrac20

I failed to find an example easily when checking twitter this way.

4Davidmanheim
A few examples of being reasonable which I found looking through quickly; https://twitter.com/GaryMarcus/status/1835396298142625991 / https://x.com/GaryMarcus/status/1802039925027881390  / https://twitter.com/GaryMarcus/status/1739276513541820428 / https://x.com/GaryMarcus/status/1688210549665075201 
Veedrac80

Blue Origin was started two years earlier (2000 v 2002), had much better funding for most of its history,

This claim is untrue. SpaceX has never had less money than Blue Origin. It is maybe true that Blue Origin had fewer obligations attached to this money, since it was exclusively coming from Bezos, rather than a mix of investment, development contracts, and income for SpaceX, but the baseline claim that SpaceX was “money-poor” is false.

Veedrac50

I need to remake the graph with more recent data, but here is a graphic of US energy additions.

https://live.staticflickr.com/65535/53977597462_2095add298_k.jpg

Nonrenewables are a walking dead at this point. I wouldn't personally tell the story of it through Musk—I think cost curves and China are a more relevant framing—but the end point is much the same.

3transhumanist_atom_understander
Yes, Tesla's role in battery storage is actually odd and niche—their importance seems to be reducing the skill level required to built a battery energy storage facility by packaging batteries into self contained modules that contain all of the required equipment (thermal management, inverter, etc). The battery cells come from Asia. The question to which Musk is relevant is not "how did we get to a world where battery storage is feasible" but "why would someone be investing in this ten or fifteen years ago when the technology was not there". It seems to me to be a case where a futurist calculation that ignores the engineering details plus seemingly naive reasoning that the rest is "just engineering" would have actually given the right answer.
Veedrac*20

LeelaKnightOdds has convincingly beaten both Awonder Liang and Anish Giri at 3+2 by large margins, and has an extremely strong record at 5+3 against people who have challenged it.

I think 15+0 and probably also 10+0 would be a relatively easy win for Magnus based on Awonder, a ~150 elo weaker player, taking two draws at 8+3 and a win and a draw at 10+5. At 5+3 I'm not sure because we have so little data at winnable time controls, but wouldn't expect an easy win for either player.

It's also certainly not the case that these few-months-old networks running a s... (read more)

Veedrac*20

Fundamentally, the story was about the failure cases of trying to make capable systems that don't share your values safe by preventing specific means by which its problem solving capabilities express themselves in scary ways. This is different to what you are getting at here, which is having those systems actually operationally share your values. A well aligned system, in the traditional ‘Friendly AI’ sense of alignment, simply won't make the choices that the one in the story did.

Veedrac42

I was finding it a bit challenging to unpack what you're saying here. I think, after a reread, that you're using ‘slow’ and ‘fast’ in the way I would use ‘soon’ and ‘far away’ (aka. referring to the time it will occur from the present). Is this read about correct?

Veedrac83

If ‘Opt into Petrov Day’ was aside something other than a big red ominous button, I would think the obvious answer is that it's a free choice and I'd be positively inclined towards it. Petrov Day is a good thing with good side effects, quite unlike launching nuclear weapons.

It is confusing to me that it is beside a big red ominous button. On the one hand, Petrov's story is about the value of caution. To quote a top comment from an older Petrov Day,

Petrov thought the message looked legit, but noticed there were clues that it wasn't.

On the other hand, ri... (read more)

2Dagon
If I didn't remember discussion from previous Petrov Days here, I'd probably have skipped it, on the heuristic that big red buttons on public websites are usually advertisements, which i'd rather not deal with.  I doubt there's any visual that could, without a LOT of explanation, make it seem like an ominous thing, rather than  an annoying attention-grab.  With the context of what petrov day is, and what LW has done in the past, it's neither - just an interesting view into what the admins think is worth doing and what users weirdly extrapolate from.
Veedrac1116

I'm interested in having Musk-company articles on LessWrong if it can be done while preserving LessWrong norms. I'm a lot less interested in it if it means bringing in sarcasm, name calling, and ungrounded motive-speculation.

2bhauth
When my crossposted blog posts aren't suitable for LessWrong, I won't be offended if the admins don't frontpage them.
Veedrac113

if, judging by looking at some economical numbers, poverty already doesn't exist for centuries, why do we feel so poor

Let's not forget that people who read LW, often highly intelligent and having well-paying jobs such as software development

This underlines what I find so incongruous about EY's argument. I think I genuinely felt richer as a child eating free school meals in the UK but going to a nice school and whose parents owned a house than I do as an obscenely-by-my-standards wealthy person in San Francisco. I'm hearing this elaborate theory to ex... (read more)

Veedrac40

Eg. a moderately smart person asking it to do something else by trying a few prompts. We're getting better at this for very simple properties but I still consider it unsolved there.

Veedrac*114

Reply to https://twitter.com/krishnanrohit/status/1794804152444580213, too long for twitter without a subscription so I threw it here, but do please treat it like a twitter comment.

rohit: Which part of [the traditional AI risk view] doesn't seem accounted for here? I admit AI safety is a 'big tent' but there's a reason they're congregated together.


You wrote in your list,

the LLMs might start even setting bad objectives, by errors of omission or commission. this is a consequence of their innards not being the same as people (either hallucinations or just not

... (read more)
4Thomas Kwa
What do you mean by "robust to even trivial selection pressure"?
Veedrac183

It took me a good while reading this to figure out whether it was a deconstruction of tabooing words. I would have felt less so if the post didn't keep replacing terms with ones that are both no less charged and also no more descriptive of the underlying system, and then start drawing conclusions from the resulting terms' aesthetics.

With regards to Yudkowsky's takes, the key thing to keep in mind is that Yudkowsky started down his path by reasoning backwards from properties ASI would have, not from reasoning forward from a particular implementation strateg... (read more)

There is no ‘the final token’ for weights not at the final layer.

Because that is where all the gradients flow from, and why the dog wags the tail.

Aggregations of things need not be of the same kind as their constituent things? This is a lot like calling an LLM an activation optimizer. While strictly in some sense true of the pieces that make up the training regime, it's also kind of a wild way to talk about things in the context of ascribing motivation to the resulting network.

I think maybe you're intending ‘next token prediction’ to mean something mor... (read more)

You're at token i in a non-final layer. Which token's output are you optimizing for? i+1?

By construction a decoder-only transformer is agnostic over what future token it should be informative to within the context limit, except in the sense that it doesn't need to represent detail that will be more cheaply available from future tokens.

As a transformer is also unrolled in the context dimension, the architecture itself is effectively required to be generic both in what information it gathers and where that information is used. Bias towards next token predict... (read more)

3gwern
I already addressed this point. If I'm in a non-final layer then I can be optimizing for arbitrary tokens within the context window, sure, and 'effectively' predicting intermediate tokens because that is the 'dominant' effect at that location... insofar as it is instrumentally useful for predicting the final token using the final layer. Because that is where all the gradients flow from, and why the dog wags the tail.

If they appear to care about predicting future tokens, (which they do because they are not myopic and they are imitating agents who do care about future states which will be encoded into future tokens), it is solely as a way to improve the next-token prediction.

I think you're just fundamentally misunderstanding the backwards pass in an autoregressive transformer here. Only a very tiny portion of the model is exclusively trained on next token prediction. Most of the model is trained on what might be called instead, say, conditioned future informativity.

5gwern
I don't think I am. ("conditioned future informativity" - informativity for what? ...the next/last token, which is the only thing taken into account by a causal loss which masks out the rest - that's the definition of it! everything else like packing or doing all the sub-sequences is an optimization and doesn't change the objective.) But feel free to expand on it and explain how the tail wags the dog in causal/decoder Transformers.

I greatly appreciate the effort in this reply, but I think it's increasingly unclear to me how to make efficient progress on our disagreements, so I'm going to hop.

If you say “Indeed it's provable that you can't have a faster algorithm than those O(n^3) and O(n^4) approximations which cover all relevant edge cases accurately” I am quite likely to go on a digression where I try to figure out what proof you're pointing at and why you think it's a fundamental barrier, and it seems now that per a couple of your comments you don't believe it's a fundamental barrier, but at the same time it doesn't feel like any position has been moved, so I'm left rather foggy about where progress has been made.

I think it's very useful th... (read more)

2[anonymous]
.

Thanks, I appreciate the attempt to clarify. I do though think there's some fundamental disagreement about what we're arguing over here that's making it less productive than it could be. For example,

The fact that this has been an extremely active area of research for over 80 years with massive real-world implications, and we're no closer to finding such a simplified heuristic.

I think both:

  1. Lack of human progress doesn't necessarily mean the problem is intrinsically unsolvable by advanced AI. Humans often take a bunch of time before proving things.
  2. It seems n
... (read more)
[anonymous]*102

.

[This comment is no longer endorsed by its author]Reply

And what reason do you have for thinking it can't be usefully approximated in some sufficiently productive domain, that wouldn't also invalidly apply to protein folding? I think it's not useful to just restate that there exist reasons you know of, I'm aiming to actually elicit those arguments here.

3[anonymous]
.

Given Claude is not particularly censored in this regard (in the sense of refusing to discuss the subject), I expect the jailbreak here to only serve as priming.

Well yes, nobody thinks that existing techniques suffice to build de-novo self-replicating nano machines, but that means it's not very informative to comment on the fallibility of this or that package or the time complexity of some currently known best approach without grounding in the necessity of that approach.

One has to argue instead based on the fundamental underlying shape of the problem, and saying accurate simulation is O(n⁷) is not particularly more informative to that than saying accurate protein folding is NP. I think if the claim is that you can... (read more)

4[anonymous]
.

Could you quote or else clearly reference a specific argument from the post you found convincing on that topic?

2[anonymous]
.

Communication overhead won't drop faster than linear.

Which is equivalent to saying if you only care about a situation where none of your observations correlate with any of your other observations and none of your actions interact with any of your observations then your observations are valueless. Which is a true but empty statement, and doesn't meaningfully affect whether there is an optimality axis that it's possible to be better on.

There is a useful generality axis and a useful optimality axis and you can meaningfully progress along both at the same time. If you think no free lunch theorems disprove this then you are confused about no free lunch theorems.

2dkirmani
Whether or not an axis is "useful" depends on your utility function. If you only care about compressing certain books from The Library of Babel, then "general optimality" is real — but if you value them all equally, then "general optimality" is fake. When real, the meaning of "general optimality" depends on which books you deem worthy of consideration. Within the scope of an analysis whose consideration is restricted to the cluster of sequences typical to the Internet, the term "general optimality" may be usefully applied to a predictive model. Such analysis is unfit to reason about search over a design-space — unless that design-space excludes all out-of-scope sequences.

Cybersecurity — this is balanced by being a market question. You invest resources until your wall makes attacks uneconomical, or otherwise economically survivable. This is balanced over time because the defense is ‘people spend time to think about how to write code that isn't wrong.’ In a world where cyber attacks were orders of magnitude more powerful, people would just spend more time making their code and computing infrastructure less wrong. This has happened in the past.

Deaths in conflicts — this is balanced by being a market question. People will only... (read more)

Veedrac*2714

This is a bad analogy. Phoning a human fails dominantly because humans are less smart than the ASI they would be trying to wrangle. Contra, Yudkowsky has even said that were you to bootstrap human intelligence directly, there is a nontrivial shot that the result is good. This difference is load bearing!

This does get to the heart of the disagreement, which I'm going to try to badly tap out on my phone.

The old, MIRI-style framing was essentially: we are going to build an AGI out of parts that are not intrinsically grounded in human values, but rather good ab... (read more)

I liked this! The game was plenty interesting and reasonably introduced. It's a fun twist on induction games with the addition of reasoning over uncertainty rather than exactly guessing a rule, though it does have the downside the relatively small number of samples can make the payoff dominated by randomness.

To offer one small piece of constructive advice on the execution, I did wish the flip history autoscrolled to the newest entry.

I think I implicitly answered you elsewhere, though I'll add a more literal response to your question here.

On a personal level, none of this is relevant to AI risk. Yudkowsky's interest in it seems like more of a byproduct of his reading choices when he was young and impressionable than anything else, which is not reading I shared. Neither he nor I think this is necessary for xrisk scenarios, with me probably being on the more skeptical side, and me believing more in practical impediments that strongly encourage doing the simple things that work, eg. conve... (read more)

Rather than focusing on where I disagree with this, I want to emphasize the part where I said that I liked a lot of the discussion if I frame it in my head differently. I think if you opened the Introduction section with the second paragraph of this reply (“In my post I have explained”), rather than first quoting Yudkowsky, you'd set the right expectations going into it. The points you raise are genuinely interesting, and tons of people have worldviews that this would be much more convincing to than Yudkowsky's.

Veedrac*82

apart from pointing out the actual physical difficulties in doing the thing

This excludes most of the potential good arguments! If you can show that large areas of the solution space seem physically unrealizable, that's an argument that potentially generalizes to ASI. For example, I think people can suggest good limits on how ASI could and couldn't traverse the galaxy, and trivially rule out threats like ‘the AI crashes the moon into Earth’, because of physical argument.

To hypothesize an argument of this sort that might be persuasive, at least to people ... (read more)

I was claiming that titotal's post doesn't appear to give arguments that directly address whether or not Yudkowsky-style ASI can invent diamondoid nanotech. I don't understand the relevance to my comment. I agree that if you find titotal's argument persuasive then whether it is load bearing is relevant to AI risk concerns, but that's not what my comment is about.

FWIW Yudkowsky frequently says that this is not load bearing, and that much seems obviously true to me also.

3[anonymous]
.
Veedrac5021

I'm not sure how to put this, but while this post is framed as a response to AI risk concerns, those concerns are almost entirely ignored in favor of looking at how plausible it is for near-term human research to achieve it, and only at the end is it connected back to AI risk via a brief aside whose crux is basically that you don't think Yudkowsky-style ASI will exist.

I like a lot of the discussion if I frame it in my head to be about what it is actually arguing for. Taking it as given, it seems instead broadly non-sequiter, as the evidence given basically doesn't relate to resolving the disagreement.

3Portia
I think the text is mostly focussed on the problems humans have run into when building this stuff, because these are known and hence our only solid empirical detailed basis, while the problems AI would run into when building this stuff are entirely hypothetical. It then makes a reasonable argument that AI probably won't be able to circumvent these problems, because higher intelligence and speed alone would not plausibly fix them, and in fact, a plausible fix might have to be slow, human-mediated, and practical. One can disagree with that conclusion, but as for the approach, what alternative would you propose when trying to judge AI risk? 
titotal2115

At no point did I ever claim that this was a conclusive debunking of AI risk as a whole, only an investigation into one specific method proposed by Yudkowksy as an AI death dealer.

In my post I have explained what DMS is, why it was proposed as a technology, how far along the research went, the technical challenges faced in it's construction, some observations of how nanotech research works, the current state of nanotech research, what near-term speedups can be expected from machine learning, and given my own best guess on whether an AGI could pull off inve... (read more)

1a3orn133

What would qualify as an evidence against how ASI can do a thing, apart from pointing out the actual physical difficulties in doing the thing?

3faul_sname
My impression is that the nanotech is a load bearing part of the "AI might kill all humans with no warning and no signs ofpreparation" story - specifically, "a sufficiently smart ASI could quickly build itself more computing substrate without having to deal with building and maintaining global supply chains, and doing so would be the optimal course of action" seems like the sort of thing that's probably true if it's possible to build self- replicating nanofactories without requiring a bunch of slow, expensive, serial real-world operations to get the first one built and debugged, and unlikely to be true if not. That's not to say human civilisation is invulnerable, just that "easy" nanotech is a central part of the "everyone dies with no warning and the AI takes over the light cone uncontested" story.

Has there been any serious study of whether mirror life—life with opposite chemical chirality—poses an existential risk?

Veedrac102

Overall this seems like a great general method. I have some critical comments below on the exact implementation, but I want to emphasize upfront that I expect minimal problems in reality, given that, as you mention, LessWrong users are generally both gentle and accurate with downvotes. It is probably not worth making the system more complicated until you see problems in practice.

I don't think I like the -1/-5/-15 recent karma distinction. Consider someone who has reasonable positive reception to their comments, albeit on lower popularity posts, say an aver... (read more)

This doesn't feel like a constructive way to engage with the zeitgeist here. Obviously Yudkowsky plus most people here disagree with you on this. As such, if you want to engage productively on this point, you should find a place better set up to discuss whether NNs uninformatively dead-end. Two such places are the open thread or a new post where you lay out your basic argument.

Veedrac1220

I think this is a pretty good and fair roundup, but I want to add as very lazy bit of personal context short of actually explaining my takes:

Both when I read the FOOM debate, and skimming over it again now, in my personal opinion Yudkowsky largely comes off better. Yudkowsky makes a few major mistakes that are clearly visible now, like being dismissive of dumb, scaled, connectionist architectures, but the arguments seem otherwise repairable. Contra, I do not know how to well defend Hanson's position.

I don't state this to claim a winner, and for sure there are people who read the arguments the other way, but only to suggest to the reader, if you have the time, consider taking a look and forming your own opinion.

1Blueberry
I don't think that's a mistake at all. Sure, they've given us impressive commercial products, but no progress towards AGI, so the dismissiveness is completely justified.

I think signalling to someone that they've missed my intended point is most valuable if they are the kind of person to take it constructively, and if they are, I have no wish to be pointing fingers any more accusatorially than the minimum amount to bring that into focus.

I think a neutral reaction is still a plenty adequate signal in the case you mention and I value that it might do less social harm, whereas a harsher reaction is at least for me less universal as I will be disinclined to use it in prosocial interactions.

I'd be hesitant to label as "Critic

... (read more)

Most of my comments on tone were meant to suggest better phrasings or, in the case of ‘Not what I meant’, iconography, not to suggest they were not valuable.

The specific issue with ‘Not what I meant’ is that the icon reads as ‘you missed’ and not ‘we missed’. Communication is a two-way street and the default react should be at least neutral and non-accusatory.

The section titles were meant very broadly and eg. you'd probably want to put both ‘Locally Valid’ and ‘Locally Invalid’ in that section next to each other even though the former is also Positive. To ... (read more)

4lionhearted (Sebastian Marshall)
Partially agreed again. I'd be hesitant to label as "Critical" pointing out that someone has an invalid argument, and having it implicitly contrasted against "Positive" — it implies they're opposites or antithetical in some way, y'know? Also, respectfully disagree with this -  Sometimes a commentor, especially someone new, is just badly off the mark. That's not a two-way street problem, it's a Well-Kept Garden problem...
Veedrac*251

Quick feedback,

  • The icons aren't all obviously interpretable to me
    • Not a crux — looks like a +, rec: ‘crux’ plus ‘🚫’
    • Please elaborate — unclear that it's a request, rec: ‘..?’
    • What's your prediction — rec: ‘
    • Examples, please — rec: ‘ex?’
    • Additional questions — rec: ‘??’
    • Obtuse — rec: remove
  • Some reactions seem tonally unpleasant:
    • Not what I meant — idea is good, but icon is accusatory
    • I already addressed this — icon is good, text could be ‘addressed elsewhere’
    • Muddled — maybe replace with ‘Unclear’?
    • Obtuse — maybe replace with ‘Too Long’?
    • Not worth getting into
... (read more)
Reply17742111
3dr_s
Yeah, honestly all these pictograms become confusing. Also I am liable to use Elephant not because I have more questions, but as a sign of approval. That was nice, here, have an elephant!
6Elizabeth
I want a bikeshed emoji so bad and if we had line-level reacts I wouldn't have had to write this comment
8lionhearted (Sebastian Marshall)
I agree that drive-by unpleasant criticisms without substance ("Obtuse") don't seem productive, but I actually think some of the mild "tonally unpleasant" ones could be very valuable. It's a way for an author to inexpensively let a commenter know that they didn't appreciate the comment. "Not what I meant" seems particularly valuable for when someone mis-summarizes or inferences wrongly what was written, and "Not worth getting into" seems useful when someone who unproductively deep on a fine-grained detail of something more macro oriented.  One challenge, though, is when you have mixed agreement with someone. I disagree on tonal unpleasantness and the grouping style - "Taboo your words" might be friendly, for instance, to keep sharpening discussion, and isn't necessarily critical. But I agree with a meta/bikeshed and clearing up some of the ambiguous ones.  I clicked both "Disagree" and "Agree" on yours for partial agreement / mixed agreement, but that seems kind of unintuitive.
2Ruby
I agree. See my response to Razied that I think they might have value, and it's interesting to see how they get used in practice. I think there's a world where people abuse them to be mean, and a world where they're used more judiciously. The ability to downvote reacts should also help here, I hope.   I was imagining something like that too. :laugh-react:

The only advantage of a CPU/GPU over an ASIC is that the CPU/GPU is programmable after device creation. If you know what calculation you want to perform you use an ASIC and avoid the enormous inefficiency of the CPU/GPU simulating the actual circuit you want to use

This has a kernel of truth but it is misleading. There are plenty of algorithms that don't naturally map to circuits, because a step of an algorithm in a circuit costs space, whereas a step of an algorithm in a programmable computer costs only those bits required to encode the task. The ineffi... (read more)

A sanity check of a counterintuitive claim can be that the argument to the claim implies things that seem unjustifiable or false. It cannot be that the conclusion of the claim itself is unjustifiable or false, except inasmuch as you are willing to deny the possibility to be convinced of that claim by argument at all.

(To avoid confusion, this is not in response to the latter portion of your comment about general cognition.)

And it says:

So true 8-bit equivalent analog multiplication requires about 100k carriers/switches

This just seems utterly wack.  Having any physical equivalent of an analog multiplication fundamentally requires 100,000 times the thermodynamic energy to erase 1 bit?  And "analog multiplication down to two decimal places" is the operation that is purportedly being carried out almost as efficiently as physically possible by... an axon terminal with a handful of synaptic vesicles dumping 10,000 neurotransmitter molecules to flood around a dendritic ter... (read more)

They typically are uniform, but I think this feels like not the most useful place to be arguing minutia, unless you have a cruxy point underneath I'm not spotting. “The training process for LLMs can optimize for distributional correctness at the expense of sample plausibility, and are functionally different to processes like GANs in this regard” is a clarification with empirically relevant stakes, but I don't know what the stakes are for this digression.

2David Johnston
I was just trying to clarify the limits of autoregressive vs other learning methods. Autoregressive learning is at an apparent disadvantage if P(Xt|Xt−1) is hard to compute and the reverse is easy and low entropy. It can “make up for this” somewhat if it can do a good job of predicting Xt from Xt−2, but it’s still at a disadvantage if, for example, that’s relatively high entropy compared to Xt−1 from Xt. That’s it, I’m satisfied.

The mathematical counterpoint is that this again only holds for sufficiently low entropy completions, which need not be the case, and if you want to make this argument against computronium suns you run into issues earlier than a reasonably defined problem statement does.

The practical counterpoint is that from the perspective of a simulator graded by simulation success, such an improvement might be marginally selected for, because epsilon is bigger than zero, but from the perspective of the actual predictive training dynamics, a policy with a success rate that low is ruthlessly selected against, and the actual policy of selecting the per-token base rate for the hash dominates, because epsilon is smaller than 1/64.

1David Johnston
Are hash characters non uniform? Then I’d agree my point doesn’t stand

This is by construction: I am choosing a task for which one direction is tractable and the other is not. The existence of such tasks follows from standard cryptographic arguments, the specifics of the limiting case are less relevant.

If you want to extrapolate to models strong enough to beat SHA256, you have already conceded EY's point as this is a superhuman task at least relative to the generators of the training data, but anyway there will still exist similar tasks of equal or slightly longer length for which it will hold again because of basic cryptogra... (read more)

1David Johnston
It’s the final claim I’m disputing - that the hashed text cannot itself be predicted. There’s still a benefit to going from e.g. 10−20 to 10−10 probability of a correct hash. It may not be a meaningful difference in practice, but there’s still a benefit in principle, and in practice it could also just generalise a strategy it learned for cases with low entropy text.
Veedrac*70

It is exactly because of the existence of GPT the predictive model, that sampling from GPT is considered simulation; I don't think there's any real tension in the ontology here.

EY gave a tension, or at least a way in which viewing Simulators as a semantic primitive, versus an approximate consequence of a predictive model, is misleading. I'll try to give it again from another angle.

To give the sort of claim worth objecting to, and I think is an easy trap to get caught on even though I don't think the original Simulators post was confused, here is a quote... (read more)

1David Johnston
I can see why your algorithm is hard for GPT — unless it predicts the follow up string perfectly, there’s no benefit to hashing correctly — but I don’t see why it’s impossible. What if it perfectly predicts the follow up?

Ah, well it seems to me that this is mostly people being miscalibrated before GPT-3 hit them over the head about it (and to a lesser extent, even then). You should be roughly likely to update in either direction only in expectation over possible observations. Even if you are immensely calibrated, you should still also a priori expect to have shortening updates around releases and lengthening updates around non-releases, since both worlds have nonzero probability.

But if you'd appreciate a tale of over-expectations, my modal timeline gradually grew for a goo... (read more)

1LGS
Thanks. I agree that in the usual case, the non-releases should cause updates in one direction and releases in the other. But in this case, everyone expected GPT-4 around February (or at least I did, and I'm a nobody who just follows some people on twitter), and it was released roughly on schedule (especially if you count Bing), so we can just do a simple update on how impressive we think it is compared to expectations. Other times where I think people ought to have updated towards longer timelines, but didn't: * Self-driving cars. Around 2015-2016, it was common knowledge that truck drivers would be out of a job within 3-5 years. Most people here likely believed it, even if it sounds really stupid in retrospect (people often forget what they used to believe). I had several discussions with people expecting fully self-driving cars by 2018. * Alpha-Star. When Alpha-star first came out, it was claimed to be superhuman at Starcraft. After fixing an issue with how it clicks in a superhuman way, Alpha-star was no longer superhuman at Starcraft, and to this day there's no bot that is superhuman at Starcraft. Generally, people updated the first time (Starcraft solved!) and never updated back when it turned out to be wrong. * That time when OpenAI tried really hard to train an AI to do formal mathematical reasoning and still failed to solve IMO problems (even when translated to formal mathematics and even when the AI was given access to a brute force algebra solver). Somehow people updated towards shorter timelines even though to me this looked like negative evidence (it just seemed like a failed attempt).
1Noosphere89
This is the best argument against a lot of the fast takeoff stories that I've seen, and it's probably one of the big failure modes of intellectuals to underestimate how much time things take in reality as opposed to their heads.

I'll know how I want to judge it better after I have more data points. I have a page of questions I plan to ask at some point.

With regards to this update specifically, recall both that I thought you thought it would fail the intersection points question when I offered the bet, and that I specifically asked for a reduced-variance version of the bet. Those should tell you something about my probabilities going into this.

6LGS
Fair enough. I look forward to hearing how you judge it after you've asked your questions. I think people on LW (though not necessarily you) have a tendency to be maximally hype/doomer regarding AI capabilities and to never update in the direction of "this was less impressive than I expected, let me adjust my AI timelines to be longer". Of course, that can't be rational, due to  the Conservation of Expected Evidence, which (roughly speaking) says you should be equally likely to update in either direction. Yet I don't think I've ever seen any rationalist ever say "huh, that was less impressive than I expected, let me update backwards". I've been on the lookout for this for a while now; if you see someone saying this (about any AI advancement or lack thereof), let me know.
Load More