All of TLW's Comments + Replies

TLW10

Interesting.

If you're stating that generic intelligence was not likely simulated, but generic intelligence in our situation was likely simulated...

Doesn't that fall afoul of the mediocrity principle applied to generic intelligence overall?

(As an aside, this does somewhat conflate 'intelligence' and 'computation'; I am assuming that intelligence requires at least some non-zero amount of computation. It's good to make this assumption explicit I suppose.)

1interstice
Sure. I just think we have enough evidence to overrule the principle, in the form of sensory experiences apparently belonging to a member of a newly-arisen intelligent species. Overruling mediocrity principles with evidence is common.
TLW20

I should probably reread the paper.

That being said:

No, it doesn't, any more than "Godel's theorem" or "Turing's proof" proves simulations are impossible or "problems are NP-hard and so AGI is impossible".

I don't follow your logic here, which probably means I'm missing something. I agree that your latter cases are invalid logic. I don't see why that's relevant.

simulators can simply approximate

This does not evade this argument. If nested simulations successively approximate, total computation decreases exponentially (or the Margolus–Levitin theorem doesn't a... (read more)

5gwern
No, it evades the argument by showing that what you take as a refutation of simulations is entirely compatible with simulations. Many impossibility proofs prove an X where people want it to prove a Y, and the X merely superficially resembles a Y. No, it evades the argument by showing that what you take as a refutation of simulations is entirely compatible with simulations. Many impossibility proofs prove an X where people want it to prove a Y, and the X merely superficially resembles a Y. No, it... No, it... No, it... Reminder: you claimed: The simulation argument does not require violating the M-L theorem to the extent it is superficially relevant and resembles an impossibility proof of simulations.
TLW10

Said argument applies if we cannot recursively self-simulate, regardless of reason (Margolus–Levitin theorem, parent turning the simulation off or resetting it before we could, etc).

In order for 'almost all' computation to be simulated, most simulations have to be recursively self-simulating. So either we can recursively self-simulate (which would be interesting), we're rare (which would also be interesting), or we have a non-zero chance we're in the 'real' universe.

2interstice
The argument is not that generic computations are likely simulated, it's about our specific situation -- being a newly intelligent species arising in an empty universe. So simulationists would take the 'rare' branch of your trilemma.
TLW10

Interesting.

I am also skeptical of the simulation argument, but for different reasons.

My main issue is: the normal simulation argument requires violating the Margolus–Levitin theorem[1], as it requires that you can do an arbitrary amount of computation[2] via recursively simulating[3].

This either means that the Margolus–Levitin theorem is false in our universe (which would be interesting), we're a 'leaf' simulation where the Margolus–Levitin theorem holds, but there's many universes where it does not (which would also be interesting), or we have a non... (read more)

9gwern
No, it doesn't, any more than "Godel's theorem" or "Turing's proof" proves simulations are impossible or "problems are NP-hard and so AGI is impossible". There are countless ways to evade this impossibility argument, several of which are already discussed in Bostrom's paper (I think you should reread the paper) eg. simulators can simply approximate, simulate smaller sections, tamper with observers inside the simulation, slow down the simulation, cache results like HashLife, and so on. (How do we simulate anything already...?) All your Margolus-Levitin handwaving can do is disprove a strawman simulation along the lines of a maximally dumb pessimal 1:1 exact simulation of everything with identical numbers of observers at every level.
1interstice
Are you saying that we can't be in a simulation because our descendants might go on to build a large number of simulations themselves, requiring too many resources in the base reality? But I don't think that weakens the argument very much, because we aren't currently in a position to run a large number of simulations. Whoever is simulating us can just turn off/reset the simulation before that happens.
TLW10

Playing less wouldn’t decrease my score

Interesting. Is this typically the case with chess? Humans tend to do better with tasks when they are repeated more frequently, albeit with strongly diminishing returns.

being distracted is one of the effects of stress.

Absolutely, which makes it very difficult to tease apart 'being distracted as a result of stress caused by X causing a drop' and 'being distracted due to X causing a drop'.

3DirectedEvolution
I see what you mean, and yes, that is a plausible hypothesis. It's hard to get a solid number, but glancing over the individual records of my games, it looks like I was playing about as much as usual. Subjectively, it doesn't feel like lack of practice was responsible. I think the right way to interpret my use of "stress" in this context is "the bundle of psychological pressures associated with exam season," rather than a psychological construct that we can neatly distinguish from, say, distractability or sleep loss. It's kind of like saying "being on an ocean voyage with no access to fresh fruits and vegetables caused me to get scurvy."
TLW390

solar+batteries are dropping exponentially in price

Pulling the data from this chart from your source:

...and fitting[1] an exponential trend with offset[2], I get:

(Pardon the very rough chart.)

This appears to be a fairly good fit[3], and results in the following trend/formula:

[4]

This is an exponentially-decreasing trend... but towards a decidedly positive horizontal asymptote.

This essentially indicates that we will get minimal future scaling, if any. $37.71/MWh is already within the given range.

For reference, he... (read more)

TLW10

Too bad. My suspects for confounders for that sort of thing would be 'you played less at the start/end of term' or 'you were more distracted at the start/end of term'.

4DirectedEvolution
Playing less wouldn’t decrease my score, and being distracted is one of the effects of stress.
TLW80

First, nuclear power is expensive compared to the cheapest forms of renewable energy and is even outcompeted by other “conventional” generation sources [...] The consequence of the current price tag of nuclear power is that in competitive electricity markets it often just can’t compete with cheaper forms of generation.

[snip chart]

Source: Lazard

This estimate does not seem to include capacity factors or cost of required energy storage, assuming I read it correctly. Do you have an estimate that does?

2PeterMcCluskey
Paul Christiano made some estimates here.
2Marius Hobbhahn
Unfortunately, we didn't find a good source for that. However, given that fossils usually don't need storage and solar+batteries are dropping exponentially in price, we think both options should be cheaper. But good estimates of that would be very welcome. 
TLW70

Thirdly, nuclear power gives you energy independence. This became very clear during Russia’s invasion of Ukraine. France, for example, had much fewer problems cutting ties with Russia than e.g. Germany. While countries might still have to import Uranium, the global supplies are distributed more evenly than with fossil fuels, thereby decreasing geopolitical relevance. Uranium can be found nearly everywhere.

Also, you can extract uranium from seawater. This has its own problems, and is still more expensive than mines currently. However, this puts a cap on the... (read more)

TLW10

How much did you play during the start / end of term compared to normal?

3DirectedEvolution
I don’t know exactly, Lichess doesn’t have a convenient way to plot that day by day. But probably roughly equal amounts. It’s my main distraction.
TLW20

Here's an example game tree:

(Kindly ignore the zeros below each game node; I'm using the dev version of GTE, which has a few quirks.)

Roughly speaking:

  1. Bob either has something up his sleeve (an exploit of some sort), or does not.
  2. Bob either:
    1. Offers a flat agreement (well, really surrender) to Alice.
    2. Offers a (binding once both sides agree to it) arbitrated (mediated in the above; I am not going to bother redoing the screenshot above) agreement by a third party (Simon) to Alice.
    3. Goes directly to war with Alice.
  3. Assuming Bob didn't go directly to war, Alice either
... (read more)
TLW10

As TLW's comment notes, the disclosure process itself might be really computationally expensive.

I was actually thinking of the cost of physical demonstrations, and/or the cost of convincing others that simulations are accurate[1], not so much direct simulation costs.

That being said, this is still a valid point, just not one that I should be credited for.

  1. ^

    Imagine trying to convince someone of atomic weapons purely with simulations, without anyone ever having detonated one[2], for instance. It may be doable; it'd be nowhere near cheap.

    Now imagine trying to d

... (read more)
TLW20

Ignore the suit of the cards. So you can draw a 1 (Ace) through 13 (King). Pulling two cards is a range of 2 to 26. Divide by 2 and add 1 means you get the same roll distribution as rolling two dice.

That's not the same roll distribution as rolling two dice[1]. For instance, rolling a 14 (pulling 2 kings) has a probability of , not [2].

(The actual distribution is weird. It's not even symmetrical, due to the division (and associated floor). Rounding to even/odd would help this, but would cause other issues.)

This also ... (read more)

TLW20

Interesting.

The other player gets to determine your next dice roll (again, either manually or randomly).

Could you elaborate here?

Alice cheats and say she got a 6. Bob calls her on it. Is it now Bob's turn, and hence effectively a result of 0? Or is it still Alice's turn? If the latter, what happens if Alice cheats again?

I'm not sure how you avoid the stalemate of both players 'always' cheating and both players 'always' calling out the other player.

Instead of dice, a shuffled deck of playing cards would work better. To determine your dice roll, just pull tw

... (read more)
1MackGopherSena
[edited]
TLW10

Interesting!

How does said binding treaty come about? I don't see any reason for Alice to accept such a treaty in the first place.

Alice would instead propose (or counter-propose) a treaty that always takes the terms that would result from the simulation according to Alice's estimate.

Alice is always at least indifferent to this, and the only case where Bob is not at least indifferent to this is if Bob is stronger than Alice's estimate, in which case accepting said treaty would not be in Alice's best interest. (Alice should instead stall and hunt for exploits, give or take.)

1benjamincosman
Simon's services are only offered to those who submit a treaty over a full range of possible outcomes. Alice could try to bully Bob into accepting a bullshit treaty ("if I win you give me X; if I lose you still give me X"), but just like today Bob has the backup plan of refusing arbitration and actually waging war. (Refusing arbitration is allowed; it's only going to arbitration and then not abiding by the result that is verboten.) And Alice could avoid committing to concessions-on-loss by herself refusing arbitration and waging war, but she wouldn't actually be in a better position by doing so, since the real-world war also involves her paying a price in the (to her mind) unlikely event that Bob wins and can extract one. Basically the whole point is that, as long as the simulated war and the agreed-upon consequences of war (minus actual deaths and other stuff) match a potential real-world war closely enough, then accepting the simulation should be a strict improvement for both sides regardless of their power differential and regardless of the end result, so both sides should (bindingly) accept.
TLW20

Let's look at a relatively simple game along these lines:

Person A either cheats an outcome or rolls a d4. Then person B either accuses, or doesn't. If person B accuses, the game ends immediately, with person B winning (losing) if their accusation was correct (incorrect). Otherwise, repeat a second time. At the end, assuming person B accused neither time, person A wins if the total sum is at least 6. (Note that person A has a lower bound of a winrate of 3/8ths simply by never cheating.)

Let's look at the second round first.

First subcase: the first roll (legi... (read more)

TLW10

What's the drawback to always accusing here?

1MackGopherSena
[edited]
TLW20

Though this does suggest a (unrealistically) high-coordination solution to at least this version of the problem: have both sides declare all their capabilities to a trusted third party who then figures out the likely costs and chances of winning for each side.

Is that enough?

Say Alice thinks her army is overwhelmingly stronger than Bob. (In fact Bob has a one-time exploit that allows Bob to have a decent chance to win.) The third party says that Bob has a 50% chance of winning. Alice can then update P(exploit), and go 'uhoh' and go back and scrub for exploits.

(So... the third-party scheme might still work, but only once I think.)

5benjamincosman
Good point, so continuing with the superhuman levels of coordination and simulation: instead of Alice and Bob saying "we're thinking of having a war" and Simon saying "if you did, Bob would win with probability p"; Alice and Bob say "we've committed to simulating this war and have pre-signed treaties based on various outcomes", and then Simon says "Bob wins with probability p by deploying secret weapon X, so Alice you have to pay up according to that if-you-lose treaty". So Alice does learn about the weapon but also has to pay the price for losing, exactly like she would in an actual war (except without the associated real-world damage).
TLW20

Conversely, if FDR wants a chicken in every pot, and then finds out that chickens don't exist, he would change his values to want a beef roast in every pot, or some such.

I do not believe his value function is "a chicken in every pot". It's likely closer to 'I don't want anyone to be unable to feed themselves', although even this is likely an over-approximation of the true utility function. 'A chicken in every pot' is one way of doing well on said utility function. If he found out that chickens didn't exist, the 'next best thing' might be a roast beef in ev... (read more)

TLW10

Demonstrating military strength is itself often a significant cost.

Say your opponent has a military of strength 1.1x, and is demonstrating it.

If you have the choice of keeping and demonstrating a military of strength x, or keeping a military of strength 1.2 and not demonstrating at all...

2benjamincosman
As an example, part of your military strength might be your ability to crash enemy systems with zero-day software exploits (or any other kind of secret weapon they don't yet have counters for). At least naively, you can't demonstrate you have such a weapon without rendering it useless. Though this does suggest a (unrealistically) high-coordination solution to at least this version of the problem: have both sides declare all their capabilities to a trusted third party who then figures out the likely costs and chances of winning for each side.
TLW10

If you allow the assumption that your mental model of what was said matches what was said, then you don't necessarily need to read all the way through to authoritatively say that the work never mentions something, merely enough that you have confidence in your model.

If you don't allow the assumption that your mental model of what was said matches what was said, then reading all the way through is insufficient to authoritatively say that the work never mentions something.

(There is a third option here: that your mental model suddenly becomes much better when you finish reading the last word of an argument.)

TLW20

I want to clarify that this is not a particularly useful type of utility function, and the post was a mostly-failed attempt to make it useful.

Fair! Here's another[1] issue I think, now that I've realized you were talking about utility functions over behaviours, at least if you allow 'true' randomness.

Consider a slight variant of matching pennies: if an agent doesn't make a choice, their choice is made randomly for them.

Now consider the following agents:

  1. Twitchbot.
  2. An agent that always plays (truly) randomly.
  3. An agent that always plays the best Nash equil
... (read more)
TLW10

Interesting!

Could you please explain why your arguments don't apply to compilers?

1Michaël Trazzi
the two first are about data, and as far as I know compilers do not use machine learning on data. third one could technically apply to compilers, though I think in ML there is a feedback loop "impressive performance -> investments in scaling -> more research", but you cannot just throw more compute to increase compiler performance (and results are less in the mainstream, less of a public PR thing)
TLWΩ120

You would get a 1.01 multiplier in productivity, that would make the speed of development 1.01x faster, especially the development of a Copilot-(N+1),

...assuming that Copilot-(N+1) has <1.01x the development cost as Copilot-N. I'd be interested in arguments as to why this would be the case; most programming has diminishing returns where e.g. eking out additional performance from a program costs progressively more development time.

2Michaël Trazzi
Some arguments for why that might be the case: -- the more useful it is, the more people use it, the more telemetry data the model has access to -- while scaling laws do not exhibit diminishing returns from scaling, most of the development time would be on things like infrastructure, data collection and training, rather than aiming for additional performance -- the higher the performance, the more people get interested in the field and the more research there is publicly accessible to improve performance by just implementing what is in the litterature (Note: this argument does not apply for reasons why one company could just make a lot of progress without ever sharing any of their progress.)
TLW10

That's a very different definition of utility function than I am used to. Interesting.

What would the utility function over behaviors for an agent that chose randomly at every timestep look like?

2Thomas Kwa
My guess is if the randomness is pseudorandom, then 1 for the behavior it chose and 0 for everything else; if the randomness is true randomness and we use Boltzmann rationality then all behaviors are equal utility; if the randomness is true and the agent is actually maximizing, then the abstraction breaks down? I want to clarify that this is not a particularly useful type of utility function, and the post was a mostly-failed attempt to make it useful.
TLW20

(Disclaimer: not my forte.)

CLIP’s task is to invent a notation system that can express the essence of (1) any possible picture, and (2) any possible description of a picture, in only a brief list of maybe 512 or 1024 floating-point numbers. 

How many bits is this? 2KiB / 16Kib? Other?

Has there been any work in using this or something similar as the basis of a high-compression compression scheme? Compression and decompression speed would be horrendous, but still.

Hm. I wonder what would happen if you trained a version on internet imag... (read more)

Ah, I now realize that I was kind of misleading in the sentence you quoted.  (Sorry about that.)

I made it sound like CLIP was doing image compression.  And there are ML models that are trained, directly and literally to do image compression in a more familiar sense, trying to get the pixel values as close to the original as possible.  These are the image autoencoders.

DALLE-2 doesn't use an autoencoder, but many other popular image generators do, such as VQGAN and the original DALLE.

So for example, the original DALLE has an autoencoder compon... (read more)

TLW30

Why is this style of pessimism repeatedly wrong?

Beware selection bias. If it wasn't repeatedly wrong, there's a good chance we wouldn't be here to ask the question!

The opposite view is that progress is a matter of luck.

Hm. I tend to not view the pessimistic side as luck so much as 'there's a finite number of useful techs, which we are rapidly burning through'.

TLW10

I don't think I make this assumption.

You don't explicitly; it's implicit in the following:

It is well known that a utility function over behaviors/policies is sufficient to describe any policy.

The VNM axioms do not necessarily apply for bounded agents. A bounded agent can rationally have preferences of the form A ~[1] B and B ~ C but A ≻[2] C, for instance[3]. You cannot describe this with a straight utility function.

  1. ^

    is indifferent to

  2. ^

    is preferred over

  3. ^
... (read more)
1Thomas Kwa
I agree that a bounded agent can be VNM-incoherent and not have a utility function over bettable outcomes. Here I'm saying you can infer a utility function over behaviors for *any* agent with *any* behavior. You can trivially do this by setting the utility gained by every action the agent actually takes to 1, and utility of every action the agent doesn't take to 0. For example for twitch-bot, the utility at each step is 1 if it twitches and 0 if it doesn't.
TLW10

Kudos for not naively extrapolating past 100% of GDP invested into AI research.

 Reduced Cost of Computation: Estimated to reduce by 50% every 2.5 years (about in-line with current trends), down to a minimum level of 1 / 106 (i.e., 0.0001%) in 50 years.

Increased Availability of Capital for AI: Estimated to reach a level of $1B in 2025, then double every 2 years after that, up to 1% of US GDP (currently would suggest $200B of available capital, and growing ~3% per year).

Our current trends in cost of computation are in combination with (seemingly) expone... (read more)

TLW10

All of this is predicated on the agent having unlimited and free access to computation.

This is a standard assumption, but is worth highlighting.

1Thomas Kwa
I don't think I make this assumption. The biggest flaw in this post is that some of the definitions don't quite make sense, and I don't think assuming infinite compute helps this.
TLW10

endowed with the of probability calculation

I suspect dropped a word in that sentence.

Maybe "endowed with the power of probability calculation", or somesuch?

2Donta Lee
Many thanks for pointing out the crack!
TLW30

So, uh, how should we start preparing for the impact of this here in the West?

Push officials to redirect corn subsidies for ethanol toward food? Keeping American corn production alive in case it suddenly was required was one of the reasons for said subsidies[1], after all.

(Of course, this doesn't help the upstream portions of the food supply chain.)

  1. ^

    ...to the best of my knowledge, although I note that I can't find anything explicitly mentioning this at a quick look.

TLW10

Alright, we are in agreement on this point.

I have a tendency to start on a tangential point, get agreement, then show the implications for the main argument. In practice a whole lot more people are open to changing their minds in this way than directly. This may be somewhat less important on this forum than elsewhere.

You stated:

In contrast, I think almost all proponents of libertarian free will would agree that their position predicts that an agent with such free will, such as a human, could always just choose to not do as they are told. If the distributio

... (read more)
1Ege Erdil
I don't know why you're talking about the Newcomb problem again. I've already said I don't see how that's relevant. Can you tell me how, in my setup, the fixed point being 50/50 means the oracle has no predictive power over the agent? If 50/50 is a fixed point then the agent clearly has predictive power, just like we have predictive power over what happens if you measure a qubit in the state (|0⟩+|1⟩)/√2. 50% doesn't imply "lack of predictive power".
TLW00

It seems to me that you're (intentionally or not) trying to find mistakes in the post.

It is obvious we have a fundamental disagreement here, and unfortunately I doubt we'll make progress without resolving this.

Euclid's Parallel Postulate was subtly wrong. 'Assuming charity' and ignoring that would not actually help. Finding it, and refining the axioms into Euclidean and non-Euclidean geometry, on the other hand...

TLW10

My point is that the fact that Omega can guess 50/50 in the scenario I set up in the post doesn't allow it to actually have good performance

That... is precisely my point? See my conclusion:

If the fixedpoint was 50/50 however, the fixedpoint is still satisfied by Omega putting in the money 50% of the time, but Omega is left with no predictive power over the outcome .

(Emphasis added.)

1Ege Erdil
Then I don't understand what you're trying to say here. Can you explain exactly what you think the problem with my setup is? I think you're just not understanding what I'm trying to say. My point is that if Omega actually has no knowledge, then predicting 50/50 doesn't allow him to be right. On the other hand, if 50/50 is actually a fixed point of g, then Omega predicting 50/50 will give him substantial predictive power over outcomes. For example, it will predict that my actions will be roughly split 50/50, when there's no reason for this to be true if Omega has no information about me; I could just always pick an action I've predetermined in advance whenever I see a 50/50 prediction.
TLW10

Paxlovid itself is free at point-of-sale (or at least the current supply is), so there's no monetary concern.

See "paxlovid is free".

Admittedly, I am not super familiar with the US hospital system, but I believe Paxlovid is only available by prescription - which itself is a nontrivial cost for many.

These patients have already been tested.

Not only did they get tested but they tested positive.

They already have covid, they aren't being asked to do something unusual and unpleasant to prevent an uncertain future harm.

They're literally being offered a prescriptio

... (read more)
2tkpwaeub
A possible reason for being cautious about prescribing Paxlovid to just anyone (regardless of underlying risk factors and severity) from a public health standpoint is the issue of rebounds. The Paxlovid might hold the infection at bay for a while, and then people could start testing positive again a few days later, so quarantine protocols become a bit more complex. https://www.bostonglobe.com/2022/04/21/metro/puzzling-phenomenon-patients-report-rebound-covid-19-symptoms-after-taking-antiviral-paxlovid/
TLW30

Another case that violates the preconditions is if the information source is not considered to be perfectly reliable.

Imagine the following scenario:

Charlie repeatedly flips a coin, and tells person A and B the results.

Alice and Bon are choosing between the following hypotheses:

  1. The coin is fair.
  2. The coin always comes up heads.
  3. The coin is fair, but person C only reports when the coin comes up heads.

Alice has a prior of 40% / 40% / 20%. Bob has a prior of 40% / 20% / 40%.

Now, imagine that Charlie repeatedly reports 'heads'. What happens?

Answer: Alice asymptote... (read more)

TLW50

I can only conclude that most people do not much care about Covid.

...or people don't have the resources to be able to afford Paxlovid.
...or 'have you had a positive test in the past 14 days; if so you cannot work'-style questions, especially combined with nontrivial false-positive rates, have lead people to avoid getting tested in the first place.
...or people think that the cost to get tests outweighs the expected benefit.
...or people think that the cost to get Paxlovid outweighs the expected benefit.
...or people have started to discount 'do X to stop Covi... (read more)

1RobertM
These seem mostly non-responsive to the described situation. * Paxlovid itself is free at point-of-sale (or at least the current supply is), so there's no monetary concern. * These patients have already been tested. * Not only did they get tested but they tested positive. * See "paxlovid is free". * They already have covid, they aren't being asked to do something unusual and unpleasant to prevent an uncertain future harm. * They're literally being offered a prescription by a doctor and turning it down.
TLW10

If Omega predicts even odds on two choices and then you always pick one you've determined in advance

(I can't quite tell if this was intended to be an example of a different agent or if it was a misconstrual of the agent example I gave. I suspect the former but am not certain. If the former, ignore this.) To be clear: I meant an agent that flips a quantum coin to decide at the time of the choice. This is not determined, or determinable, in advance[1]. Omega can predict  here fairly easily, but not .

There's a big difference between those two

... (read more)
1Ege Erdil
It's intended to be an example of a different agent. I don't care much about the Newcomb problem setting since I think it's not relevant in this context. My point is that the fact that Omega can guess 50/50 in the scenario I set up in the post doesn't allow it to actually have good performance and it's easy to tell this by running any standard hypothesis test. So I don't get how your Newcomb setup relates to my proposed setup in the post.
TLW00

I think this is a terminological dispute

Fair.

and is therefore uninteresting.

Terminology is uninteresting, but important.

There is a false proof technique of the form:

  1. A->B
  2. Example purportedly showing A'
  3. Therefore B
  4. [if no-one calls them out on it, declare victory. Otherwise wait a bit.]
  5. No wait, the example shows A (but actually shows A'')
  6. [if no-one calls them out on it, declare victory. Otherwise wait a bit.]
  7. No wait, the example shows A (but actually shows A''')
  8. [if no-one calls them out on it, declare victory. Otherwise wait a bit.]
  9. No wait, the example shows
... (read more)
1Ege Erdil
As far as I can see my argument was clear from the start and nobody seems to have been confused by this point of it other than you. I'll admit I'm wrong if some people respond to my comment by saying that they too were confused by this point of my argument and your comment & my response to it helped clear things up for them. It seems to me that you're (intentionally or not) trying to find mistakes in the post. I've seen you do this in other posts as well and have messaged you privately about it, but since you said you'd rather discuss this issue in public I'm bringing it up here. Any post relies on some amount of charity on the part of the reader to interpret it correctly. It's fine if you're genuinely confused about what I'm saying or think I've made a mistake, but your behavior seems more consistent with a fishing expedition in which you're hunting for technically wrong statements to pick on. This might be unintentional on your part or a reflection of my impatience with this kind of response, but I find it exhausting to have to address these kinds of comments that appear to me as if they are being made in bad faith.
TLW10

 being continuous does not appear to actually help to resolve predictor problems, as a fixed point of 50%/50% left/right[1] is not excluded, and in this case Omega has no predictive power over the outcome [2].

If you try to use this to resolve the Newcomb problem, for instance, you'll find that an agent that simply flips a (quantum) coin to decide does not have a fixed point in , and does have a fixed point in , as expected, but said fixed point is 50%/50%... which means the Omega is wrong exactly half the time. You could replac... (read more)

1Ege Erdil
You're overcomplicating the problem. If Omega predicts even odds on two choices and then you always pick one you've determined in advance, it will be obvious that Omega is failing to predict your behavior correctly. Imagine that you claim to be able to predict the probabilities of whether I will choose left or right, and then predict 50% for both. If I just choose "left" every time then obviously your predictions are bad - you're "well calibrated" in the sense that 50% of your 50% predictions come true, but the model that both of my choices have equal probability will just be directly rejected. In contrast, if this is actually a fixed point of g, I will choose left about half the time and right about half the time. There's a big difference between those two cases and I can specify an explicit hypothesis test with p-values, etc. if you want, though it shouldn't be difficult to come up with your own.
TLW10

The result is that the function  is continuous

Absolutely. I continue to agree with you on this.

The part I was noting was incorrect was "our understanding of physics seems to exclude effects that are truly discontinuous." (emphasis added).

 is continuous.  is not[1].

 

  1. ^

    Or rather, may not be.

1Ege Erdil
I think this is a terminological dispute and is therefore uninteresting. My argument only requires g to be continuous and nothing else.
TLW*240

These are mostly combinations of a bunch of lower-confidence arguments, which makes them difficult to expand a little. Nevertheless, I shall try.

1. I remain unconvinced of prompt exponential takeoff of an AI.

...assuming we aren't in Algorithmica[1][2]. This is a load-bearing assumption, and most of my downstream probabilities are heavily governed by P(Algorithmica) as a result.

...because compilers have gotten slower over time at compiling themselves.
...because the optimum point for the fastest 'compiler compiling itself' is not to turn on all optimizations... (read more)

3Heighn
Wow, thank you so much! (And I apologize for the late reaction.) This is great, really. Indeed. It seems FOOM might require a world in which processors can be made arbitrarily tiny. I have been going over all your points and find them all very interesting. My current intuition on "recursive self-improvement" is that deep learning may be about the closest thing we can get, and that performance of those will asymptote relatively quickly when talking about general intelligence. As in, I don't expect it's impossible to have super-human deep learning systems, but I don't expect with high probability that there will be an exponential trend smashing through the human level.
TLW40

For anyone else that can't read this quickly, this is what it looks like, un-reversed:

The rain in Spain stays mainly in the get could I Jages' rain in and or rain in the quages or, rain in the and or rain in the quages or rain in the and or rain in the quages or, rain in the and or rain in the quages or rain in the and or rain in the quages or rain in the and or rain in the quages or rain in the and or rain in the quages or rain in the and or rain in the quages or rain in the and or rain in the quages or rain in the and or rain in the quages or rain in the

... (read more)
6Dustin
The rain in Spain stays mainly in the plain.  This is a famous line from the play Pygmalion by George Bernard Shaw. In the play, a character named Henry Higgins is teaching a lower-class woman named Eliza Doolittle how to speak proper English. He tells her that the rain in Spain stays mainly in the plain in order to help her remember the correct pronunciation of the word "plain."
TLWΩ570

Ah, but perhaps your objection is that the difficulty of the AI alignment problem suggests that we do in fact need the analog of perfect zero correlation in order to succeed.

My objection is actually mostly to the example itself.

As you mention:

the idea is not to try ang contain a malign AGI which is already not on our side. The plan, to the extent that there is one, is to create systems that are on our side, and apply their optimization pressure to the task of keeping the plan on-course.

Compare with the example:

Suppose we’re designing some secure electronic

... (read more)
7abramdemski
Fair enough! I admit that John did not actually provide an argument for why alignment might be achievable by "guessing true names". I think the approach makes sense, but my argument for why this is the case does differ from John's arguments here.
TLW10

Sorry, are we talking about effects that are continuous, or effects that are discontinuous but which have probability distributions which are continuous?

I was rather assuming you meant the former considering you said 'effects that are truly discontinuous.'.

Both of your responses are the latter, not the former, assuming I am understanding correctly.

*****

You can't get around the continuity of unitary time evolution in QM with these kinds of arguments.

And now we're into the measurement problem, which far better minds than mine have spent astounding amounts of effort on and not yet resolved. Again, assuming I am understanding correctly.

1Ege Erdil
We're talking about the continuity of the function g. I define it in the post, so you can check the post to see exactly what I'm talking about. This has nothing to do with how you settle the measurement problem. As I say in the post, a quantum Turing machine would have the property that this g I've defined is continuous, even if it's a version that can make measurements on itself mid-computation. That doesn't change the fact that g is continuous, roughly because the evolution before the measurement is unitary, and so perturbing the initial state by a small amount in L2-norm perturbs the probabilities of collapsing to different eigenstates by small amounts as well. The result is that the function g:M→M is continuous, even though the wave-function collapse is a discontinuous operation on the Hilbert space of states. The conclusion generalizes to any real-world quantum system, essentially by the same argument.
TLW00

I think this is interesting because our understanding of physics seems to exclude effects that are truly discontinuous.

This is not true. An electron and a positron will, or will not, annihilate. They will not half-react.

For example, real-world transistors have resistance that depends continuously on the gate voltage

This is incorrect. It depends on the # of electrons, which is a discrete value. It's just that most of the time transistors are large enough that it doesn't really matter. That being said, it's absolutely important for things like e.g. flash mem... (read more)

1Ege Erdil
The Feynman diagrams of that process give you a scattering amplitude which will tell you the rate at which that process is going to occur. The probability of it occuring will be a continuous as a function on the Hilbert space. In quantum mechanics, even if the states of a system are quantized/discrete, the probability of you being in those states behaves continuously under unitary evolution or collapse. You can't get around the continuity of unitary time evolution in QM with these kinds of arguments.
TLW30

In any case, my world model says that an AGI should actually be able to recursively self-improve before reaching human-level intelligence. Just as you mentioned, I think the relevant intuition pump is "could I FOOM if I were an AI?" Considering the ability to tinker with my own source code and make lots of copies of myself to experiment on, I feel like the answer is "yes."

Counter-anecdote: compilers have gotten ~2x better in 20 years[1], at substantially worse compile time. This is nowhere near FOOM.

  1. ^

    Proebsting's Law gives an 18-year doubling time. The 200

... (read more)
TLW20

It may be worthwhile to extend this to doctors who aren't perfect - that is are only correct most of the time - and see what happens.

Load More