Wiki Contributions

Comments

Sorted by
rk30

I think this can be resolved by working in terms of packages of property (in this case, uninterrupted ownership of the land), where the value can be greater than the sum of its parts. If someone takes a day of ownership, they have to be willing to pay in excess of the difference between "uninterrupted ownership for 5 years" and "ownership for the first 3 years and 2 days", which could be a lot. Certainly this is a bit of a change from Harberger taxes / needs to allow people to put valuations on extended periods.

It also doesn't really resolve Gwern's case below, where the value to an actor of some property might be less than the amount of value they have custody over via that property.

rk10

To check understanding: if in the first timeline, we use a radiation that doesn't exceed double the heteropneum's EFS, then there remains one timeline. But if we do, there are multiple timelines that aren't distinguishable ... except that the ones with <2x the EFS can't have been the original timeline, because otherwise there wouldn't be branching. I guess I'm confused

rk10

I'm confused by the predictions of death rates for the global population -- seems like that's what would happen only if the 50% of world population is infected all at once. Is it just exponential growth that's doing the work there? I'm also confused about how long contagion is well-modelled as exponential

rk20

To the extent this is a correct summary, I note that it's not obvious to me that agents would sharpen their reasoning skills via test cases rather than establishing proofs on bounds of performance and so on. Though I suppose either way they are using logic, so it doesn't affect the claims of the post

rk90

Here is my attempt at a summary of (a standalone part of) the reasoning in this post.

  • An agent trying to get a lot of reward can get stuck (or at least waste data) when the actions that seem good don't plug into the parts of the world/data stream that contain information about which actions are in fact good. That is, an agent that restricts its information about the reward+dynamics of the world to only its reward feedback will get less reward
  • One way an agent can try and get additional information is by deductive reasoning from propositions (if they can relate sense data to world models to propositions). Sometimes the deductive reasoning they need to do will only become apparent shortly before the result of the reasoning is required (so the reasoning should be fast)
    • The nice thing about logic is that you don't need fresh data to produce test cases: you can make up puzzles! As an agent will need fast deductive reasoning strategies, they may want to try out the goodness of their reasoning strategies on puzzles they invent (to make sure they're fast and reliable (if they hadn't proved reliability))
  • In general, we should model things that we think agents are going to do, because that gives us a handle on reasoning about advanced agents. It is good to be able to establish what we can about the behaviour of advanced boundedly-rational agents so that we can make progress on the alignment problem etc
rk20

Somehow you're expecting to get a lot of information about task B from performance on task A

Are "A" and "B" backwards here, or am I not following?

Answer by rk30

is true iff one of (i) is false or (ii) is true. Therefore, if is some true sentence, for any . Here, is .

rk50

Most of the rituals were created by individuals that did actually understand the real reasons for why certain things had to happen

This is not part of my interpretation, so I was surprised to read this. Could you say more about why you think this? (Either why you think this being argued for in Vaniver's / Scott's posts or why you believe it is fine; I'm mostly interested in the arguments for this claim).

For example, Scott writes:

How did [culture] form? Not through some smart Inuit or Fuegian person reasoning it out; if that had been it, smart European explorers should have been able to reason it out too.

And quotes (either from Scholar's Stage or The Secret of Our Success):

It’s possible that, with the introduction of rice, a few farmers began to use bird sightings as an indication of favorable garden sites. On-average, over a lifetime, these farmers would do better – be more successful – than farmers who relied on the Gambler’s Fallacy or on copying others’ immediate behavior.

Which, I don't read as the few farmers knowing why they should use bird sightings.

Or this quote from Xunzi in Vaniver's post:

One performs divination and only then decides on important affairs. But this is not to be regarded as bringing one what one seeks, but rather is done to give things proper form.

Which doesn't sound like Xunzi understanding the specific importance of a given divination (I realise Xunzi is not the originator of the divinatory practices)

rk10

This link (and the one for "Why do we fear the twinge of starting?") is broken (I think it's an admin view?).

(Correct link)

rk20

Yes, you're quite right!

The intuition becomes a little clearer when I take the following alternative derivation:

Let us look at the change in expected value when I increase my capabilities. From the expected value stemming from worlds where I win, we have . For the other actor, their probability of winning decreases at a rate that matches my increase in probability of winning. Also, their probability of deploying a safe AI doesn't change. So the change in expected value stemming fro m worlds where they win is .

We should be indifferent to increasing capabilities when these sum to 0, so .

Let's choose our units so . Then, using the expressions for from your comment, we have .

Dividing through by we get . Collecting like terms we have and thus . Substituting for we have and thus

Load More