All of Seed's Comments + Replies

Seed50

Yudkowsky + Wolfram Debate

Some language to simplify some of the places where the debate got stuck.

Is-Ought

Analyzing how to preserve or act on preferences is a coherent thing to do, and it's possible to do so without assuming a one true universal morality. Assume a preference ordering, and now you're in the land of is, not ought, where there can be a correct answer (highest expected value).

Is There One Reality?

Let existence be defined to mean everything, all the math, all the indexical facts. "Ah, but you left out-" Nope, throw that in too. Everything. Exis... (read more)

Seed20

I can't really get why would one need to know which configuration gave rise to our universe.

This was with respect to feasibility of locating our specific universe for simulation at full fidelity. It's unclear if it's feasible, but if it were, that could entail a way to get at an entire future state of our universe.

I can't see why we would need to "distinguish our world from others"

This was only a point about useful macroscopic predictions any significant distance in the future; prediction relies on information which distinguishes which world we're in.

For n

... (read more)
Seed50

reevaluate how you're defining all the terms that you're using

Always a good idea. As for why I'm pointing to EV: epistemic justification and expected value both entail scoring rules for ways to adopt beliefs. Combining both into the same model makes it easier to discuss epistemic justification in situations with reasoners with arbitrary utility functions and states of awareness.

Knowledge as mutual information between two models induced by some unspecified causal pathway allows me to talk about knowledge in situations where beliefs could follow from arbitra... (read more)

Seed10

Serious thinkers argue for both trying to slow down (PauseAI), and for defensive acceleration (Buterin, Aschenbrenner, etc)

Yeah, I'm in both camps. We should do our absolute best to slow down how quickly we approach building agents, and one way is leveraging AI that doesn't rely on being agentic. It offers us a way to do something like global compute monitoring and could possibly also alleviate short-term incentives satisfiable by building agents, by offering a safer avenue. Insofar as a global moratorium stopping all large model research is feasible, we s... (read more)

Seed30

The problem with that and many arguments for caution is that people usually barely care about possibilities even twenty years out. 

It seems better to ask what would people do if they had more tangible options, such that they could reach a reflective equilibrium which explicitly endorses particular tradeoffs. People mostly pick not caring about possibilities twenty years out due to not seeing how their options constrain what happens in twenty years. This points to not treating their surface preferences as central insofar as they are not following from ... (read more)

3Seth Herd
All good points. I agree that people will care more if their decisions clearly matter in producing that future. This isn't easy to apply to the AGI situation, because what actions will help which outcomes is quite unclear and vigorously argued. Serious thinkers argue for both trying to slow down (PauseAI), and for defensive acceleration (Buterin, Aschenbrenner, etc). And it's further complicated in that many of us think that accelerating will probably produce a better world in a few years, then shortly after that, humanity is dead or sadly obsolete. This pits short-term directly against long-term concerns. I very much agree that helping people imagine either a very good or a very bad future will cause them to care more about it. I think that's been established pretty thoroughly in the decision-making empirical literature. Here I'm reluctant to say more than "futures so good they're difficult to imagine" since the my actual predictions sound like batshit-crazy scifi to most people right now. Sometimes I say things like people won't have to work and global warming will be easy to solve; then people fret about what they'd do with their time if they didn't have to work. I've also tried talking about dramatic health extension, to which people question how much longer they'd want to live any (except old people, who never do - but they're ironically exactly the ones who probably won't benefit from AGI-designed life extension). That's all specific points in agreement with your take that really good outcomes are hard for modern humans to conceive. I agree that describing good futures is worth some more careful thinking. One thought is that it might be easier for most folks to imagine a possible dystopian outcome, in which humans aren't wiped out but made obsolete and simply starve to death when they can't compete with AI wages for any job. I don't think that's the likeliest catastrophe, but it seems possible and might be a good point of focus.
Seed50

He was talking about academic philosophers.

This was a joke referencing academic philosophers rarely being motivated to pick satisfying answers in a time-dependent manner.  

Are you saying that the mechanism of correspondence is an "isomorphism"? Can you please describe what the isomorphism is?

An isomorphism between two systems indicates those two systems implement a common mathematical structure -- a light switch and one's mental model of the light switch are both constrained by having implemented this mathematical structure such that their central beh... (read more)

0Zero Contradictions
I'm not going to respond to most of what you wrote here because I think this will be an unproductive discussion. What I will say is that I think it would help to reevaluate how you're defining all the terms that you're using. Many of your disagreements with the OP essay are semantic in nature. I believe that you will arrive at a richer and more nuanced understanding of epistemology if you learn the definitions used in the OP essay and the author's blog and use those terms to understand epistemology instead. Many of the things that you wrote in your comment seem confused. As for how you're using subjective and objective, I recognize that there are various dictionary definitions for those two terms, but I believe that the most coherent ones that are the most useful for explaining epistemology are the ones that specifically relate to the subject | object dichotomy. You're disagreeing with the statement "knowledge is subjective" because you're not defining "subjective" according to the subject | object dichotomy. I've also written a webpage that might help some of these concepts. You mentioned JTB in your response, and I've written a section explaining why JTB is not an adequate way to define knowledge at all.
Seed*20

While many computations admit shortcuts that allow them to be performed more rapidly, others cannot be sped up.

In your game of life example, one could store larger than 3x3 grids and get the complete mapping from states to next states, reusing them to produce more efficient computations. The full table of state -> next state permits compression, bottoming out in a minimal generating set for next states. One can run the rules in reverse and generate all of the possible initial states that lead to any state without having to compute bottom-up for eve... (read more)

2Clément L
Thanks for your comment. §1 : Okay, you mean something like this, right? I think you're right, maybe the game of life wasn't the best example then. §2 : I think I agree, but I can't really get why would one need to know which configuration gave rise to our universe. § 3 : I'm not sure if i'm answering adequately, but I meant many chaotic phenomena, which probably include stars transforming into supernovae. In that case we arguably can't precisely predict the time of the transformation without fully computing "low level phenomena". But still I can't see why we would need to "distinguish our world from others". For now I'm not sure to see where you're going after that, i'm sorry ! Maybe i'll think about it again and get it later.
Seed3-1

>They leave those questions "to the philosophers"

Those rascals. Never leave a question to philosophers unless you're trying to drive up the next century's employment statistics.

But why would there exist something outside a brain that has the same form as an idea? And even if such facts existed, how would ideas in the mind correspond to them? What is the mechanism of correspondence?

The missing abstraction here is isomorphism. Isomorphisms describe things that can be true in multiple systems simultaneously. How would the behavior of a light switch corresp... (read more)

2Zero Contradictions
He was talking about academic philosophers. Are you saying that the mechanism of correspondence is an "isomorphism"? Can you please describe what the isomorphism is? As the essay explains, knowledge doesn't correspond to reality. Knowledge represents reality. For example, it was considered "true" hundreds of years ago that the Sun revolved around the Earth. Everyone was as strongly convinced that that was a fact as we are now of the opposite belief. Would a Geocentrist be accurate if he confidently claimed that Geocentrism corresponds to reality? Of course not (from our perspective). Instead, what can be accurately said from both perspectives is that the belief that the Sun revolves around the Earth represents reality from the perspective of the people who believe it. If a person believes in the barycentric coordinates theory, then they would judge the Geocentrism theory to be false. They are using a different model of reality to make that judgment. I believe in the barycentric coordinates theory, and that theory represents reality to me, from my perspective. I could claim that barycentric coordinates corresponds to reality, but what does that mean? In practice, people who say "X corresponds to reality" are essentially saying "I think X is true". This is problematic because saying "X is true" is also defined as "X corresponds to reality". The so-called "correspondence" is not well-defined. People can't even agree on what the "correct correspondences" to reality are. A proper theory of knowledge has to explain why a Geocentrist judges the barycentric coordinates theory to be false. The answer is that a geocentrist and a proponent for barycentric coordinates both have different models of reality. Each person believes that their model of reality is true knowledge. To each person, their knowledge represents reality, from their perspective. That's possible because knowledge is subjective. For more information, see: What is Subjectivity? The problem with the Corres
Seed10

You are Elon Musk instead of whoever you actually are.

This is a combination of descriptions only locally accurate in two different worlds and not coherent as a thought experiment asking about the one world fitting those descriptions.

4Seth Herd
Agreed. The premise that physical reality is identical is false, unless you're a hardcore non-physicalist of some sort. Something physical needs to be piping the sensory experience from Musk's body to your brain, because there's little question that your brain is what makes (or whose unfolding pattern, is) your conscious experience.
Seed10

Conditional prediction markets could resolve to the available options weighted by calibration on similar subjects of the holders in unconditional markets, rather than N/A. Such markets might end up looking like predicting what well-calibrated people will pick, or following on after they bet (implying not expecting significant well-calibrated disagreement). Well-calibrated people could then expect to earn a profit by betting in conditional markets if they bet closer to the consensus than the market does, partly weighted in their favor for being better calibrated relative to the whole market.

Seed30

I'm glad you wrote this, it adds some interesting context that was unfamiliar to me for this market I opened around a week ago: https://manifold.markets/dogway/which-is-the-earliest-year-well-hav#wji33pv4fcj

I was entertaining the possibility of a powder or fluid-based metal as an input to a 3D printer which works today for fabricating metal components and seems likely to improve significantly with time. I was considering this avenue to be the most likely way that the threshold of full fidelity-preserving self-reproduction is passed, but I have no expertise... (read more)

4Carl Feynman
Yeah, I looked at various forms of printing from powder as a productive system.  The problem is that the powder is very expensive, more expensive than most of the parts that can be produced from it.  And it can’t produce some parts— like ball bearings or cast iron— so you need tools to make those.  And by the time you add those in, it turns out you don’t need powder metallurgy.
Seed20

I think it's pretty clear that any foundations are also subject to justificatory work

EV is the boss turtle at the bottom of the turtle stack. Dereferencing justification involves a boss battle.

there's some work to be done to make them seem obvious

There's work to show how justification for further things follows from a place where EV is in the starting assumptions, but not to take on EV as an assumption in the first place, as people have EV-calculatingness built into their behaviors as can be noticed to them.

Sometimes—unavoidably, as far as I can tell—those

... (read more)
Seed10

Some beliefs do not normatively require justification;

Beliefs have to be justified on the basis of EV, such that they fit in a particular way into that calculation, and justification comes from EV of trusting the assumptions. Justification could be taken to mean having a higher EV for believing something, and one could be justified in believing things that are false. Any uses of justification to mean something not about EV should end up dissolving; I don't think justification remains meaningful if separated.

Some justifications do not rest on beliefs

Justifi... (read more)

Seed10

Multiple argument chains without repetition can demonstrate anything a circular argument can. No beliefs are constrained when a circular argument is considered relative to the form disallowing repetition (which could avoid costly epicycles). The initial givens imply the conclusion, and they carry through to every point in the argument, implying the whole.

Seed*10

One trusts proofs contextually, as a product of the trusts of the assumptions that led to it in the relevant context. Insofar as Bayesianism requires justification, it can be justified as a dependency in EV calculations. 

We're not going to find a set of axioms which just seem obvious to all humans once articulated.

People understand EV intuitively as a justification for believing things, so this doesn't ring true to me.

The premise A can be contingently true rather than tautologically.

True, I should have indicated I was rejecting it on the basis of repe... (read more)

Seed30

I think it's fair to say that the most relevant objection to circular arguments is that they are not very good at convincing someone who does not already accept the conclusion.

All circular reasoning which is sound is tautological and cannot justify shifting expectation.

The "Regress Argument" in epistemology

The point is, you have to live with at least one of:

No branch of this disjunction applies. Justifications for assumptions bottom out in EV of the reasoning, and so are justified when the EV calculation is accurate. A reasoner can accept less than perfect... (read more)

5abramdemski
Yes, I agree: the essay doesn't really contain a significant argument for this point. "Seem necessary in practice" is more of an observation, a statement of how things seem to me. The closest thing to a positive argument for the conclusion is this: And this, which is basically the same argument: I also cite Eliezer stating a similar conclusion:
2abramdemski
As mentioned in AnthonyC's comment, circular arguments do constrain beliefs: they show that everything in the circle comes as a package deal. Any point in the circle implies the whole.
2abramdemski
I can easily interpret this as falling into branches of the disjunction, and I am not sure how to interpret it as falling into none of the branches. It seems most naturally like a "justification doesn't always rest on further beliefs" type view ("the value of reasoning bottoms out in territory, not map").
3abramdemski
Does your perspective on this also imply that mathematical proofs should never shift one's beliefs? It sounds like you are assuming logical omniscience. Also, it is possible for a circular argument like "A; A -> B; so, B; and also, B -> A; therefore, A" to be sound without being tautological. The implications can be contingently true rather than tautologically. The premise A can be contingently true rather than tautologically.
Seed10

'Self' is "when the other agent is out of range" and 'Other' is "when the other agent is out of range and you see it teleport to a random space". It's unclear to me what reducing the distance between these representations would be doing other than degrading knowledge of the other agent's position. The naming scheme seems to suggest that the agent's distinction of self and other is what is degrading, but that doesn't sound like it's the case. I doubt this sort of loss generalizes to stable non-deceptive behavior in the way that more purely defining the agent's loss in terms of a coalition value for multiple agents that get lower value for being deceived would.

2Marc Carauleanu
You are correct that the self observation happens when the other agent is out of range/when there is no other agent than the self in the observation radius to directly reason about. However, the other observation is actually something more like: “when you, in your current position and environment, observe the other agent in your observation radius”.  Optimising for SOO incentivises the model to act similarly when it observes another agent to when it only observes itself. This can be seen as a distinction between self and other representations being lowered even though I agree that these RL policies likely do not have very expressive self and other representations.  Also agree that in this toy environment, SOO would not perform better than simply rewarding the agents to not be deceptive. However, as the training scenario becomes more complex, eg, by having the other agent be a real human and not being able to trust purely behavioural metrics, SOO is preferred because it does not require being able to incorporate the human’s utility (which is hard to specify) in a shared reward function. Alternatively, it works on the latents of the model, which makes it suitable for the scenarios in which we cannot trust the output of the model. 
Seed32

I appreciate the speculation about this.

redesigning and going through the effort of replacing it isn't the most valuable course of action on the margin. 

Such effort would most likely be a trivial expenditure compared to the resources those actions are about acquiring, and wouldn't be as likely to entail significant opportunity costs as in the case of humans taking those actions, as AIs could parallelize their efforts when needed.

The number of Von Neumann probes one can produce should go up the more planetary material is used, so I'm not sure the adequ... (read more)

2Cole Wyeth
I agree with most of this. I would be modestly surprised, but not very surprised, if an A.G.I. could cause build a Dyson sphere causing the sun to be dimmed by >20% in less than a couple decades (I think a few percent isn't enough to cause crop failure), but within a century is plausible to me.   I don't think we would be squashed for our potential to build a competitor. I think that a competitor would no longer be a serious threat once an A.G.I. seized all available compute. I give a little more credence to various "unknown unknowns" about the laws of physics and the priorities of superintelligences implying that an A.G.I. would no longer care to exploit the resources we need. Overall rationalists are right to worry about being killed by A.G.I. 
Seed10

Human Intelligence Enhancement via Learning:

Intelligence enhancement could entail cognitive enhancements which increase rate / throughput of cognition, increase memory, use of BCI or AI harnesses which offload work / agency or complement existing skills and awareness.

In the vein of strategies which could eventually lead to ASI alignment by leveraging human enhancement, there is an alternative to biological / direct enhancements which attempt to influence cognitive hardware, and instead attempt to externalize one's world model and some of the agency necessa... (read more)

4JBlack
Grabby aliens doesn't even work as an explanation for what it purports to explain. The enormous majority of conscious beings in such a universe-model are members of grabby species who have expanded to fill huge volumes and have a history of interstellar capability going back hundreds of billions of years or more. If this universe model is correct, why is this not what we observe?
Seed10

'Alignment' has been used to refer to both aligning a single AI model, and the harder problem of aligning all AIs. This difference in the way the word alignment is used has led to some confusion. Alignment is not solved by aligning a single AI model, but by using a strategy which prevents catastrophic misalignment/misuse from any AI.

Seed30

The original alignment thinking held that explaining human values to AGI would be really hard.

The difficulty was suggested to be in getting an optimizer to care about what those values are pointing to, not to understand them[1]. If in some instances the values mapped to doing something unwise, using an optimizer that understood those values might fail to constrain away from doing something unwise. Getting a system to use extrapolated preferences as behavioral constraints is a deeper problem than getting a system to reflect surface preferences. The high p(d... (read more)