quetzal_rainbow

Wiki Contributions

Comments

Sorted by

First season of Game of Thrones was released in 2011, and first book was written in 1996, I think we got rrrrreally desensitized here.

(I am genuinely curious about reasons behind downvotes)

Idea for experiment: take a set of coding problems which have at least two solutions, say, recursive and non-recursive. Prompt LLM to solve them. Is it possible to predict which solution LLM will generate from activations due to first token generation?

If it is possible, it is the evidence against "weak forward pass".

I've seen LLMs generating text backwards. Theoretically, LLM can keep pre-image in activations, calculate hash and then output in order hash, pre-image.

Answer by quetzal_rainbow80

"Frontiers" is known to publish various sorts of garbage. If somebody comes to you to argue some controversial point with article from Frontiers, you can freely assume it to be wrong.

Bandits have obvious cost: if they kill all farmers, from whom are they going to take stuff?

(This text should be understood as written from within the frame of OP: I do not endorse it fully, I just accept the premise and infer conclusions)

The reason why extreme suffering and injustice exist is interpersonal justification of Goodness of Reality hypothesis: the first thing upon hearing it is to dismiss it as naive, romantic, squamish, lacking skin in the game. It is easy to forgive when you have never been wronged.  Enduring extreme suffering gives the possibility to answer: "We suffered and we forgive, why can't you?"

  1. Humans write computer viruses for far more money than price of token generation.
  2. Quoting the bill:

“Critical harm” does not include any of the following: (A) Harms caused or materially enabled by information that a covered model or covered model derivative outputs if the information is otherwise reasonably publicly accessible by an ordinary person from sources other than a covered model or covered model derivative.

I don't understand how one part of the post relates to another. Yeah, sure, computational irreducibility of the world can make understanding of the world impossible and this would be sad. But I don't see what it has to do with "Platonic laws of physics". Current physics, if anything else, is sort of antiplatonic: it claims that there are several dozens of independent entities, actually existing, called "fields", which produce the entire range of observable phenomena via interacting with each other, and there is no "world" outside this set of entities. "Laws of nature" are just "how this entities are". Outside very radical skepticism I don't know any reasons to doubt this worldview.

Multiple comments in one trenchcoat:

I am fine with your iterated mugging? "Forever" without discount implies infinities which is not nice, but we can assume that I'll live 50 years more, so I either need to pay 100*365*50 = 182500$ with delay or get 18250000$ with delay, which sounds like pretty good bet? The only issue I see here is necessity to suspend prior disbelief in such bets and reasonable doubt in honesty of person who does not warn in advance about future payments.

***

Note that betting is a sort of counterfactual mugging in itself: you agree to pay in case of one outcome conditional on payment in case of the other outcome. If resulting outcome is not win and there is no causal enforcement, then your decision to pay or not to pay is similar to decision to pay or not to pay in counterfactual mugging.

***

I think that the simplest way to ignore weird hypotheses is to invoke computing power.

Roughly: there are many worlds that require less computing power to consider which are also more probable than lizard world, so you redirect computing power from considering lizard world to get more utility.

Correspondingly, if lizard world is sufficiently similar to our world and has similar restrictions on computing power, it will likely not consider our world, reducing incentive to analyze it further.

Conversively, if we imagine the world with vastly more computing power than we have, we likely won't be able to analyze it, which incentivizes such world to not try to influence our policies.

To finish, if we imagine the world where it is possible to actually build literal Solomonoff induction which is capable to consider all computable universes in finite time, such world is possibly sufficiently weird to justify considering all possible weird hypotheses, because such world probably has multiple AIXIs which have multiple different utility functions and different priors and run multiple computable copies of you in all sorts of weird circumstances, even if we don't invoke any multiversal weirdness. (It probably sucks to be finite computable structure is such worlds.)

In the end, actual influence on our updateless decisions is limited by very narrow set of worlds, like "different outcomes of real analogs of Prisoner's dilemma with people on the same planet with you" and we don't need to worry about lizard worlds until we solve all our earthly problems and start to disassemble galaxies into computronium.

Load More