I very much appreciate this post, because it strongly resonates with my own experience of laziness and willpower. Reading this post feels less like learning something new and more like an important reminder.
This is not quite accurate. You can't uniformly pick a random rational number from 0 to 1, because there are countably many such numbers, and any probability distribution you assign will have to add up to 1. Every probability distribution on this set assigns a nonzero probability to every number.
You can have a uniform distribution on an uncountable set, such as the real numbers between 0 and 1, but since you can't pick an arbitrary element of an uncountable set in the real world this is theoretical rather than a real-world issue.
As far as I know, any mathematical case in which something with probability 0 can happen does not actually occur in the real world in a way that we can observe.
On the other hand, the more you get accustomed to a pleasurable stimulus, the less pleasure you receive from it over time (hedonic adaptation). Since this happens to both positive and negative emotions, it seems to me that there is a kind of symmetry here. To me this suggests that decreasing prediction error results in more neutral emotional states rather than pleasant states.
I disagree that all prediction error equates to suffering. When you step into a warm shower you experience prediction error just as much as if you step into a cold shower, but I don't think the initial experience of a warm shower contains any discomfort for most people, whereas I expect the cold shower usually does.
Furthermore, far more prediction error is experienced in life than suffering. Simply going for a walk leads to a continuous stream of prediction error, most of which people feel pretty neutral about.
This reminds me of a lot of discussions I've had with people where we seem to be talking past each other, but can't quite pin down what the disagreement is.
Usually we just end up talking about something else instead that we both seem to derive value from.
It seems to me that the constraints of reality are implicit. I don't think "it can be done by a human" is satisfied by a method requiring time travel with a very specific form of paradox resolution. It sounds like you're arguing that the Church-Turing thesis is simply worded ambiguously.
It looks Deutschian CTCs are similar to a computer that can produce all possible outputs in different realities, then selectively destroy the realities that don't solve the problem. It's not surprising that you could solve the halting problem in such a framework.
Our symbolic conception of numbers is already logarithmic, as order of magnitude corresponds to the number of digits. I think an estimate of a product based on an imaginary slide rule would be roughly equivalent to estimating based on the number of digits and the first digit.
Similar to point 2: I find that reading a book in the morning helps my mood. Particularly a physical fiction book.
I've definitely noticed the pattern of habits seeming to improve my life without them feeling like they are improving my life. On a similar note, a lot of habits seem easy to maintain while I'm doing them and obviously beneficial, but when I stop I have no motivation to continue. I don't know why that is, but my hope is that if I notice this hard enough it will become easier for me to recognize that I should do the thing anyway.
I read some of the post and skimmed the rest, but this seems to broadly agree with my current thoughts about AI doom, and I am happy to see someone fleshing out this argument in detail.
[I decided to dump my personal intuition about AI risk below. I don't have any specific facts to back it up.]
It seems to me that there is a much larger possibility space of what AIs can/will get created than the ideal superintelligent "goal-maximiser" AI put forward in arguments for AI doom.
The tools that we have depend more on the specific details of the underlying mechanic...
Perhaps they could be next to the "Reply" button, and fully contained in the comment's container?
The answer is pretty clear with Bayes' Theorem. The world in which the coin lands heads and you get the card has probability 0.0000000005, and the world in which the coin lands tails has probability 0.5. Thus you live a world with a prior probability of 0.5000000005, so the probability of the coin being heads is 0.0000000005/0.5000000005, or a little under 1 in a billion.
Given that the worst case scenario of losing the bet is saying you can't pay it and losing credibility, you and Adam should take the bet. If you want to (or have to) actually commit to pay...
Personally I would be interested in a longer post about whatever you have to say about the battery and battery design. You could make a sequence, so that it can be split into multiple posts.
I assume work is output/time. If a machine is doing 100% of the work, then the human's output is undefined since the time is 0.
Some properties that I notice about semistable equilibria:
There's still the problem that two people can't occupy the same space at the same time, so we need people to be able to swap places instantly. This then requires some coordination, which is mentioned below.
Some commenters have mentioned economy of scale—It can be more efficient to pool together resources to make a bunch of one thing at a time. For example, people want paperclips but they could get them much faster if they operate a massive paperclip-making machine rather than everyone making their own individually. I think this is already covered though, a...
I wonder if there's a different potential takeaway here than "find what feels rewarding". Duhig’s story makes me think of a perspective I've learned from TEAM-CBT: Bad habits (and behavioural patterns in general) are there for a reason, as a solution to some other problem. An important first step to changing your behaviour is to understand the reasons why not to change, and then really consider what is worth changing. It sounds to me that Duhig figured out what problem eating cookies was trying to solve.
At least, that's the theory as I understand it. I hav...
I've learned to be resilient against AI distortions, but 'octagonal red stop sign' really got me. Which is ironic, you'd think that prompt would be particularly easy for the AI to handle. The other colours and shapes didn't have a strong effect, so I guess the level of familiarity makes a difference.
I think the level of nausea is a function of the amount of meaning that is being distorted, eg. distorted words, faces or food have a much stronger effect than warped clock faces or tables, for example. (I would also argue there is more meaning to the shape of a golf club than a clock face.)
I think the complement sandwich can be useful as a stepping stone to good communication. That said, I think of it as a narrow formulation of a more general (and less precisely defined) approach to conversation that I might point to with phrases like "work with people where they are at" and "be aware of the emotions that your words induce in other people". There was an article on LessWrong that I can't find, arguing that clear communication is worded to pre-emptively avoid likely misunderstandings and misconceptions. The idea I'm pointing to is like that, b... (read more)