Hi I'm helping organising the Stockholm LW meetup but I need more karma to be able to post, upboats plz.
Update on LW 2.0: user interviews scheduled for this week, work on the design underway, as well as some extra features. The broad plan is something like the following: user interviews / alpha testing to find the breaking UX bugs and get the design squared away, a closed beta to find more bugs and make sure the experience with multiple people doing stuff on the site is good / how we expect it to be, and then an open beta to give the broader community a chance to see it and find things for us to fix before it goes live at lesswrong.com. A core part of this process is making sure that there's consensus that it's actually worth switching.
Some random barely-edited thoughts on my experience with weight loss:
In the midst of a diet where I will lose 15 lbs (15.9lb, from 185.8 lb to 169.9, to be exact) in 40 days.
I have 95% certainty I will reach this goal in the appointed time. Even if I don't reach exactly 169.9lb, I'll be close, so whether or not I hit the exact number is arbitrary for my purposes. (I'm losing some weight to see if it helps a lingering back injury.)
I'm just eating a disciplined diet and working out according to a consistent schedule.
My diet is simple and not starvation-y at all. Most people wouldn't do it because it's repetitive (I literally eat the same thing nearly everyday so I can know my calorie intake without any counting.)
My workout isn't hard but most people wouldn't do it because...I don't know why, it's just my experience that people won't. It's 4-5 days per week of 30-60 minutes cardio and 30-60 minutes of weight training. I have a back injury that's limiting me, so it's nothing terribly rigorous.
...
In my years at health clubs, talking to health-club-going people, I've seen all the evidence I'll ever need to believe, basically, the Calories In / Calories Out model of weight loss is corre...
Why does patternism [the position that you are only a pattern in physics and any continuations of it are you/you'd sign up for cryonics/you'd step into Parfit's teleporter/you've read the QM sequence]
not imply
subjective immortality? [you will see people dying, other people will see you die, but you will never experience it yourself]
(contingent on the universe being big enough for lots of continuations of you to exist physically)
I asked this on the official IRC, but only feep was kind enough to oblige (and had a unique argument that I don't think everyone ...
There's a free market idea that the market rewards those who provide value to society. I think I've found a simple counterexample.
Imagine a loaf of bread is worth 1 dollar to consumers. If you make 100 loaves and sell them for 99 cents each, you've provided 1 dollar of value to society, but made 99 dollars for yourself. If you make 100 loaves and give them away to those who can't afford it, you've provided 100 dollars of value to society, but made zero for yourself. Since the relationship is inverted, we see that the market doesn't reward those who provide...
I often feel like upvotes on LW correspond more to the "insightfulness" of a post, rather than its perceived instrumental value. Unsure how I feel about this because if I'm relying on upvotes as a social incentive to write things, this shapes what I write in directions that might not be directly useful (IMO) to the most people.
Another week, another Open thread, another problem:
https://protokol2020.wordpress.com/2017/05/14/chesslike-problem/
Is there a good reason, that I am not seeing that there isn't a society for AGIrisk?
It would do various meta things around AGIrisk like
Outreach to AI students to inform them and measure the spread of ai safety ideas
Co-ordinate with the research institutes to provide experts for the media/government
Provide opsec advice for researchers to keep their dangerous results hidden.
Is there some nice game-theoretic solution that deals with the 'free rider problem', in the sense of making everyone pay in proportion to their honest valuation? Like how Vickery Auctions reveal honest prices, or Sperner's lemma can help with envy-free rent division?
I've watched Stuart Russel's TED talk on AI risk, and my gut reaction to it was "do you want to be paperclips? this is how you become paperclips!". It goes completely against the grain of the view that has been expressed on this blog as of few years ago. But, then again, AI is hard, and there might be some recent developments that I have missed. What is the current state of the research? What does EY and his camarilla think about the state of the problem as of now?
I'm searching for a quote. It goes something like this:
"In nearly every contest there comes a point where one competitor has decided that they are going to lose. Sometimes it's near the end; sometimes it's right at the start. After that point, everything they do will be aimed at bringing that result to pass."
And then continues in that vein for a bit. I don't have the wording close enough to correct for Google to get me what I'm looking for, though. And I could swear I've seen it quoted here before. Does someone else remember the source?
The last bigger Windows update left my computer without a driver for the GPU. Hardware acceleration on websites like https://human.biodigital.com/index.html didn't work. Unfortunately, it took a few months notice the specific problem. I installed the open source tool Snappy Driver Installer and it fixed the issue. It also installed proper drivers for other hardware.
Does anybody here know any personally successful techniques or strategies on handling regret? I have some regrets from my past that occasional come and bother me sometimes whenever I study.
I think it does imply subjective immortality. I'll bite that bullet. Therefore, you should sign up for cryonics.
Consciousness isn't continuous. There can be interruptions, like falling asleep or undergoing anesthesia. A successor mind/pattern is a conscious pattern that remembers being you. In the multiverse, any given mind has many many successors. It doesn't have to follow immediately, or even have to follow at all, temporally. At the separations implied even for a Tegmark Level I multiverse, past and future are meaningless distinctions, since there can be no interactions.
You are your mind/pattern, not your body. A mind/pattern is independent of substrate. Your unconscious, sleeping self is not your successor mind/pattern. It's an unconscious object that has a high probability of creating your successor (i.e. it can wake up). Same with your cryonicically-preserved corpsicle, though the probability is lower.
Any near-death event will cause grievous suffering to any barely-surviving successors, and grief and loss to friends and relatives in branches where you (objectively) don't survive. I don't want to suffer grievous injury, because that would hurt. I also don't want my friends and relatives to suffer my loss. Thus, I'm reluctant to risk anything that may cause objective death.
But, the universe being a dangerous place, I can't make that risk zero. By signing up for cryonics, I can increase the measure of successors that have a good life, even after barely surviving.
In the Multiverse, death isn't all-or-none, black or white. A successor is a mind that remembers being you. It does not have to remember everything. If you take a drug that causes you to not form long-term memory of any event today, have you died by the next day? Objectively, no. Your friends and relatives can still talk to "you" the next day. Subjectively, partially. Your successors lack certain memories. But people forget things all the time.
Being mortal in the multiverse, you can expect that your measure of successors will continue to diminish as your branches die, but the measure never reaches absolute zero. Eventually all that remains are Bolzman Brains and the like. The most probable Boltzman brain successors only live long enough to have a "single" conscious qualia of remembering being you. The briefest of conscious thoughts. Their successors remember that thought and may have another random thought. You can eventually expect an eternity of totally random qualia and no control at all over your experience.
This isn't Hell, but Limbo. Suffering is probably only a small corner of possible qualia-space, but so is eudaimonia. After an eternity you might stumble onto a small Botzlman World where you have some measure of control over your utility for some brief time, but that world will die, and your successors will again be only Boltzman brains.
I can't help that some of my successors from any given moment are Boltzman brains. But I don't want my only successors to be Boltzman Brains, because they don't increase my utility. Therefore, cryonics.
See the Measure Problem of cosmology. I'm not certain of my answer, and I'd prefer not to bet my life on it, but it seems more likely than not. I do not believe that Boltzman Brains can be eliminated from cosmology, only that they have lesser measure than evolved beings like us. This is because of the Trivial Theorem of Arithmetic: almost all natural numbers are really damn huge. The universe doesn't have to be infinite to get a Tegmark Level I multiverse. It just has to be sufficiently large.
Are people close to you aware that this is a reason that you advocate cryonics?
If it's worth saying, but not worth its own post, then it goes here.
Notes for future OT posters:
1. Please add the 'open_thread' tag.
2. Check if there is an active Open Thread before posting a new one. (Immediately before; refresh the list-of-threads page before posting.)
3. Open Threads should start on Monday, and end on Sunday.
4. Unflag the two options "Notify me of new top level comments on this article" and "