Hi I'm helping organising the Stockholm LW meetup but I need more karma to be able to post, upboats plz.
Update on LW 2.0: user interviews scheduled for this week, work on the design underway, as well as some extra features. The broad plan is something like the following: user interviews / alpha testing to find the breaking UX bugs and get the design squared away, a closed beta to find more bugs and make sure the experience with multiple people doing stuff on the site is good / how we expect it to be, and then an open beta to give the broader community a chance to see it and find things for us to fix before it goes live at lesswrong.com. A core part of this process is making sure that there's consensus that it's actually worth switching.
Some random barely-edited thoughts on my experience with weight loss:
In the midst of a diet where I will lose 15 lbs (15.9lb, from 185.8 lb to 169.9, to be exact) in 40 days.
I have 95% certainty I will reach this goal in the appointed time. Even if I don't reach exactly 169.9lb, I'll be close, so whether or not I hit the exact number is arbitrary for my purposes. (I'm losing some weight to see if it helps a lingering back injury.)
I'm just eating a disciplined diet and working out according to a consistent schedule.
My diet is simple and not starvation-y at all. Most people wouldn't do it because it's repetitive (I literally eat the same thing nearly everyday so I can know my calorie intake without any counting.)
My workout isn't hard but most people wouldn't do it because...I don't know why, it's just my experience that people won't. It's 4-5 days per week of 30-60 minutes cardio and 30-60 minutes of weight training. I have a back injury that's limiting me, so it's nothing terribly rigorous.
...
In my years at health clubs, talking to health-club-going people, I've seen all the evidence I'll ever need to believe, basically, the Calories In / Calories Out model of weight loss is corre...
Why does patternism [the position that you are only a pattern in physics and any continuations of it are you/you'd sign up for cryonics/you'd step into Parfit's teleporter/you've read the QM sequence]
not imply
subjective immortality? [you will see people dying, other people will see you die, but you will never experience it yourself]
(contingent on the universe being big enough for lots of continuations of you to exist physically)
I asked this on the official IRC, but only feep was kind enough to oblige (and had a unique argument that I don't think everyone ...
There's a free market idea that the market rewards those who provide value to society. I think I've found a simple counterexample.
Imagine a loaf of bread is worth 1 dollar to consumers. If you make 100 loaves and sell them for 99 cents each, you've provided 1 dollar of value to society, but made 99 dollars for yourself. If you make 100 loaves and give them away to those who can't afford it, you've provided 100 dollars of value to society, but made zero for yourself. Since the relationship is inverted, we see that the market doesn't reward those who provide...
I often feel like upvotes on LW correspond more to the "insightfulness" of a post, rather than its perceived instrumental value. Unsure how I feel about this because if I'm relying on upvotes as a social incentive to write things, this shapes what I write in directions that might not be directly useful (IMO) to the most people.
Another week, another Open thread, another problem:
https://protokol2020.wordpress.com/2017/05/14/chesslike-problem/
Is there a good reason, that I am not seeing that there isn't a society for AGIrisk?
It would do various meta things around AGIrisk like
Outreach to AI students to inform them and measure the spread of ai safety ideas
Co-ordinate with the research institutes to provide experts for the media/government
Provide opsec advice for researchers to keep their dangerous results hidden.
Is there some nice game-theoretic solution that deals with the 'free rider problem', in the sense of making everyone pay in proportion to their honest valuation? Like how Vickery Auctions reveal honest prices, or Sperner's lemma can help with envy-free rent division?
I've watched Stuart Russel's TED talk on AI risk, and my gut reaction to it was "do you want to be paperclips? this is how you become paperclips!". It goes completely against the grain of the view that has been expressed on this blog as of few years ago. But, then again, AI is hard, and there might be some recent developments that I have missed. What is the current state of the research? What does EY and his camarilla think about the state of the problem as of now?
I'm searching for a quote. It goes something like this:
"In nearly every contest there comes a point where one competitor has decided that they are going to lose. Sometimes it's near the end; sometimes it's right at the start. After that point, everything they do will be aimed at bringing that result to pass."
And then continues in that vein for a bit. I don't have the wording close enough to correct for Google to get me what I'm looking for, though. And I could swear I've seen it quoted here before. Does someone else remember the source?
The last bigger Windows update left my computer without a driver for the GPU. Hardware acceleration on websites like https://human.biodigital.com/index.html didn't work. Unfortunately, it took a few months notice the specific problem. I installed the open source tool Snappy Driver Installer and it fixed the issue. It also installed proper drivers for other hardware.
Does anybody here know any personally successful techniques or strategies on handling regret? I have some regrets from my past that occasional come and bother me sometimes whenever I study.
I know the usual arguments why rewarding value creation is a good idea, and I'm not trying to argue against that. Instead, my first comment points out how markets don't always reward value creation. They do a more complicated and recursive thing. You can think of it as "if your product is awesome but your customers are poor, you're screwed". Only even worse because there's feedback effects, where value creators can become poor just because other value creators are poor etc.
To put it yet another way, in a PageRank-like system the utility will tend to clump together, leading to inequality and monopolies. That's more or less what the system does. The original PageRank had some nasty attractors after people started gaming it, which is why Google is tweaking it all the time. Someone on HN once made a cheeky comment saying "the end game of capitalism is one corporation selling everyone oxygen", and I can see how blindly trusting a recursive system to be fair can get you there.
Also, your St. Paul quote is a famous Soviet slogan, instantly familiar to anyone who was born in the USSR, so I chuckled a bit when you used it to defend free markets :-)
I disagree with your first comment about the $100 and the loaves, as I said, because you are overly simplifying. For example, even aside from the things I already mentioned, you also ignore the fact that the person needs to spend money or goods in order to produce the loaves.
That said, you might be able to refine that example or come up with another; I certainly do not think that markets infallibly have the result of rewarding value creation. I agree that free markets leads to that kind of inequality and that this is a not particularly great aspect of it....
If it's worth saying, but not worth its own post, then it goes here.
Notes for future OT posters:
1. Please add the 'open_thread' tag.
2. Check if there is an active Open Thread before posting a new one. (Immediately before; refresh the list-of-threads page before posting.)
3. Open Threads should start on Monday, and end on Sunday.
4. Unflag the two options "Notify me of new top level comments on this article" and "