All of dmav's Comments + Replies

dmav
68

Unfortunately, comparing the returns isn't a great way of evaluating the portfolio compared to the S&P 500. You should really be comparing their Sharpe ratios (or just annualized tstat). If you have, for example, 5% annualized returns on $x in excess of the risk-free rate, you can just double your pnl to 10% by borrowing $x more money and investing it (assuming you can borrow at a competitive rate). Why not do that? Well, you'll also have more variance in your portfolio. Probably what you really care about is risk-adjusted returns.

The most common way t... (read more)

dmav
10

This wouldn't really solve much of the problem though, since ETFs are still pretty expressive. For example, when they have a sense for whether an important clean-energy bill will pass or fail, they could buy/sell a clean-energy-tracking ETF.

Some ETFs are pretty high-weight Nvidia, so it would be pretty easy to still trade it indirectly, albeit a little bit less efficiently.

And honestly even the S&P500 will still move a lot based on various policy outcomes.

1AnnaJo
I agree that ETFs are pretty expressive and they can absolutely select into ETFs that directly correlate with their pet policies. But stock movements are often correlated across different sectors, so this would curb the most egregious carve-outs and concentrated benefits that politicians can confer on certain companies. Like an ETF made up of electric cars etc. would still generally correlate with economic growth. 
dmav
10

Just so you know, this is still missing on your personal site.
Also the image here doesn't exist on your personal site's post.
Thanks for writing all these wonderful resources Neel!

dmav
20

You probably also want to do some kind of normalization here based on how many total posts the user has upvoted. (So you can't just i.e. upvote everything.) (You probably actually care about something a little different from the accuracy of their upvoted-as-predictions on average though...)

dmav
80

Here's a good/accessible blog post that does a pretty good job discussing this topic. https://ericneyman.wordpress.com/2019/09/17/least-squares-regression-isnt-arbitrary/

dmav
72

I think that this is true of the original version of alphastar, but they have since trained a new version on camera inputs and with stronger limitations on apm (22 actions/5s) (Maybe you'd want some kind of noise applied to the inputs still, but I think the current state is much closer to human-like playing conditions.) See: https://www.deepmind.com/blog/alphastar-grandmaster-level-in-starcraft-ii-using-multi-agent-reinforcement-learning

2Lone Pine
Ah I didn't know they had upgraded it. I'm much more satisfied that SC2 is solved now.
dmav
20

In other words, we should be telling children 'be careful of roads/cars' (including on Halloween) Not 'be careful of Halloween'

I agree with the post, but I will point out that you really do need to emphasize the utility per micromort here. If you keep your utility constant, it is the total risk that matters. Just like if you were going to go on a long car ride tomorrow (on safer-than-usual roads, but not enough to outweigh the total driving) and someone points out you're much more likely to die than usual - sure, you can point out 'ah yes, but the chance I... (read more)

2jefftk
Yes, but even then not unduly so. Child pedestrian deaths per FARS are about 0.5 per 100k (320 per 72M), which puts them below guns (5 per 100k), driving (4 per 100k), drugs (3 per 100k), suffocation (2 per 100k), drowning (1 per 100k), etc: https://www.jefftk.com/goldstick-causes-of-child-deaths.jpg. Total deaths under age 18 are about 37k/y, so 320 pedestrian deaths is about 1% of deaths. Worth careful teaching, but still fine to let them walk to their friends' houses once they're ready. Not sure I understand what this paragraph is getting at? But it's also dark and chaotic.
dmav
41

As you kind of say - there are already (at least decently smart/competent) people trying to do (almost) all of these things. For many of these projects, joining current efforts is probably a better allocation than starting your own effort, and most of the value to be added is if you're in the 99.5th+ %-ile (?) for the 'skills needed.' (or sometimes there's just not enough people working on a problem, or sometimes there's a place to add value if you're willing to do annoying work other people don't want to do - these are both rarer though, in the current fu... (read more)

8Orpheus16
I disagree with part of this-- I think people often see 1 or 2 projects in space and assume the space is covered. I also think people are generally overrating the orgs that currently exist. In general, I expect that if there are 3-7 teams focused on a particular issue, 0-2 will be successful.  I think some orgs are worth joining and some people are better off joining existing orgs/projects. But I think it's very easy to see something that looks-like-its-solving-the-problem (and maybe its mission statement even says it's going to solve the problem) but forget that Solving The Problem Is Really Hard and Many Organizations Don't Live Up To Their Mission Statement. (I may write up a longer thing about this at some point with more details. But for now take this as "Akash's intuitions/observations.") Also I agree with the part about how much of the value is if you're 99+ %ile for skills needed (though I think people in that category are often deterred from doing things because they assume they're not in that category. I think people imagine that 99+ %ile means you're already some superhero with highly relevant experiences and tons of status. See also hero licensing). So on the margin I would want more people trying to test out if they are indeed 99+ %ile at a few different things.
dmav
31

Note that your prediction isn't interesting. Each year, conditioned on a doomsday not happening, it would be pretty weird for the date(s) to not have moved forward. 
Do you instead mean to say that you guess that the date will move forward each year by more than a year, or something like that?

1mukashi
The date can move forward because people might update for shorter timelines after seeing the improvements in AI
dmav
10

Here are some objections I have to your post:
How are you going to specify the amount of optimization pressure the AI exerts on answering a question/solving a problem? Are you hoping to start out training a weaker AI that you later augment? 
If so, I'd be concerned about any distributional shifts in its optimization process that occur during that transition
If not, it's not clear to me how you have the AI 'be safe' through this training process.

At the point where you, the human, is labeling data to train the AI to identify concepts with measurements/feat... (read more)

dmav
10

I think the question of you/Adele miscommunicating is mostly under-specification of what features you want your test-AGI to have.

  • If you throttle its ability to optimize for its goals, see EY and Adele's arguments.

  • If you don't throttle in this way, you run into goal-specification/constraint-specification issues, instrumental convergence concerns and everything that goes along with it.

I think most people here will strongly feel a (computationally) powerful AGI with any incentives is scary, and that any test-versions should require using at-most a mu... (read more)