All of RyanCarey's Comments + Replies

This covers pretty well the altruistic reasons for/against working on technical AI safety at a frontier lab. I think the main reason for working at a frontier lab, however, is not altruistic. It's that it offers more money and status than working elsewhere - so it would be nice to be clear-eyed about this.

To be clear, on balance, I think it's pretty reasonable to want to work at a frontier lab, even based on the altruistic considerations alone. 

What seems harder to justify altruistically, however, is why so many of us work on, and fund the same kinds ... (read more)

4Guive
It's important to be careful about the boundaries of "the same sort of safety work." For example, my understanding is that "Alignment faking in large language models" started as a Redwood Research project, and Anthropic only became involved later. Maybe Anthropic would have done similar work soon anyway if Redwood didn't start this project. But, then again, maybe not. By working on things that labs might be interested in you can potentially get them to prioritize things that are in scope for them in principle but which they might nevertheless neglect. 
3bilalchughtai
Agreed that this post presents the altruistic case. I discuss both the money and status points in the "career capital" paragraph (though perhaps should have factored them out).

I don't mean this as a criticism - you can both be right - but this is extremely correlated to the updates made by the average Bay Area x-risk reduction-enjoyer over the past 5-10 years, to the extent that it almost could serve as a summary.

5LawrenceC
That seems correct, at least directionally, yes.

It may be useful to know that if events all obey the Markov property (they are probability distributions, conditional on some set of causal parents), then the Reichenbach Common Cause Principle follows (by d-separation arguments) as a theorem. So any counterexamples to RCCP must violate the Markov property as well.

There's also a lot of interesting discussion here.

The idea that "Agents are systems that would adapt their policy if their actions influenced the world in a different way." works well on mechanised CIDs whose variables are neatly divided into object-level and mechanism nodes: we simply check for a path from a utility function F_U to a policy Pi_D. But to apply this to a physical system, we would need a way to obtain such a partition those variables. Specifically, we need to know (1) what counts as a policy, and (2) whether any of its antecedents count as representations of "influence" on the world (and af... (read more)

3tom4everitt
Agree, the formalism relies on a division of variable. One thing that I think we should perhaps have highlighted much more is Appendix B in the paper, which shows how you get a natural partition of the variables from just knowing the object-level variables of a repeated game. A spinal reflex would be different if humans had evolved in a different world. So it reflects an agentic decision by evolution. In this sense, it is similar to the thermostat, which inherits its agency from the humans that designed it. Same as above. One thing that I'm excited about to think further about is what we might call "proper agents", that are agentic in themselves, rather than just inheriting their agency from the evolution / design / training process that made them. I think this is what you're pointing at with the ant's knowledge. Likely it wouldn't quite be a proper agent (but a human would, as we are able to adapt without re-evolving in a new environment). I have some half-developed thoughts on this.

Nice. I've previously argued similarly that if going for tenure, AIS researchers might places that are strong in departments other than their own, for inter-departmental collaboration. This would have similar implications to your thinking about recruiting students from other departments. But I also suggested we should favour capital cities, for policy input, and EA hubs, to enable external collaboration. But tenure may be somewhat less attractive for AIS academics, compared to usual, in that given our abundant funding, we might have reason to favour Top-5 ... (read more)

Feature suggestion. Using highlighting for higher-res up/downvotes and (dis)agreevotes.

Sometimes you want to indicate what part of a comment you like or dislike, but can't be bothered writing a comment response. In such cases, it would be nice if you could highlight the portion of text that you like/dislike, and for LW to "remember" that highlighting and show it to other users. Concretely, when you click the like/dislike button, the website would remember what text you had highlighted within that comment. Then, if anyone ever wants to see that highlighting... (read more)

The title suggests (weakly perhaps) that the estimates themselves peer-reviewed. Would be clearer to write "building on" peer reviewed argument, or similar.

3michaelcohen
Thank you. I've changed the title.

Hi Orellanin,

In the early stages, I had in mind that the more info any individual anon-account revealed, the more easily one could infer what time they spent at Leverage, and therefore their identity. So while I don't know for certain, I would guess that I created anonymoose to disperse this info across two accounts.

When I commented on the Basic Facts post as anonymoose, It was not my intent to contrive a fake conversation between two entities with separate voices. I think this is pretty clear from anonymoose's comment, too - it's in the same bulleted and ... (read more)

"A Russian nuclear strike would change the course of the conflict and almost certainly provoke a "physical response" from Ukraine's allies and potentially from the North Atlantic Treaty Organization, a senior NATO official said on Wednesday.

Any use of nuclear weapons by Moscow would have "unprecedented consequences" for Russia, the official said on the eve of a closed-door meeting of NATO's nuclear planning group on Thursday.

Speaking on condition of anonymity, he said a nuclear strike by Moscow would "almost certainly be drawing a physical response from ma... (read more)

The reasoning is that retaliating is US doctrine - they generally respond to hostile actions in-kind, to deter them. If Ukraine got nuked, the level of outrage would place intense pressure on Biden to do something, and the hawks would become a lot louder than the doves, similar to after the 9/11 attacks. In the case of Russia, the US has exhausted most non-military avenues already. And US is a very militaristic country - they have many times bombed countries (Syria, Iraq, Afghanistan, Libya) for much less. So military action just seems very likely. (Involv... (read more)

2RyanCarey
"A Russian nuclear strike would change the course of the conflict and almost certainly provoke a "physical response" from Ukraine's allies and potentially from the North Atlantic Treaty Organization, a senior NATO official said on Wednesday. Any use of nuclear weapons by Moscow would have "unprecedented consequences" for Russia, the official said on the eve of a closed-door meeting of NATO's nuclear planning group on Thursday. Speaking on condition of anonymity, he said a nuclear strike by Moscow would "almost certainly be drawing a physical response from many allies, and potentially from NATO itself". "-Reuters https://news.yahoo.com/russian-nuclear-strike-almost-certainly-144246235.html" I have heard of talk that the US might instead arm Ukraine with tactical nukes of its own, although I think that would be at least comparably risky as military retaliation.

I think your middle number is clearly too low. The risk scenario does not require that NATO trigger article 5 necessarily, but just that they carry out a strategically significant military response, like eliminating Russia's Black Sea Fleet, nuking, or creating a no-fly zone. And Max's 80% makes more sense than your 50% for he union of these possibilities, because it is hard to imagine that the US would stand down without penalising the use of nukes.

I would be at maybe .2*.8*.15=.024 for this particular chain of events leading to major US-Russia nuclear war.

4Ege Erdil
I don't think it's hard to imagine, I can imagine it quite easily. 80% just seems overconfident to me on this question. NATO has no actual obligation to respond to any nuclear use in Ukraine, and I don't see why you're so confident that NATO would respond to Russian use of e.g. tactical nukes in Ukraine by attacking Russia directly. It's not that I think this is unlikely, but in my opinion 80% is just too high of a confidence in what NATO would do in such an unprecedented situation. That said, this is the part of Tegmark's forecast that I disagree with the least, because the difference between 50% and 80% is quite small for the purposes of this calculation. I think it's much more important for him to justify his 30% and 70%, and I assume you would agree with me about that.

All of these seem to be good points, although I haven't given up on liquidity subsidy schemes yet.

Some reports are not publicised in order not to speed up timelines. And ELK is a bit rambly - I wonder if it will get subsumed by much better content within 2yr. But I do largely agree.

It would be useful to have a more descriptive title, like "Chinchilla's implications for data bottlenecks" or something.

It's noteworthy that the safety guarantee relies on the "hidden cost" (:= proxy_utility - actual_utility)  of each action being bounded above. If it's unbounded, then the theoretical guarantee disappears. 

1Adam Jermyn
Good point! And indeed I am skeptical that there are useful bounds on the cost...

For past work on causal conceptions of corrigibility, you should check out this by Jessica Taylor. Quite similar.

It seems like you're saying that the practical weakness of forecasters vs experts is their inability to make numerous causal forecasts. Personally, I think the causal issue is the main issue, whereas you think it is that the predictions are so numerous. But they are not always numerous - sometimes you can affect big changes by intervening at a few pivot points, such as at elections. And the idea that you can avoid dealing with causal interventions by conditioning on every parent is usually not practical, because conditioning on every parent/confounder mean... (read more)

You could test this to some extent by asking the forecasters to predict more complicated causal questions. If they lose most of their edge, then you may be right.

3Ege Erdil
Speaking as a forecaster, my first instinct when someone does that is to go on Google and look up whatever information may be relevant to the question that I've been asked. If I'm not allowed to do that and the subject is not one I'm expert in then I anticipate doing quite badly, but if I'm allowed a couple hours of research and the question is grounded enough I think most of the edge the expert has over me would disappear. The problem is that an expert is really someone who has done that for their whole field for thousands of hours, but often if your only goal is to answer a very specific question you can outsource most of the work away (e.g. download Sage and run some code on it to figure out the group of rational points of an elliptic curve). So I don't think what you said really tests my claim, since my argument for why understanding is important when you want to have impact is that in the real world it's prohibitively expensive to set up conditional prediction markets on all the different actions you could take at a given moment every time you need to make a decision. As davidad puts it, "it's mostly that conditional prediction markets are too atomized/uncompressed/incoherent". I still think your experiment would be interesting to run but depending on how it were set up I can imagine not updating on it much even if forecasters do about as well as experts.

I don't think the capital being locked up is such a big issue. You can just invest everyone's money in bonds, and then pay the winner their normal return multiplied by the return of the bonds.

A bigger issue is that you seem to only be describing conditional prediction markets, rather than ones that truly estimate causal quantities, like P(outcome|do(event)). To see this, note that the economy will go down IF Biden is elected, whereas it is not decreased much by causing Biden to be elected. The issue is that economic performance causes Biden to be unpopular... (read more)

I would do thumbs up/down for good/bad, and tick/cross for correct/incorrect.

What do you want to spend most of your time on? What do you think would be the most useful things to spend most of your time on (from a longtermist standpoint)?

2lsusr
I think the most important thing for me to spend my time on is AI Alignment. I have PM'd you a link to a proposal I could use funding on.

You say two things that seem in conflict with one another.

[Excerpt 1] If a system is well-described by a causal diagram, then it satisfies a complex set of statistical relationships. For example ... To an evidential decision theorist, these kinds of statistical relationships are the whole story about causality, or at least about its relevance to decisions. 
[Excerpt 2] [Suppose] that there is a complicated causal diagram containing X and Y, such that my beliefs satisfy all of the statistical relationships implied by that causal diagram. EDT recommends

... (read more)
2tailcalled
"The parents of X" is stuff like the observations that agent has made, as well as the policy the agent uses. It is bog-standard for EDT to use this in its decisions, and because of the special nature of those variables, it does not require knowing an entire causal model.
1Ege Erdil
The EDT agent needs to know the inputs to its own decision process to even make decisions at all, so I don't think there's a causal implication there. Obviously no decision theory can get off the ground if it's not permitted to have any inputs. It's just that in a causal model the inputs to a decision process would have to be causal arrows going from the inputs to the decision-maker. If by "coinciding for decisions that are in the support" you mean what I think that means, then that's true re: actions that never happen, but it's not clear why actions that never happen should influence your assessment of how a decision theory works. Implicitly when you do anything probabilistic you assume that sets of null measure can be thrown away without changing anything.

Many people don't realize how effective migraine treatments are. High-dose aspirin, tryptans, and preventers all work really well, and can often reduce migraine severity by 50-90%.

Also, most don't yet realise how effective semaglutide is for weight loss, due to the fact that weight loss drugs have generally been much less effective, or had much worse side-effects previously.

Balding treatments (finasteride and topical minoxodil) are also pretty good for a lot of people.

3moridinamael
Many people don't even realize that they have migraines, and treat their recurring headaches with NSAIDs or acetaminophen, instead of the vastly more effective triptans. And as you say, few are aware of the new and miraculous CGRP inhibitor class of migraine preventative drugs.

Another possibility is that most people were reluctant to read, summarise, or internalise Putin's writing on Ukraine due to finding it repugnant, because they aren't decouplers.

Off the top of my head, maybe it's because Metaculus' presents medians, and the median user neither investigates the issue much, nor trusts those who do (Matt Y, Scott A) and just roughly follows base rates. I also feel there was some wishful thinking, and that to some extent, the fullness of the invasion was at least somewhat intrinsically surprising.

5RyanCarey
Another possibility is that most people were reluctant to read, summarise, or internalise Putin's writing on Ukraine due to finding it repugnant, because they aren't decouplers.

Nice idea. But if you set C at like 10% of the correct price, then you're going to sell 90% of the visas on the first day for way too cheap, so you can lose almost all of the market surplus.

3Measure
My thought was that you'd set C at your best guess clearing price or maybe a bit higher. You could instead go with 10x the clearing price and plan to not sell many before the last month but maybe get a bit more revenue overall.

Yeah I think in practice auctioning every day or two would be completely adequate - that's much less than the latency involved in dealing with lawyers and other aspects of the process. So now I'm mostly just curious about whether there's a theory built up for these kinds of problems in the continuous time case.

Yes. And, the transformer-based WordTune is complementary - better for copyediting/rephrasing, rather than narrow grammatical correctness.

We do not have a scientific understanding of how to tell a superintelligent machine to "solve problem X, without doing something horrible as a side effect", because we cannot describe mathematically what "something horrible" actually means to us...

Similar to how utility theory (from von Neumann and so on) is excellent science/mathematics despite our not being able to state what utility is. AI Alignment hopes to tell us how to align AI, not the target to aim for. Choosing the target is also a necessary task, but it's not the focus of the field.

2Gunnar_Zarncke
It is not a quote but a paraphrasing of what the OP might agree on about AI security.

In terms of trying to formulate rigorous and consistent definitions, a major goal of the Causal Incentives Working Group is to analyse features of different problems using consistent definitions and a shared framework. In particular, our paper "Path-specific Objectives for Safer Agent Incentives" (AAAI-2022) will go online in about month, and should serve to organize a handful of papers in AIS.

1mocny-chlapik
Thanks, this looks very good.

Exactly. Really, the title should be "Six specializations makes you world-class at a combination of skills that is probably completely useless." Really, productivity is a function of your skills. The fact that you are "world class" in a random combination of skills is only interesting if people are systematically under-estimating the degree to which random skills can be usefully combined. If there are reasons to believe that, then I would be interested in reading about it.

Transformer models (like GPT-3) are generators of human-like text, so they can be modeled as quantilizers. However, any quantiliser guarantees are very weak, because they quantilise with very low q, equal to the likelihood that a human would generate that prompt.

The most plausible way out seems to be for grantmakers to grant money conditionally on work being published as open source. Some grantmakers may benefit from doing this, despite losing some publication prestige, because the funded work will be read more widely, and the grantmaker will look like they are improving the scientific process. Researchers lose some prestige, but gain some funding. Not sure how well this has worked so far, but perhaps we could get to the world where this works, if we're not already there.

Elsevier journals allow individual authors to make their own published paper "immediately and permanently free for everyone to read and download" for a fee. In the last Elsevier journal I submitted a paper to, the fee was $2,400.

I think this means that a grant conditioned on open-access publishing would just mean that authors will have to pay the fee if they publish in an Elsevier journal -- this makes it more like a tax (paid out of grant money) than a ban. Not sure if that would make it more or less effective, on net, though.

It would be useful to have a clarification of these points, to know how different of an org you actually encountered, compared to the one I did when I (briefly) visited in 2014.

It is not true that people were expected to undergo training by their manager.

OK, but did you have any assurance that the information from charting was kept confidential from other Leveragers? I got the impression Geoff charted people who he raised money from, for example, so it at least raises the question whether information gleaned from debugging might be discussed with that pers... (read more)

Thanks for your courage, Zoe!

Personally, I've tried to maintain anonymity in online discussion of this topic for years. I dipped my toe into openly commenting last week, and immediately received an email that made it more difficult to maintain anonymity - I was told "Geoff has previously speculated to me that you are 'throwaway', the author of the 2018 basic facts post". Firstly, I very much don't appreciate my ability to maintain anonymity being narrowed like this. Rather, anonymity is a helpful defense in any sensitive online discussion, not least this o... (read more)

8orellanin
Hi! In the past few months I've been participating in Leverage Research/EA discourse on Twitter. Now there is one Twitter thread discussing your involvement as throwaway/anonymoose: https://twitter.com/KerryLVaughan/status/1585319237018681344 (with a subthread starting at https://twitter.com/ohabryka/status/1586084766020820992 discussing anti-doxxing norms and linking back to EA Forum comments). One piece of information that's missing is why you used two throwaway accounts instead of one (and in particular, why you used one to reply to the other one, as alleged by Kerry Vaughan in https://twitter.com/KerryLVaughan/status/1585319243985424384 ). Can you tell me about your reasoning behind that decision? (If that matters, I am not affiliated with any Leverage-adjacent org and I am not a throwaway account for a different EA Forum user.)
4Evan_Gaensbauer
Leverage Research hosted a virtual open house and AMA a couple weeks ago for their relaunch as a new kind of organization that has been percolating for the last couple years. I attended. One subject Geoff and I talked about was the debacle that was the article in The New York Times (NYT) on Scott Alexander from several months ago. I expressed my opinion that: 1.  Scott Alexander could have managed his online presence much better than he did on and off for a number of years. 2. Scott Alexander and the rationality community in general could have handled the situation much better than they did. 3. Those are parts of this whole affair that too few in the rationality community have been willing to face, acknowledge or discuss about what can be learned from mistakes made. 4. Nonetheless, NYT was the instigating party in whatever of the situation constituted a conflict between NYT, and Scott Alexander and his supporters, and NYT is the party that should be held more accountable and is more blameworthy if anyone wants to make it about blame. Geoff nodded, mostly in agreement, and shared his own perspective on the matter that I won't share. Yet if Geoff considers NYT to have done one or more things wrong in that case,  You yourself, Ryan, never made any mistake of posting your comments online in a way that might make it easier for someone else to de-anonymize you. If you made any mistake, it's that you didn't anticipate how adeptly Geoff would apparently infer or discern your identity. I expect why it wouldn't be so hard for Geoff to have figured it out it was you because you would have shared information about the internal activities at Leverage Research you are one of only a small number of people would have had access to.  Yet that's not something you should not have had to anticipate. A presumption of good faith in a community or organization entails a common assumption that nobody would do that to their other peers. Whatever Geoff himself has been thinking about

What's frustrating about still hearing noisy debate on this topic, so many years later, is that Leverage being a really bad org seems overdetermined at this point. On the one hand, if I ranked MIRI, CFAR, CEA, FHI, and several startups I've visited, in terms of how reality-distorting they can be, Leverage would score ~9, while no other would surpass ~7. (It manages to be nontransparent and cultlike in other ways too!).  While on the other hand, their productive output was... also like a 2/10? It's indefensible. But still only a fraction of the relevan

... (read more)

As in, 5+ years ago, around when I'd first visited the Bay, I remember meeting up 1:1 with Geoff in a cafe. One of the things I asked, in order to understand how he thought about EA strategy, was what he would do if he wasn't busy starting Leverage. He said he'd probably start a cult, and I don't remember any indication that he was joking whatsoever. I'd initially drafted my comment as "he told me, unjokingly", except that it's a long time ago, so I don't want to give the impression that I'm quite that certain.

He's also told me, deadpan, that he would like to be starting a cult if he wasn't running Leverage.

6matt
I've read this comment several times, and it seems open to interpretation whether RyanCarey is mocking orthonormal for presenting weak evidence by presenting further obviously weak evidence, or whether RyanCarey is presenting weak evidence believing it to be strong. Just to lean on the scales a little here, towards readers taking from these two comments (Ryan's and orthonormal's) what I think could (should?) be taken from them… An available interpretation of orthonormal's comment is that orthonormal: 1. had a first impression of Geoff that was negative, 2. then backed that first impression so hard that they "[hurt their] previously good friendships with two increasingly-Leverage-enmeshed people" (which seems to imply: backed that first impression against the contrary opinions of two friends who were in a position to gather increasingly overwhelmingly more information by being in a position to closely observe Geoff and his practices), 3. while telling people of their first impression "for the entire time since" (for which, absent other information about orthonormal, it is an available interpretation that orthonormal engaged in what could be inferred to be hostile gossip based on very little information and in the face of an increasing amount of evidence (from their two friends) that their first impression was false (assuming that orthonormal's friends were themselves reasonable people)). 4. (In this later comment) orthonormal then reports interacting with Geoff "a few times since 2012" (and reports specific memory of one conversation, I infer with someone other than Geoff, about orthonormal’s distrust of Leverage) (for which it is an available interpretation that orthonormal gathered much less information than their "Leverage-enmeshed" friends would have gathered over the same period, stuck to their first impression, and continued to engage in hostile gossip). Those who know orthonormal may know that this interpretation is unreasonable given their knowledge o

Your comparison does a disservice to the human's sample efficiency in two ways: 

  1. You're counting diverse data in the human's environment, but you're not comparing their performance on diverse tasks. Human's are obviously better than GPT3 at interactive tasks, walking around, etc. For either kind of fair comparison text data & task, or diverse data & task, the human has far superior sample efficiency.
  2. "fancy learning techniques" don't count as data. If the human can get mileage out of them, all the better for the human's sample efficiency.

So you seem to have it backwards when you say that the comparison that everyone is making is the "bad" one.

1Daniel Kokotajlo
Thanks. Hmmm. I agree with #2, and should edit to clarify. I meant "fancy learning techniques that we could also do with our AIs if we wanted," but maybe I'll just avoid that can of worms for now. For #1: We don't know how well a human-sized artificial neural net would perform if it was trained on the quantity and variety of data that humans have. We haven't done the experiment yet. However, my point is that for all we know it's entirely possible that such a neural net would perform at about human level on all the tasks humans do. The people who are saying that modern neural nets are significantly less sample-efficient than humans are committed to denying this. (Or if they aren't, then I don't know what we are arguing about anymore?) They are committed to saying that we can extrapolate from e.g. GPT-3's performance vs. training data to conclude that we'd need something trained a lot longer than a human (on similar-to-human-lifetime data) to reach human performance. One way they might run this argument is to point out that GPT-3 has already seen more text than any human ever. My reply is that if a human had seen as much text as GPT-3, and only text, nothing else they probably would have poor performance as well, certainly on every task that wasn't a text-based task! Sorry for this oblique response to your point, if it is insufficient I can make a more direct one.

I think this becomes a lot clearer if we distinguish between total and marginal thinking. GPT-3's total sample efficiency for predicting text is poor:

  • To learn to predict text, GPT-3 has to read >1000x as much text as a human can learn in their lifetime.
  • To learn to win at go, AlphaGo has to play >100x times as many games as a human could play in their lifetime.

But on-the-margin, it's very sample efficient at learning to perform new text-related tasks:

  • GPT-3 can learn to perform a new text-related task as easily as a human can.

Essentially, what's happen... (read more)

That does help, thanks. However, now that I understand better what people are saying, I think it's wrong:

The comparison they are making is as follows:

GPT-3Human
Pre-trained on 3x10^11 tokens of textPre-trained on 3x10^8 tokens of text (fermi estimate based on WMP 300 so maybe 500 tokens per minute, 10 hours per week reading, 52 weeks a year, over 20 years of life)
Able to read a new fact once or twice and then learn it / remember it.Able to read a new fact once or twice and then learn it / remember it

However, I think this is a bad comparison, because it igno... (read more)

Can you clarify whether you're talking about "30% of X" i.e. 0.3*X, or "30% off X", i.e. 0.7*X?

Thanks. These algorithms seem like they would be better for passing the independence of clone alternatives criterion.

I imagine you could catch useful work with i) models of AI safety, or ii) analysis of failure modes, or something, though I'm obviously biased here.

The implication seems to be that this RFP is for AIS work that is especially focused on DL systems. Is there likely to be a future RFP for AIS research that applies equally well to DL and non-DL systems? Regardless of where my research lands, I imagine a lot of useful and underfunded research fits in the latter category.

3abergal
This RFP is an experiment for us, and we don't yet know if we'll be doing more of them in the future. I think we'd be open to including research directions we think that are promising that apply equally well to both DL and non-DL systems-- I'd be interested in hearing any particular suggestions you have. (We'd also be happy to fund particular proposals in the research directions we've already listed that apply to both DL and non-DL systems, though we will be evaluating them on how well they address the DL-focused challenges we've presented.)

To me, his main plausible x important claim is that performance is greatly improved by subject specialisation from age <5. The fact that many geniuses enter their fields late doesn't falsify this, since that isn't humdrum at all - barely one in a million kids specialise in that way. I think that people who enter a field such as CS at age 30, rather than at age 20 do have a mild disadvantage, maybe 0.5SD. So I wouldn't be surprised if starting at age 4, rather than at age 20 gave you another 1-2SD advantage. Of course, subjects like CS do benefit a lot f... (read more)

4gwern
I don't think that's true either, though. Early specialization requires solving an almost impossible prediction problem (it's difficult enough to know what would be the 'right' field for a teenager or young adult, how are you going to do it for a <5yo? This is the same reason that high-IQ elementary schools can't work); people, nevertheless, continue to try to do with Polgar says, and yet, we don't see kids trained from toddlerhood dominating the elite reaches of every field. Early training is of pretty dubious value: when we look at early childhood interventions like Headstart, the gains fade out, and there are plenty of places like, I believe, Finland, which start education late and see no problem from this. (I think Scott also discussed this for homeschooling and in his graduation post.) "T-shaped" expertise requires a lot of exploration to gain breadth and figure out where to specialize, and for every Polgar, there's a late bloomer (iirc, Epstein in The Sports Gene - which I liked far more than Bring Up Genius - gives many athletic examples, and made it a major focus of his 2019 Range: Why generalists triumph in a specialized world which I haven't read yet); and you have newer results like "What Makes a Champion? Early Multidisciplinary Practice, Not Early Specialization, Predicts World-Class Performance", Gullich et al 2021, which find the opposite of this claim: (I favor their "multiple-sampling-and-functional-matching hypothesis": when I read biographies, the importance of 'fitting' in a specific field that one can be obsessive about and which matches one's unique profile, seems like a critical and often underrated factor in going from being a highly talented and competent researcher, to a researcher someone would be reading or writing a bio about.)
2Viliam
There is a tension between specialization and keeping your options open. But also, many people are far from the Pareto boundary. For most kids, the alternative to learning chess (or CS or whatever) is watching cartoons, not studying something else. Specialization has the risk that the entire field may become irrelevant, so it doesn't matter if you are world's number 1. Or you may find out one day that you hate the job, but you can't do anything else. Polgár made a successful bet. All three of his girls made it to the top. But without the benefit of hindsight, I would say that his strategy was quite risky: suppose the girls instead become "merely" 10th best in their country; now what? -- Perhaps the idea was that they could later switch from chess to something else, and become good-but-not-genius at the other thing. (Also, the girls were born during socialism, where full employment was the official goal, so not being able to find a job was not an actual risk. Arguably, in socialism it made a lot of sense to become a specialist on something that brings prestige, because the regime would redistribute the money.)

Have you considered just doing some BTC/BTC-PERP arbitrage, or betting on politics and sports? You'd probably learn what skills they're looking for, gain some of them, and make money while you're at it...

1KaragounisZ
I have no idea even how to begin doing those things. But they do sound like good things to try to figure out. I am hoping that having a mentor would speed up this process considerably.

Thanks for these thoughts about the causal agenda. I basically agree with you on the facts, though I have a more favourable interpretation of how they bear on the potential of the causal incentives agenda. I've paraphrased the three bullet points, and responded in reverse order:

3) Many important incentives are not captured by the approach - e.g. sometimes an agent has an incentive to influence a variable, even if that variable does not cause reward attainment. 

-> Agreed. We're starting to study "side-effect incentives" (improved name pending), whic... (read more)

I think moving to the country could possibly be justified despite harms to recruitment and the rationality community, but in the official MIRI explanations, the downsides are quite underdiscussed.

Load More