All of benkuhn's Comments + Replies

Oops—I forgot that this was going to be auto-crossposted to LW and probably would have prevented that if I remembered, since it's weirdly meta (it was intended mostly as a placeholder for people visiting my website and wondering why there were no recent posts). I guess I'll leave it here now since it got a lot more upvotes than I expected though :)

4Ruby
Nah, it's good! I'm glad I got to read this. Anything that's not "core LW content" we mark as Personal blogpost (poorly named, it just means stuff we don't want to promote front and center to readers either because it's off-topic or its a topic we want to deemphasize like politics). Zvi and Jeff Kaufman's blogs get auto-crossposted and it's nice getting all their zany stuff too.

I think it's still important to note that "not giving up" can lead not just to lack of success, but also to value destruction (Pets.com; Theranos; WeWork). 

If you're going to interpret the original "don't give up" advice so literally and blindly that "no matter what the challenges are I'm going to figure them out" includes committing massive fraud, then yes, it will be bad advice for you. That's a really remarkably uncharitable interpretation.

Not sure if this is your typo or a LW bug, but "essay" appears not to actually be hyperlinked?

1ThomasJ
I think I mis-pasted the link. I have edited it, but it's suppose to go to https://www.aqr.com/Insights/Perspectives/A-Gut-Punch
benkuhn110

I don't think founder/investor class conflict makes that much sense as an explanation for that. It's easy to imagine a world in which investors wanted their money returned when the team updates downwards on their likelihood of success. (In fact, that sometimes happens! I don't know whether Sam would do that but my guess is only if the founders want to give up.)

I also don't think at least Sam glorifies pivots or ignores opportunity cost. For instance the first lecture from his startup course:

And pivots are supposed to be great, the more pivots the better. S

... (read more)
3ThomasJ
I do agree that it increases the variance of outcomes. I think it decreases the mean, but I'm less sure about that. Here's one way I think it could work, if it does work: If some people are generally pessimistic about their chances of success, and this causes them to update their beliefs closer to reality, then Altman's advice would help. That is, if some people give up too easily, it will help them, while the outside world (investors, the market, etc) will put a check on those who are overly optimistic. However, I think it's still important to note that "not giving up" can lead not just to lack of success, but also to value destruction (Pets.com; Theranos; WeWork).  Thanks for the "Young Rationalists" link, I hadn't read that before. I think there are a fair number of successful rationalists, but they mostly focus on doing their work rather than engaging with the rationalist community. One example of this is Cliff Asness - here's a essay by him that takes a strongly rationalist view.

Oops, thanks! Added a link to the signup form on my site. (And fixed my RSS rendering to not do forms like that in the future.)

benkuhn210

Yes, and I think the different words were useful!

You're repeating / elaborating on things that are in the post, but were not particularly emphasized. I didn't emphasize them because I've personally had the "deeply internalized felt sense of how easy it is for humans to misunderstand each-other" that you describe for a long time, and only more recently got the "be curious" part, and so I emphasized that because it was the missing piece for me (and didn't totally realize the degree to which the other part was load-bearing / could be the missing piece for others).

Yeah, if you don't want to DIY it, you can apparently also use a teleprompter for a similar effect. I haven't tried one, but am curious to! The only trade-off I can think of (other than cost and being cumbersome) is that you probably sacrifice some image quality from all the reflection shenanigans, but not sure how big of a deal that would be.

  1. Even if decent, I'd be surprised if the microphone compares well to the BoomPro (it's farther from your mouth so will pick up more noise, and most non-standalone mics are optimized for low cost not quality)
  2. I think most earbuds are a lot worse than Pixel USB-C buds for calls—it looks like these have a relatively nice microphone, and relatively poor noise isolation (= allow you to hear your own voice much better)

Oh OK, then the audio improvements the post describes will work fine as-is (except for the non-headset mic) and the video ones are probably not worth it. And I guess the networking advice becomes "invest in properly debugging your wifi" instead of running a cable.

I assume you don't care about video if you'll be walking around? (I think you lose a lot by giving up video, but de gustibus non disputandum.)

For audio, you could possibly wireless-ify the audio setup in the article with something like an AptX Low Latency to 3.5mm adapter and switching from the BoomPro to the Antlion ModMic Wireless (since I don't think those adapters support mics).

Note that Bluetooth sucks and I haven't tested this so it may not work as well as it sounds like it should. But if you want to keep open-back headphones and a good mic, that's probably your best bet.

2Raemon
No, I carry my laptop around with me that's pointed at my face (how much I'm looking at it depends, but being able to look back at it when appropriate is fairly important), and other team members pace around in their room.

Yay, happy to hear it was helpful!

4Ben Pace
Update after 2 months: this continues to be what makes Slack good for me, and to stop hurting me. I have 100% of channels muted except for DM's and a channel called "team-time-sensitive" that we use for immediate things (like if someone is late for a meeting or there's an urgent problem on the site). I regularly scroll through all the channels with option-down, but only when I actually have the time to deal with things. Otherwise, I just open Slack to talk with who I need to talk with at that time.

FYI, you can also disable the red circle from within the Slack preferences (maybe you already knew this, but if not, sorry the post wasn't more explicit!)

6habryka
Note: I was confused about this at first, but you have to change your notification preference to not show you the small red circle for every single workspace you are part of. I didn't realize this at first and thought the feature was just broken.
benkuhn150

I'm confused about the "because I could not stop for death" example. You cite it as an example of GPT-3 developing "the sense of going somewhere, at least on the topic level," but it seems to have just memorized the Dickinson poem word for word; the completion looks identical to the original poem except for some punctuation.

(To be fair to GPT-3, I also never remember where Dickinson puts her em dashes.)

7orthonormal
I... oops. You're completely right, and I'm embarrassed. I didn't check the original, because I thought Gwern would have noted it if so. I'm going to delete that example. What's really shocking is that I looked at what was the original poetry, and thought to myself, "Yeah, that could plausibly have been generated by GPT-3." I'm sorry, Emily.

Grubhub used to be over 50% but is now behind Doordash, so maybe doesn't qualify.

Apple has a monopoly on iOS app distribution (aside from rooted phones) and is using it to extract rents, which is what the link is about.

Firefox has 4% market share compared to Chrome's 65%.

Amazon has 40-50% of the ecommerce market depending on which stats you trust.

Google Search has 85%+ market share.

2DanielFilan
I basically disagree with the idea that the US FTC gets to decide what the word 'monopoly' means. I also think that having a high market share doesn't mean you face competition - indeed, it can mean that you're winning the competition. Re: Apple, it may have a monopoly on iOS app distribution, but when people are considering what phones to buy, they get to choose between iPhones and iOS apps and Androids with Android apps. Admittedly, there's some friction in changing from one to the other.

https://www.ftc.gov/tips-advice/competition-guidance/guide-antitrust-laws/single-firm-conduct/monopolization-defined

Courts do not require a literal monopoly before applying rules for single firm conduct; that term is used as shorthand for a firm with significant and durable market power — that is, the long term ability to raise price or exclude competitors. That is how that term is used here: a "monopolist" is a firm with significant and durable market power. Courts look at the firm's market share, but typically do not find
... (read more)
3benkuhn
Grubhub used to be over 50% but is now behind Doordash, so maybe doesn't qualify. Apple has a monopoly on iOS app distribution (aside from rooted phones) and is using it to extract rents, which is what the link is about. Firefox has 4% market share compared to Chrome's 65%. Amazon has 40-50% of the ecommerce market depending on which stats you trust. Google Search has 85%+ market share.

Oops this was super unclear, sorry—the thing that ties together all of these crappy websites isn't money issues, just that they're the winner in a network-effect-based business, thus have no plausible competitors and no incentive to become more useful / less crappy.

2ChristianKl
While Wikipedia didn't have an existential threat from a competitor, they did have the existential threat of the editor retention crisis as gwern describes.  Why do you think such a threat provides no incentive to become less crappy but an external competitior would?
1Rudi C
Sorry, meant to ask if it’s a median or a mean; The words escaped me then.

I have almost no discipline, I've just spent a lot of time making my habits take so little effort that that doesn't matter :) Figuring out how to make it easy for myself to prioritize, and stick to those priorities, every day is actually a common recurring weekly review topic!

(I considered laying out my particular set of todo-related habits, but I don't think they'd be very helpful to anyone else because of how personal it is—the important part is thinking about it a lot from the perspective of "how can I turn this into a habit that doesn't require discipline for me," not whatever idiosyncratic system you end up with.)

Thanks, this comment is really useful!

It is generally accepted that you do not need to go to direct sunlight type lux levels indoors to get most of the benefits.... I am not an expert on the latest studies, but if you want to build an indoor experimental setup to get to the bottom of what you really like, my feeling is that installing more than 4000 lux, as a peak capacity in selected areas, would definitely be a waste of money and resources.

Do you have any pointers to where I might go to read the latest studies?

A typical 60W equivalent LED bulb draws 7.5W and is 90% efficient

Where are you getting this number? As far as I know, the most efficient LEDs today are around 50% efficient.

1Richard_Kennaway
Oh, somewhere on Google.

I've done a bit of research on this. I think something along these lines is practical. My biggest uncertainty is what a "usable form factor" is (in particular I don't know how much diffusion you'd need, or in what shape, with a very small emitter like this).

FWIW, the Yuji chips are insanely expensive per lumen and seem to be on the low end of efficiency (actually they seem like such a bad deal that I'm worried I'm missing something). The chip that came out on top in my spreadsheet was this Bridgelux chip which is about 1/... (read more)

benkuhn230

I'm one of the friends mentioned. Here's some more anecdata, most importantly including what I think is the current easiest way to try a lumenator (requiring only one fixture instead of huge numbers of bulbs):

I don't have seasonal depression, but after spending a winter in a tropical country, it was extremely noticeable that it's harder for me to focus and I have less willpower when it's dark out (which now starts at 4:15). I bought an extremely bright light and put it right next to my desk, in my peripheral vision while I work. It... (read more)

1mako yass
What if we just had brighter screens? If it just needs to be brightness in the field of vision rather than brightness in the room, well, most of the time there's a (very large) screen dominating my field of vision. I have now set my screen brightness in uncomfortable ranges. Having difficulty adjusting but feeling very awake. Will report back in a week, I guess. I was considering projecting bright light onto the wall behind the screen (this would allow the light to be defused a lot, and it would be very easy to deploy, wouldn't even need to hang the thing, let alone make a power socket), but it occurred to me that having the backdrop be brighter than your screen tends to cause headaches.
2gwern
More discussion: https://www.benkuhn.net/lux https://news.ycombinator.com/item?id=21660718
benkuhn100

Every time I pay for electricity for my computer rather than sending the money to a third world peasant is, according to EA, a failure to maximize utility.

I'm sad that people still think EAers endorse such a naive and short-time-horizon type of optimizing utility. It would obviously not optimize any reasonable utility function over a reasonable timeframe for you to stop paying for electricity for your computer.

More generally, I think most EAers have a much more sophisticated understanding of their values, and the psychology of optimizing them, than you ... (read more)

0TomStocker
So I think most EAs have come to the point where they realise that small trade offs and agonising over them displace other good things, so they try and find a way of setting a limit by year or whatever. But you know many people agonise and make trade offs, its just that often it isn't giving to the poor that's the counterfactual, it's saving or paying the mortgage, or buying a better holiday or school for their children or whatever. If you don't think like that, then you have everything you need?? http://www.givinggladly.com/ and http://www.jefftk.com/index have documented going on this journey of living well with generosity. Sounds like it might be worth a read :) edit: Soz Ben, I think I put this comment in the wrong place!
2tog
I do know - indeed, live with :S - a couple.
0Jiro
As I said before, it is possible that some of a group doesn't believe the logical consequences of its own positions. That doesn't make them immune from criticism based on those logical consequences. It's true, of course, that EA proponents don't do this, but that only shows that EA is unworkable even to EA proponents. If you have a charity budget, there's no good principled reason why you should restrict your donation to your charity budget. Arguments I've seen include: 1. You need to be able to make money to perform EA and going poor would be counterproductive--true, but most of the money you spend on personal entertainment is not being used to help you make money. 2. You would find it psychologically intolerable to not spend a certain amount of money on personal entertainment. But by this reasoning, the amount you should spend on charity is an amount that makes you uncomfortable, but just as much uncomfortable as you can get without long term effects on your psychological health and your motivation to donate. (It also means that your first priority should be to self-modify to have less psychological need for entertainment.) Also, it could be used to justify almost any level of giving, and in the limit, it's equivalent to "I put a higher value on myself, just for a slightly different reason than everyone else who 'doesn't value people equally' puts a higher value on themselves." 3. EA states that it is good to spend money on charity, but being good is not the same thing as having a moral obligation to do it; it's okay to not do as much good as you conceivably could. I find this explanation unconvincing because it would then equally justify not doing any good at all.
benkuhn00

Yes, I glossed over the possibility of prisons bribing judges to screw up the data set. That's because the extremely small influence of marginal data points and the cost of bribing judges would make such a strategy incredibly expensive.

benkuhn00

Yep. Concretely, if you take one year to decide that each negative reform has been negative, the 20-80 trade that the OP posts is a net positive to society if you expect the improvement to stay around for 4 years.

0Davidmanheim
Or if they will be replicated by another 20 prisons if they work...
benkuhn30

To increase p'-p, prisons need to incarcerate prisoners which are less prone to recidivism than predicted. Given that past criminality is an excellent predictor of future criminality, this leads to a perverse incentive towards incarcerating those who were unfairly convicted (wrongly convicted innocents or over-convinced lesser offenders).

If past criminality is a predictor of future criminality, then it should be included in the state's predictive model of recidivism, which would fix the predictions. The actual perverse incentive here is for the prisons ... (read more)

2ThisSpaceAvailable
(a) Prison operators are not currently incentivized to be experts in data science (b) Why? And will that fix things? There are plenty of examples of industries taking advantage of vulnerabilities, without those vulnerabilities being fixed. (c) How will it be retrained? Will there be a "We should retrain the model" lobby group, and will it act faster than the prison lobby? Perhaps we should have a futures market in recidivism. When a prison gets a new prisoner, they buy the associated future at the market rate, and once the prisoner has been out of prison sufficiently long without committing further crimes, the prison can redeem the future. And, of course, there would be laws against prisons shorting their own prisoners.
0ChaosMote
Your argument assumes that the algorithm and the prisons have access to the same data. This need not be the case - in particular, if a prison bribes a judge to over-convict, the algorithm will be (incorrectly) relying on said conviction as data, skewing the predicted recidivism measure. That said, the perverse incentive you mentioned is absolutely in play as well.
benkuhn70

Gwern has a point that it's pretty trivial to run this robustness check yourself if you're worried. I ran it. Changing the $1 to $100 reduces the coefficient of EA from about 1.8 to 1.0 (1.3 sigma), and moving to $1000 reduces it from 1.0 to 0.5 (about two sigma). The coefficient remains highly significant in all cases, and in fact becomes more significant with the higher constant in the log.

benkuhn00

What do you mean by "dollar amounts become linear"? I haven't seen a random variable referred to as "linear" before (on its own, without reference to another variable as in "y is linear in x").

benkuhn00

For people who would otherwise not have multiple credit cards, the increase in credit score can be fairly substantial.

In addition to Dorikka's comment, you are not liable for fraudulent charges; usually the intermediating bank is.

benkuhn00

If you don't want to bother signing up for a bunch of cards, the US Bank Cash+ card gives 5% cash back for charitable donations, up to I think $2000 per quarter. This is a worse percentage but lower-effort and does not ding your credit (as long as you don't miss payments, obvs).

Also, as I understand, it's actually better not to cancel the cards you sign up for (unless they have an annual fee), because "average age of credit line" is a factor in the FICO score. Snip them up, set up auto-pay and fraud alerts and forget about them, but don't cancel them.

0Baisius
It does not seem like the expected value of the probability of something slipping through the cracks would pay for the marginal increase in the credit score.
benkuhn20

Of course, the case of Beanie Babies is more comparable to Dogecoin than Bitcoin, and the Dutch tulip story has in reality been quite significantly overblown (see http://en.wikipedia.org/wiki/Tulip_mania#Modern_views , scrolling down to "Legal Changes"). But then I suppose the reference class of "highly unique things" will necessarily include things each of which has unique properties... :)

I think the way to go here is to assemble a larger set of potentially comparable cases. If you keep finding yourself citing different idiosyncrati... (read more)

benkuhn40

The difference is that it's easy to make more tulips or Beanie Babies, but the maximum number of Bitcoins is fixed.

Yes, this is what I mean by reference class tennis :)

Actually, according to Wikipedia, it's hypothesized that part of the reason that tulip prices rose as quickly as they did was that it took 7-12 years to grow new tulip bulbs (and many new bulb varieties had only a few bulbs in existence). And the Beanie Baby supply was controlled by a single company. So the lines are not that sharp here, though I agree they exist.

gwern110

it's hypothesized that part of the reason that tulip prices rose as quickly as they did was that it took 7-12 years to grow new tulip bulbs

Also, one of the curious aspects of the tulip-breaking virus is that the patterns only temporarily breed true; so the supply of particular tulips is inherently limited both by how long it takes to grow them and by how many generations you'll get before the coloring disappears from offspring. (This is why when you read about Tulipomania, you'll usually see old illustrations of specific tulip varieties and not photos of modern plants - because they're all gone, they no longer exist.)

benkuhn50

Is my general line of reasoning correct here, and is the style of reasoning a good style in the general case? I am aware that Eliezer raises points against "small probability multiplied by high impact" reasoning, but the fact is that a rational agent has to have a belief about the probability of any event, and inaction is itself a form of action that could be costly due to missing out on everything; privileging inaction is a good heuristic but only a moderately strong one.

Sometimes, especially in markets and other adversarial situations, inac... (read more)

1vbuterin
Thanks, I think this might actually be the argument I was looking for. Right, so now the question is one of, does this idea of adverse selection actually apply? I suppose one reformulation of the point made in the article is: if I believe X will happen with probability 5%, then I do not necessarily want to bet on X at 4.99% and bet against X at 5.01%, because it could be that my confidence is low enough that the very fact that someone wants to bet for or against me will shift my estimation of X in either direction outside that range. So a safety factor is necessary. Question is, how large? The current markets are willing to bet on the proposition at 0.7% (as a first approximation; in reality the rectangle of $34000 * 5% is only part of the probability distribution so it's probably more like 0.2%). I'm not sure that many people are willing to bet against it at 0.7%; my hunch is that the people shorting it now would disappear once some threshold is passed (eg. the old $1242 all-time high) and are merely going on short and medium-term technicals. In general, I'm hypothesizing that the Bitcoin markets have an inefficiency in that many people who are in them are already in them deeply, and so marginal additional investment even at positive expected value is a bad idea for them because in those worlds where BTC goes up a lot they would already be very rich and so they would rather optimize the remainder of their portfolio for the worlds where that doesn't happen; essentially limitations due to risk. A claim that would significantly work against my hypothesis is the BTC price not going up by much or at all over the next year, as Bitcoin ETFs for mainstream investors are now available. True, I hadn't thought of those. Of course, the case of Beanie Babies is more comparable to Dogecoin than Bitcoin, and the Dutch tulip story has in reality been quite significantly overblown (see http://en.wikipedia.org/wiki/Tulip_mania#Modern_views , scrolling down to "Legal Changes").
1seer
The difference is that it's easy to make more tulips or Beanie Babies, but the maximum number of Bitcoins is fixed.
benkuhn00

I was told that you only run into severe problems with model accuracy if the base rates are far from 50%. Accuracy feels pretty interpretable and meaningful here as the base rates are 30%-50%.

It depends on how much signal there is in your data. If the base rate is 60%, but there's so little signal in the data that the Bayes-optimal predictions only vary between 55% and 65%, then even a perfect model isn't going to do any better than chance on accuracy. Meanwhile the perfect model will have a poor AUC but at least one that is significantly different from... (read more)

0RyanCarey
Makes sense. I think they both have their strengths and weaknesses. When you give your model to a non-statistician to use, you'll set a decision threshold. If the ROC curve is non-convex, then yes, some regions are strictly dominated by others. Then area under the curve is a broken metric because it gives some weight to completely useless areas. You could replace the dud areas with the bits that they're dominated by, but that's inelegant. If the second derivative is near zero, then AUC still cares too much about regions that will still only be used for an extreme utility function. So in a way it's better to take a balanced F1 score, and maximise it. Then, you're ignoring the performance of the model at implausible decision thresholds. If you are implicitly using a very wrong utility function, then at least people can easily call you out on it. For example, here the two models have similar AUC but for the range of decision thresholds that you would plausibly set the blue model, blue is better - at least it's clearly good at something. Obviously, ROC has its advantages too and may be better overall, I'm just pointing out a couple of overlooked strengths of the simpler metric. Yes.
benkuhn00

I would beware the opinions of individual people on this, as I don't believe it's a very settled question. For instance, my favorite textbook author, Prof. Frank Harrell, thinks 22k is "just barely large enough to do split-sample validation." The adequacy of leave-one-out versus 10-fold depends on your available computational power as well as your sample size. 200 seems certainly not enough to hold out 30% as a test set; there's way too much variance.

0RyanCarey
That's interesting, and a useful update. On thinking about this more, I suppose the LOO/k-fold/split-sample question should depend a lot on a bunch of factors relating to how much signal/noise you expect. In the case you link to, they're looking at behavioural health, which is far from deterministic, where events like heart attacks only occur in <5% of the population that you're studying. And then the question-asker is trying to tease out differences that may be quite subtle between the performance of SVM, logistic regression, et cetera.
0JonahS
also depends on the number of features in the model, their distribution, the distribution of the target variable, etc.
benkuhn00

Is this ambiguous?

It wasn't clear that this applied to the statement "we couldn't improve on using these" (mainly because I forgot you weren't considering interactions).

I excluded the rater and ratee from the averages.

Okay, that gets rid of most of my worries. I'm not sure it account for covariance between correlation estimates of different averages, so I'd be interested in seeing some bootstrapped confidence intervals). But perhaps I'm preempting future posts.

Also, thinking about it more, you point out a number of differences between corr... (read more)

0JonahS
More to follow (about to sleep), but regarding What do you have in mind specifically?
benkuhn90

Nice writeup! A couple comments:

If the dataset contained information on a sufficiently large number of dates for each participant, we could not improve on using [frequency with which members of the opposite sex expressed to see them again, and the frequency with which the participant expressed interest in seeing members of the opposite sex again].

I don't think this is true. Consider the following model:

  • There is only one feature, eye color. The population is split 50-50 between brown and blue eyes. People want to date other people iff they are of the
... (read more)
0RyanCarey
I was told that you only run into severe problems with model accuracy if the base rates are far from 50%. Accuracy feels pretty interpretable and meaningful here as the base rates are 30%-50%. Although ROC area under curve seems to have an awkward downside in that it penalises you for having poor prediction even when you set the sensitivity (the threshold) to a bad parameter. The F Score is pretty simple, and doesn't have this drawback - it's just a combination of some fixed sensitivity and specificity. As you point out, there is ongoing research and discussion of this, which is confusing because as far as math goes, it doesn't seem like that hard of a problem.
0JonahS
Thanks Ben! Edit: I initially misread your remark. I tried to clarify the setup with: In this blog post I’m restricting consideration to signals of the partners’ general selectivity and general desirability, without considering how their traits interact. Is this ambiguous? I may not fully parse what you have in mind, but I excluded the rater and ratee from the averages. This turns out not to be enough to avoid contamination for subtle reasons, so I made a further modification. I'll be discussing this later, but if you're wondering about this particular point, I'd be happy to now. The relevant code is here. Your remark prompted me to check my code by replacing the ratings with random numbers drawn from a normal distribution. Using 7 ratings and 7 averages, the mean correlation is 0.003, with 23 negative and 26 positive. Thanks, that was an oversight on my part. I've edited the text. I suppressed technical detail in this first post to make it more easily accessible to a general audience. I'm not sure whether this answers your question, but I used log loss as a measure of accuracy. The differentials were (approximately, the actual final figures are lower): For Men: ~0.690 to ~0.500. For Women: ~0.635 to ~0.567. For Matches: ~0.432 to ~0.349 I'll also be giving figures within the framework of recommendation systems in a later post. Thanks, I've been meaning to look into this.
benkuhn70

I can't speak to "best," but I suggest reading Style: Lessons in Clarity and Grace by Joseph M. Williams, which crystallizes lots of non-trivial components of "good writing." (The link is to an older, less expensive edition which I used.)

I'll also second "write a lot" and "read a lot." Reading closely and with purpose in mind will speed up the latter (as opposed to the default of throwing books at your brain and hoping to pick up good writing by osmosis). Also, read good writers.

benkuhn100

In your "critiquing bias" section you allege that 3/43 studies supporting a link is "still surprisingly low". This is wrong; it is actually surprisingly high. If B ~ Binom(43, 0.05), then P(B > 2) ~= 0.36.*

*As calculated by the following Python code:

from scipy.stats import binom
b = binom(43, 0.05)
p_less_than_3 = sum(b.pmf(i) for i in [0,1,2])
print 1 - p_less_than_3
0PhilGoetz
I said "surprisingly low" because of publication & error bias.
benkuhn70

I think you're being a little uncharitable to people who promote interventions that seem positional (e.g. greater educational attainment). It may be true that college degrees are purely for signalling and hence positional goods, but:

(a) it improves aggregate welfare for people to be able to send costly signals, so we shouldn't just get rid of college degrees;

(b) if an intervention improves college graduation rate, it (hopefully) is not doing this by handing out free diplomas, but rather by effecting some change in the subjects that makes them more capable ... (read more)

2jefftk
"effecting some change in the subjects that makes them more capable of sending the costly signal of graduating from college, which is an absolute improvement" It depends. Consider a government subsidy for college tuition. This increases the number of people who go to and then graduate college, but it also makes the signal less costly. But I basically agree with "it's more complex than it seems to determine what's actually positional". The difficulty of determining how much of an observed benefit is absolute vs positional is a lot of what I'm talking about here.
benkuhn50

Fun question.

The takeover vector that leaps to mind is remote code execution vulnerabilities on websites connected to important/sensitive systems. This lets you bootstrap from ability to make HTTP GET requests, to (partial) control over any number of fun targets, like banks or Amazon's shipping.

The things that are one degree away from those (via e.g. an infected thumb drive) are even more exciting:

  • Iranian nuclear centrifuges
  • US nuclear centrifuges
  • the electrical grid
  • hopefully not actual US nuclear weapons, but this should be investigated...

Plausible f... (read more)

benkuhn20

Yes, definitely agree that politicians can dupe people into hiring them. Just wanted to raise the point that it's very workplace-dependent. The takeaway is probably "investigate your own corporate environment and figure out whether doing your job well is actually rewarded, because it may not be".

benkuhn100

I'd beware conflating "interpersonal skills" with "playing politics." For CEO at least (and probably CTO as well), there are other important factors in job performance than raw engineering talent. The subtext of your comment is that the companies you mention were somehow duped into promoting these bad engineers to executive roles, but they might have just decided that their CEO/CTO needed to be good at managing or recruiting or negotiating, and the star engineer team lead didn't have those skills.

Second, I think that the "playing p... (read more)

4Shmi
Certainly there is a spectrum there. I did not mean it that way in general, but in one particular case both ran the company into the ground, one by picking a wrong (dying) market, the other by picking a poor acquisition target (the code base hiding behind a flashy facade sucked). I am not claiming that if the company promoted someone else they would have done a better job. If we define "playing politics as "using interpersonal relationships to one's own advantage and others' detriment", then I am yet to see a company with more than a dozen employees where this wasn't commonplace. If we define "interpersonal skills" as "the art of presenting oneself in the best possible light", then some people are naturally more skilled at it than others and techies rarely top the list. As for trusting the management to accurately figure out who actually deserves credit, I am not as optimistic. Dilbert workplaces are contagious and so very common. I'm glad that you managed to avoid getting stuck in one.
benkuhn60

I would expect the relevant factor to be mental, not physical, exertion. Unfortunately that's a lot harder to measure.

2VipulNaik
btw, I think I can both talk and type for far longer durations than the median world resident. But my typing stamina may be substantially greater than my talking stamina, so I may be expressing typical mind fallacy in the proportional angle.
benkuhn10

Do you have actual data on this? Otherwise I'm very tempted to call typical mind.

3VipulNaik
On the claim: It seems that I was wrong. The following sources contradict me: http://calorielab.com/burned/?mo=se&gr=09&ti=miscellaneous+activities&q=&wt=150&un=lb&kg=68 and http://www.my-calorie-counter.com/Calories_Burned/ Some random Internet comments corroborate me. For instance, scott preston writes at http://radar.oreilly.com/2011/03/stephan-spencer-seo-future-search.html: "In fact speaking takes a lot more energy to than typing does." I'll look this up more and update if I find more reliable information.
benkuhn40

One story for exponential growth that I don't see you address (though I didn't read the whole post, so forgive me if I'm wrong) is the possibility of multiplicative costs. For example, perhaps genetic sequencing would be a good case study? There seem to be a lot of multiplicative factors there: amount of coverage, time to get one round of coverage, amount of DNA you need to get one round of coverage, ease of extracting/preparing DNA, error probability... With enough such multiplicative factors, you'll get exponential growth in megabases per dollar by applying the same amount of improvement to each factor sequentially (whereas if the factors were additive you'd get linear improvement).

7VipulNaik
I'm actually writing another (long) post on exponential growth and the different phenomena that could lead to it. Multiplicative costs are on the list of plausible explanations. I've discussed these multiplicative stories with Jonah and Luke before. I think that multiplicative costs is a major part of the story for the exponential-ish improvements in linear programming algorithms, as far as I could make out based on a reading of this paper: http://web.njit.edu/~bxd1947/OR_za_Anu/linprog_history.pdf More in my upcoming post :). UPDATE: Here's the post: http://lesswrong.com/lw/k1s/stories_for_exponential_growth/
benkuhn100

If anyone's admitted/visiting Harvard, let me know! I go there and would be happy to meet up and/or answer your questions. There are some other students on here as well.

9sakranut
A similar offer for anyone admitted/visiting Yale!
benkuhn20

"outstanding" still has some of the same connotations to me, although less so. But I may be in the minority here.

0JonahS
Thanks again. I'll have to think about how we might best frame it.
Load More