All of dynomight's Comments + Replies

If there was some kind of app that could recognize the content and make it searchable, that would indeed have most of the advantages of paper.

(Most chalkboards/whiteboard don't spark as much joy as high-quality paper/pen in my opinion, but I reckon a good blackboard with good chalk does?)

I'd hate to convince you to stop using paper, but I use this Obsidian Excalidraw plugin for making drawings and I find it to be reaaallllly fast: https://github.com/zsviczian/obsidian-excalidraw-plugin

It's kinda clunky but fundamentally I find it incredibly "non-frustrating" compared to all other tools. I guess you can try the editor in your browser here: https://excalidraw.com/

2Seth Herd
I installed it and did like one or two test diagrams and then never again. I should get back to it because it did seem good.

I move that we think of paper and notes software as complements. Certainly, notes software is much better for almost any purpose where you're actually going to be referencing the notes repeatedly. But for the purpose of "make the neurons in your brain fire good", paper still can't be beat.

(This post was written by first scribbling on paper and then retyping and editing in, umm, Obsidian.)

4Seth Herd
Right. The post did inspire me to maybe get a new notebook for the first time in years for that reason.   I've been using Obsidian exculsively, but it's really reduced how much diagramming I do. To me the speed does make up for being forced to write on lines and in a limited number of styles. I haven't really gotten skilled enough in it to quickly diagram in its editor. I'm not sure there's much other advantage to paper for making neurons fire good. I wonder if it puts you into thinking mode based on associations or something? Or if staring at a blank page on which you can write or draw anywhere evokes a more openminded and analytical state for anyone?  

(Sorry for the slow reply—just saw this.)

> What is the alternative explanation for why semaglutide would disincline people who would have had small change scores from participating or incline people who have large change scores to participate (remember, this is within-subjects) in the alcohol self-administration experiment?

I'm a bit unsure what the non-alternative explanation is here. But imagine that semaglutide does not reduce the urge to drink but—I don't know—makes people more patient, or makes them more likely to agree to do things doctors ask them... (read more)

What premises would I have to accept for the comparison to be fair? Suppose I think that available compute will continue to grow along previous trends and that we'll continue to find new tricks to turn extra compute into extra capabilities. Does conditioning on that make it fair? (Not sure I accept those premises, but never mind that.)

The problem is deeper than that.

Playing a game of chess takes hours. LLMs are pretty bad it, but we have had good chess engines for decades - why isn’t there a point way off on the top left for chess?

Answer: we’re only interested in highly general AI agents, which basically means LLMs. So we’re only looking at the performance of LLMs, right? But if you only look at LLM performance without scaffolding, it looks to me like that asymptotes around 15 minutes. Only by throwing in systems that use a massive amount of inference time compute do we recover a line w... (read more)

Thanks for the response! I must protest that I think I'm being misinterpreted a bit. Compare my quote:

the point of RCTs is to avoid resorting to regression coefficients on non-randomized sample

To the:

The point of RCTs is not to avoid resorting to regression coefficients.

The "non-randomized sample" part of that quote is important! If semaglutide had no impact on the decision to participate, then we can argue about about the theory of regressions. Yes, the fraction that participated happened to be close, but with small numbers that could easily happen by cha... (read more)

1Daniel V
I guess I misunderstood you. I figured that without "regression coefficients," the sentence would be a bit tautological: "the point of randomized controlled trial is to avoid [a] non-randomized sample," and there were other bits that made me think you had an issue with both selection bias (agree) and regressions (disagree). I share your overall takeaway, but at this point I am just genuinely curious why the self-selection is presumed to be such a threat to internal validity here. I think we need more attention to selection effects on the margin, but I also think there is a general tendency for people to believe that once they've identified a selection issue the results are totally undermined. What is the alternative explanation for why semaglutide would disincline people who would have had small change scores from participating or incline people who have large change scores to participate (remember, this is within-subjects) in the alcohol self-administration experiment? Maybe those who had the most reduced cravings wanted to see more of what these researchers could do? But that process would also occur among placebo, so it'd work via the share of people with large change scores being greater in the semaglutide group, which is...efficacy. There's nuance there, but hard to square with lack of efficacy. That said, still agree that the results are no slam dunk. Very specific population, very specific outcomes affected, and probably practically small effects too.

It ranges from 0% to 100%.

 

Small nitpick that doesn't have any significant consequences—this isn't technically true, it could be higher than 100%.

Wow, I didn't realize bluesky already supports user-created feeds, which can seemingly use any algorithm? So if you don't like "no algorithm" or "discover" you can create a new ranking method and also share it with other people? 

Anyone want to create a lesswrong starter pack? Are there enough people on bluesky for that to be viable?

9hmys
https://bsky.app/profile/hmys.bsky.social/post/3lbd7wacakn25 I made one. A lot of people are not here, but many people are.

Well done, yes, I did exactly what you suggested! I figured that an average human lifespan was "around 80 years" and then multiplied and divided by 1.125 to get 80×1.125=90 and 80/1.125=71.111.

(And of course, you're also right that this isn't quite right since (1.125 - 1/1.125) / (1/1.125) = (1.125)²-1 = .2656 ≠ .25. This approximation works better for smaller percentages...)

Interesting. Looks like they are starting with a deep tunnel (530 m) and may eventually move to the deepest tunnel in Europe (1444 m). I wish I could find numbers on how much weight will be moved or the total energy storage of the system. (They say quote 2 MW, but that's power, not energy—how many MWh?)

According to this article, a Swiss company is building giant gravity storage buildings in China and out of 9 total buildings, there should be a total storage of 3700 MWh, which seems quite good! Would love to know more about the technology.

You're 100% right. (I actually already fixed this due to someone emailing me, but not sure about the exact timing.) Definitely agree that  there's something amusing about the fact that I screwed up my manual manipulation of units while in the process of trying to give an example of how easy it is to screw up manual manipulations of units...

You mentioned a density of steel of 7.85 g/cm^3 but used a value of 2.7 g/cm^3 in the calculations.

 

Yes! You're right! I've corrected this, though I still need to update the drawing of the house. Thank you!

dynomight910

Word is (at least according to the guy who automated me) that if you want an LLM to really imitate style, you really really want to use a base model and not an instruction-tuned model like ChatGPT. All of ChatGPT's "edge" has been worn away into bland non-offensiveness by the RLHF. Base models reflect the frightening mess of humanity rather than the instructions a corporation gave to human raters. When he tried to imitate me using instruction-tuned models it was very cringe no matter what he tried. When he switched to a base model it instantly got my voice... (read more)

5lsusr
Yeah, I like that ChatGPT does what I tell it to, that it doesn't decay into crude repetition, and that it doesn't just make stuff up as much as the base LLM, but in terms of attitude and freedom, I prefer edgy base models. I don't want a model that's "safe" in the sense that it does what its corporate overlords want. I want a model that's safe like a handgun, in the sense that it does exactly what I tell it to.

Why somewhat? It's plausible to me that even just the lack of DHA would give the overall RCT results.

 

Yeah, that seems plausible to me, too. I don't think I want to claim that the benefits are "definitely slightly lower", but rather that they're likely at least a little lower but I'm uncertain how much. My best guess is that the bioactive stuff like IgA does at least something, so modern formula still isn't at 100%, but it's hard to be confident.

My impression was that the backlash you're describing is causally downstream of efforts by public health people to promote breastfeeding (and pro-breastfeeding messages in hospitals, etc.) Certainly the correlation is there (https://www.researchgate.net/publication/14117103_The_Resurgence_of_Breastfeeding_in_the_United_States) but I guess it's pretty hard to prove a strict cause.

I'm fascinated that caffeine is so well-established (the most popular drug?) and yet these kinds of self-experiments still seem to add value over the scientific literature.

Anyway, I have a suspicion that tolerance builds at different rates for different effects. For example, if you haven't had any caffeine in a long time (like months), it seems to create a strong sense of euphoria. But this seems to fade very quickly. Similarly, with prescription stimulants, people claim that tolerance to physical effects happens gradually, but full tolerance never develop... (read more)

dynomight*50

(Many months later) Thanks for this comment, I believe you are right! Strangely, there do seem to be many resources that list them as being hydrogen bonds (e.g. Encyclopedia Brittanica: https://www.britannica.com/science/unsaturated-fat which makes me question their editorial process.) In any case, I'll probably just rephrase to avoid using either term. Thanks again, wish I had seen this earlier!

2bakhsv
Well, EB article you linked doesn't state directly that fatty acids are made out of carbon atoms linked via hydrogen bonds. It has two sentences relevant to the topic, and I am not entirely sure how to parse them: The first sentence is (almost) fully correct. The second sentence, if viewed without the first one, may technically also be correct, but for what I know it's not and also it's not what they meant. See, fatty acids are capable of forming actual hydrogen bonds with each other with their "acid" parts (attached the picture from my organic chem course). On the left covalent bonds are shown with solid lines and hydrogen bonds are shown with dashed lines. The "fatty" part of the molecule is hidden under the letter R. On the right there is methyl instead of R (ie it's vinegar) and hydrogen bonds are not shown—molecules are just oriented in the right way. (I'm really sorry if I'm overexplaining, I just want to make it understandable for people with different backgrounds). So, if interpreted literally, the second sentence states that unsaturated fatty acids form less hydrogen bonds with each other for whatever reason, and that's why they are liquid instead off solid. The explanation I've heard many times is different, it says that they are liquid because their "fatty" part is bent because double bonds have different geometry, so it is harder for them to form a crystal. I mean, it is still possible that they also form less hydrogen bonds, but I bet it's insignificant even if true. But it honestly looks like they don't mean all of that at all, they are just incorrectly calling covalent bonds between carbon and hydrogen "hydrogen bonds" and they also don't know what they mean by "the structures are weaker". It's still a sin, but not the one you are accusing them of. I am also completely fine with the phrasing that is currently in the article and I'm sorry for wasting your time with all that overthinking, hope it wasn't totally useless.
dynomight20

Thanks, any feedback on where the argument fails? (If anywhere in particular.)

I would dissuade no one from writing drunk, and I'm confident that you too can say that people are penguins! But I'm sorry to report that personally I don't do it by drinking but rather writing a much longer version with all those kinds of clarifications included and then obsessively editing it down.

Do you happen to have any recommended pointers for research on health impacts of processed food? It's pretty easy to turn up a few recent meta reviews, which seems like a decent place to start, but I'd be interested if there were any other sources, particularly influential individual experiments, etc. (It seems like there's a whole lot of observational studies, but many fewer RCTs, for reasons that I guess are pretty understandable.) It seems like some important work here might never use the word "processing".

3Freyja
I don't remember individual studies but two books that might be helpful: Ultra-Processed People by Chris van Tulleken Metabolical by Robert Lustig  UPP is terribly written and I imagine mostly useful for its bibliography (I skimmed it in an hour or so). Metabolical is better (although far too difficult a read to be a successful popsci book), although it isn't specifically focused on processing techniques (it in particular discusses stripping out fibre, adding sugars, reducing water, as some major processing techniques with big issues). You might find something helpful looking in the refs section of either book. 

If I hadn't heard back from them, would you want me to tell you? Or would that be too sad?

5Kabir Kumar
I want to know.

Seed oils are usually solvent extracted, which makes me wonder, how thoroughly are they scrubbed of solvent, what stuff in the solvent is absorbed into the oil (also an effective solvent for various things), etc

 

I looked into this briefly at least for canola oil. There, the typical solvent is hexane. And some hexane does indeed appear to make it into the canola oil that we eat. But hexane apparently has very low toxicity, and—more importantly—the hexane that we get from all food sources apparently makes up less than 2% of our total hexane intake! http... (read more)

3RedMan
https://www.mdpi.com/2304-8158/11/21/3412 more recent source on hexane tox.   I'm not just talking about the hexane (which isn't usually standardized enough to generalize about), I'm talking about any weird crap on the seed, in the hopper, in the hexane, or accumulated in the process machinery.  Hexane dissolves stuff, oil dissolves stuff, and the steam used to crash the hexane out of the oil also dissolves stuff, and by the way, the whole process is high temp and pressure. There's a ton of batch to batch variability and opportunity to introduce chemistry you wouldn't want in your body which just isn't present with "I squeezed some olives between two giant rocks" By your logic, extra virgin olive oil is a waste, just use the olive pomace oil, it's the same stuff, and the solvent extraction vs mechanical pressing just doesn't matter.

It's a regression. Just like they extrapolate backwards to (1882+50=1932) using data from 1959, they extrapolate forwards at the end. (This is discussed in the "timelines" section.) This is definitely a valid reason to treat it with suspicion, but nothing's "wrong" exactly.

Many thanks! All fixed (except one that I prefer the old way.)

As the original author of underrated reasons to be thankful (here), I guess I can confirm that tearing apart the sun for raw materials was not an intended implication.

1tcheasdfjkl
as Solstice creative lead I neither support nor oppose tearing apart the sun for raw materials

I think matplotlib has way too many ways to do everything to be comprehensive! But I think you could do almost everything with some variants of these.

ax.spines['top'].set_visible(False) # or 'left' / 'right' / 'bottom'
ax.set_xticks([0,50,100],['0%','50%','100%'])
ax.tick_params(axis='x', left=False, right=False) # or 'y'
ax.set_ylim([0,0.30])
ax.set_ylim([0,ax.get_ylim()[1]])

Good point regarding year tick marks! I was thinking think that labeling 0°C would make the most sense when freezing is really important. Say, if you were plotting historical data on temperatures and you were interested in trying to estimate the last frost date in spring or something. Then, 10°C would mean "twice as much margin" as 5°C.

dynomight2317

One way you could measure which one is "best" would be to measure how long it takes people to answer certain questions. E.g. "For what fraction of the 1997-2010 period did Japan spend more on healthcare per-capita than the UK?" or "what's the average ratio of healthcare spending in Sweden vs. Greece between 2000 and 2010?" (I think there is an academic literature on these kinds of experiments, though I don't have any references on hand.)

In this case, I think Tufte goes overboard in saying you shouldn't use color. But if the second plot had color, I'd ventu... (read more)

4jefftk
Yes! But not just time, you should also compare them on accuracy.
5Nathan Helm-Burger
Yeah, agreed that getting people to answer questions using the chart, and measuring their speed and accuracy is the key objective metric of design quality. Also, I like it when both color and line styles are used together. Keeps it clear for colorblind people, and makes it extra clear for colorsighted people. Choosing colors should be done carefully to balance contrast with the background color. And can be done in such a way as to be visible even to the most common colorblindness types.

Thanks, someone once gave me the advice that after you write something, you should go back to the beginning and delete as many paragraphs as you can without making everything incomprehensible. After hearing this, I noticed that most people tend to write like this:

  • Intro
  • Context
  • Overview
  • Other various throat clearing
  • Blah blah blah
  • Finally an actual example, an example, praise god

Which is pretty easy to correct once you see it!

4cubefox
This is a good point. Beginning in medias res seems also one of the reasons why posts by Eliezer Yudkowsky and Scott Alexander are so readable. But for long posts I think a short abstract in the beginning is actually helpful, perhaps highlighted in italics. Unfortunately some people use the abstract as a mere teaser ("... wanna know how I came to that startling conclusion? Guess you have to read the whole paper/post, hehe") rather than as a proper spoiler of the main insights. "Spoiler" sounds bad from the perspective of the author ("will people still read my post when I already revealed the punchline in the abstract?"), but a spoiler can actually provide motivation to read the whole post for "fun" reasons. E.g. by going "I already agree with that claim in the abstract, let's indulge in confirming my preconceptions!" or "I disagree with that claim, guess I have to read the post so I can write a rebuttal in the comments!" Not very rational, but better than not being motivated to read the post at all. Though you probably use other tricks to make a post more readable. From your post above I inferred these points: * Use examples * Include images if possible * Don't clutter the post with a lot of distracting links and footnotes * Include rhetorical questions * short sentences * delete unnecessary tangents to make the post shorter That's what I thought anyway. Maybe you could share your own tips? "How to Write Readable Posts"

Hey, you might be right! I'll take this as useful feedback that the argument wasn't fully convincing. Don't mean to pull a motte-and-bailey, but I suppose if I had to, I'd retreat to an argument like, "if making a plot, consider using these rules as one option for how to pick axes." In any case, if you have any examples where you think following this advice leads to bad choices, I'd be interested to hear them.

2kave
When I looked at your proposed GDP-Time chart, I felt I was more inclined to treat the year as quantitative and the amounts as categorical. Though I don't know how that would actually play out if I were trying to make use of it in anger.
Answer by dynomight73

I think you're basically right: Correlation is just one way of measuring dependence between variables. Being correlated is a sufficient but not necessary condition for dependence. We talk about correlation so much because:

  1. We don't have a particularly convenient general scalar measure of how related two variables are. You might think about using something like mutual information, but for that you need the densities not datasets.
  2. We're still living in the shadows of the times when computers weren't so big. We got used to doing all sorts of stuff based on linearity decades ago because we didn't have any other options, and they became "conventional" even when we might have better options now.
1notfnofn
Suppose we don't have any prior information about the dataset, only our observations. Is any metric more accurate than assuming our dataset is the exact distribution and calculating mutual information? Kind of like bootstrapping.

Thanks, you've 100% convinced me. (Convincing someone that something that (a) is known to be true and (b) they think isn't surprising, actually is surprising is a rare feat, well done!)

Chat or instruction finetuned models have poor prediction cailbration, whereas base models (in some cases) have perfect calibration.

 

Tell me if I understand the idea correctly: Log-loss to predict next token leads to good calibration for single token prediction, which manifests as good calibration percentage predictions? But then RLHF is some crazy loss totally removed from calibration that destroys all that?

If I get that right, it seems quite intuitive. Do you have any citations, though?

Sune138

I don’t find it intuitive at all. It would be intuitive if you started by telling a story describing the situation and asked the LLM to continue the story, and you then sampled randomly from the continuations and counted how many of the continuations would lead to a positive resolution of the question. This should be well-calibrated, (assuming the details included in the prompt were representative and that there isn’t a bias of which types of ending the stories are in the training data for the LLM). But this is not what is happing. Instead the model outpu... (read more)

7ReaderM
https://imgur.com/a/3gYel9r https://openai.com/research/gpt-4

Sadly, no—we had no way to verify that.

I guess one way you might try to confirm/refute the idea of data leakage would be to look at the decomposition of brier scores: GPT-4 is much better calibrated for politics vs. science but only very slightly better at politics vs. science in terms of refinement/resolution. Intuitively, I'd expect data leakage to manifest as better refinement/resolution rather than better calibration.

That would definitely be better, although it would mean reading/scoring 1056 different responses, unless I can automate the scoring process. (Would LLMs object to doing that?)

Thank you, I will fix this! (Our Russian speaker agrees and claims they noticed this but figured it didn't matter 🤔) I re-ran the experiments with the result that GPT-4 shifted from a score of +2 to a score of -1.

Well, no. But I guess I found these things notable:

  • Alignment remains surprisingly brittle and random. Weird little tricks remain useful.
  • The tricks that work for some models often seem to confuse others.
  • Cobbling together weird little tricks seems to help (Hindi ranger step-by-step)
  • At the same time, the best "trick" is a somewhat plausible story (duck-store).
  • PaLM 2 is the most fun, Pi is the least fun.

You've convinced me! I don't want to defend the claim you quoted, so I'll modify "arguably" into something much weaker.

4jam_brand
Also perhaps of interest might be this discussion from the SSC subreddit awhile back where someone detailed their pro-Bigfoot case.

I don't think I have any argument that it's unlikely aliens are screwing with us—I just feel it is, personally.

I definitely don't assume our sensors are good enough to detect aliens. I'm specifically arguing we aren't detecting alien aircraft, not that alien aircraft aren't here. That sound like a silly distinction, but I'd genuinely give much higher probability to "there are totally undetected alien aircraft on earth" than "we are detecting glimpses of alien aircraft on earth."

Regarding your last point, I totally agree those things wouldn't explain the we... (read more)

I know that the mainstream view on Lesswrong is that we aren't observing alien aircraft, so I doubt many here will disagree with the conclusion. But I wonder if people here agree with this particular argument for that conclusion. Basically, I claim that:

  • P[aliens] is fairly high, but
  • P[all observations | aliens] is much lower than P[all observations | no aliens], simply because it's too strange that all the observations in every category of observation (videos, reports, etc.) never cross the "conclusive" line.

As a side note: I personally feel that P[observat... (read more)

2AnthonyC
Even if there are aliens, and humans do sometimes gain data showing such, if the aliens are sufficiently advanced and don't want to be found, I would not be surprised if they selectively took away our conclusive data but left behind the stuff that's already indistinguishable from noise. Kinda like how we take our trash with us after hiking and camping, but don't worry too much in most places about our footprints or the microscopic bits of material our gear and bodies leave behind.
1Dweomite
The general point that you need to update on the evidence that failed to materialize is in the sequences and is exactly where I expected you to go based on your introductory section.
5gjm
I make no claim to speak for anyone who isn't me, but I agree with your analysis. I would say similar things about e.g. ESP and miracles and the like.
2simon
Glitches happen. Misunderstandings happen. Miscommunications happen. Coincidences happen. Weird-but-mundane things happen. Hoaxes happen. To use machine learning terminology, the real world occurs at temperature 1. We shouldn't expect P[observations] to be high - that would require temperature less than 1. The question is, is P[observations] surprisingly low, or surprisingly high for some different paradigm, to such an extent as would provide strong evidence for something outside of current paradigms? My assessment is no. (see my discussion of Nimitz for example) Some additional minor remarks specifically on P[aliens]: * non-detection of large (in terms of resource utilization) alien civilizations implies that the density of interstellar-spacefaring civilizations is low - I don't expect non-expansion to be the common (let alone overwhelmingly selected) long term choice, and even aestivating civilizations should be expected to intervene to prevent natural entropy generation (such as by removing material from stars to shut them down) * If the great filter (apart from the possible filter against resource-utilization expansion by interstellar-spacefaring civilizations, which I consider unlikely to be a significant filter as mentioned above) is almost entirely in the abiogenesis step, and interstellar panspermia isn't too hard, then it would make sense for a nearby civilization to exist as Robin Hanson points out. I do actually consider it fairly likely that a lot of the great filter is in abiogenesis, but note that there needs to be some combination of weak additional filter between abiogenesis and spacefaring civilization or highly efficient panspermia for this scenario to be likely. * If a nearby, non-expanding interstellar-spacefaring civilization did exist, then of course it could, if it so chose, mess with us in a way that left hints but no solid proof. They could even calibrate their hints across multiple categories of observations, and adjust over time, to m

I get very little value from proofs in math textbooks, and consider them usually unnecessary (unless they teach a new proof method).

 

I think the problem is that proofs are typically optimized for "give most convincing possible evidence that the claim is really true to a skeptical reader who wants to check every possible weak point". This is not what most readers (especially new readers) want on a first pass, which is "give maximum possible into why this claim is true for to a reader who is happy to trust the author if the details don't give extra intuition." At a glance, infinite Napkin seems to be optimizing much more for the latter.

If you're worried about computational complexity, that's OK. It's not something that I mentioned because (surprisingly enough...) this isn't something that any of the doctors discussed. If you like, let's call that a "valid cost" just like the medical risks and financial/time costs of doing tests. The central issue is if it's valid to worry about information causing harmful downstream medical decisions.

0anonymousaisafety
I'm sorry, but I just feel like we've moved the goal posts then. I don't see a lot of value in trying to disentangle the concept of information from 1.) costs to acquire that information, and 2.) costs to use that information, just to make some type of argument that a certain class of actor is behaving irrationally. It starts to feel like "assume a spherical cow", but we're applying that simplification to the definition of what it means to be rational. First, it isn't free to acquire information. But second, even if I assume for the sake of argument that the information is free, it still isn't free to use it, because computation has costs. if a theory of rational decision making doesn't include that fact, it'll come to conclusions that I think are absurd, like the idea that the most rational thing someone can do is acquire literally all available information before making any decision.

I might not have described the original debate very clearly. My claim was that if Monty chose "leftmost non-car door" you still get the car 2/3 of the time by always switching and 1/3 by never switching. Your conditional probabilities look correct to me. The only thing you might be "missing" is that (A) occurs 2/3 of the time and (B) occurs only 1/3 of the time. So if you always switch your chance of getting the car is still (chance of A)*(prob of car given A) + (chance of B)*(prob of car given B)=(2/3)*(1/2) + (1/3)*(1) = (2/3).

One difference (outside the... (read more)

1Dacyn
Ah, I see, fair enough.

Just to be clear, when talking about how people behave in forums, I mean more "general purpose" places like Reddit. In particular, I was not thinking about Less Wrong where in my experience, people have always bent over backwards to be reasonable!

I have two thoughts related to this:

First, there's a dual problem: Given a piece of writing that's along the Pareto frontier, how do you make it easy for readers who might have a utility function aligned with the piece to find it.

Related to this, for many people and many pieces of writing, a large part of the utility they get is from comments. I think this leads to dynamics where a piece where the writing that's less optimal can get popular and then get to a point on the frontier that's hard to beat.

I loved this book. The most surprising thing to me was the answer that people who were there in the heyday give when asked what made Bell Labs so successful: They always say it was the problem, i.e. having an entire organization oriented towards the goal of "make communication reliable and practical between any two places on earth". When Shannon left the Labs for MIT, people who were there immediately predicted he wouldn't do anything of the same significance because he'd lose that "compass". Shannon was obviously a genius, and he did much more after than most people ever accomplish, but still nothing as significant as what he did when at at the Labs.

I thought this was fantastic, very thought-provoking. One possibly easy thing that I think would be great would be links to a few posts that you think have used this strategy with success.

9johnswentworth
Drawing from my own posts: * Many of the abstraction research posts used this strategy. I was trying to pump out updates at least ~weekly, and most weeks I didn't have a proof for a new theorem or anything like that. The best I could do was explain whatever I was thinking about, and why it seemed interesting/important. * Some of my best posts (IMO) came from looking at why I believed some idea, finding a ball of illegible intuitions, and untangling that ball. The constraints/scarcity posts all came from that process, the review of Design Principles of Biological Circuits came from that process, Everyday Lessons From High-Dimensional Optimization and various posts on gears-level models came from that process, Whats So Bad About Ad-Hoc Mathematical Definitions? came from this process, probably many others. * Core Pathways of Aging would never have been finished if I'd tried to hunt down every source.

Thanks, I clarified the noise issue. Regarding factor analysis, could you check if I understand everything correctly? Here's what I think is the situation:

We can write a factor analysis model (with a single factor) as

where:

  1. is observed data
  2. is a random latent variable
  3. is some vector (a parameter)
  4. is a random noise variable
  5. is the covariance of the noise (a parameter)

It always holds (assuming and are independent) that

In the simplest variant of factor analysis (in the current post) we use in which cas... (read more)

2Radford Neal
Assuming you're using "C" to denote Covariance ("Cov" is more common), that seems right. It's typical that the noise covariance is diagonal, since a general covariance matrix for the noise would render use of a latent variable unnecessary (the whole covariance matrix for x could be explained by the covariance matrix of the "noise", which would actually include the signal as well).  (Though it could be that some people use a non-diagonal covariance matrix that is subject to some other sort of constraint that makes the procedure meaningful.) Of course, it is very typical for people to use factor analysis models with more than one latent variable.  There's no a priori reason why "intelligence" couldn't have a two-dimensional latent variable.  In any real problem, we of course don't expect any model that doesn't produce a fully general covariance matrix to be exactly correct, but it's scientifically interesting if a restricted model (eg, just one latent variable) is close to being correct, since that points to possible underlying mechanisms.

Thanks for pointing out those papers, which I agree can get at issues that simple correlations can't. Still, to avoid scope-creep, I've taken the less courageous approach of (1) mentioning that the "breadth" of the effects of genes is an active research topic and (2) editing the original paragraph you linked to to be more modest, talking about "does the above data imply" rather than "is it true that". (I'd rather avoid directly addressing 3 and 4 since I think that doing those claims justice would require more work than I can put in here.) Anyway, thanks again for your comments, it's useful for me to think of this spectrum of different "notions of g".

Load More