All of Elliot Callender's Comments + Replies

How much would you say (3) supports (1) on your model? I'm still pretty new to AIS and am updating from your model.

I agree that marginal improvements are good for fields like medicine, and perhaps so too AIS. E.g. I can imagine self-other overlap scaling to near-ASI, though I'm doubtful about stability under reflection. I'll put 35% we find a semi-robust solution sufficient to not kill everyone.

Given my model, I think 20% generalizability is worth a person's time. Given yours, I'd say 1% is enough.

I think that the distribution of success probability of typ... (read more)

I strongly think cancer research has a huge space and can't think of anything more difficult within biology.

I was being careless / unreflective about the size of the cancer solution space, by splitting the solution spaces of alignment and cancer differently; nor do I know enough about cancer to make such claims. I split the space into immunotherapies, things which target epigenetics / stem cells, and "other", where in retrospect the latter probably has the optimal solution. This groups many small problems with possibly weakly-general solutions into a "bott... (read more)

9Ariel
Thank you, that was very informative. I don't find the "probability of inclusion in final solution" model very useful, compared to "probability of use in future work" (similarly for their expected value versions) because 1. I doubt that central problems are a good model for science or problem solving in general (or even in the navigation analogy). 2. I see value in impermanent improvements (e.g. current status of HIV/AIDS in rich countries) and in future-discounting our value estimations. 3. Even if a good description of a field as a central problem and satalite problems exists, we are unlikely to correctly estimate it apriori, or estimate the relevance of a solution to it. In comparison, predicting how useful a solution is to "nearby" work is easier (with the caveat that islands or cliques of only internaly-useful problems and solutions can arise, and do in practice). Given my model, I think 20% generalizability is worth a person's time. Given yours, I'd say 1% is enough.

I think general solutions are especially important for fields with big solution spaces / few researchers, like alignment. If you were optimizing for, say, curing cancer, it might be different (I think both the paradigm-and subproblem-spaces are smaller there).

From my reading of John Wentworth's Framing Practicum sequence, implicit in his (and my) model is that solution spaces for these sorts of problems are apriori enormous. We (you and I) might also disagree on what apriori feasibility would be "weakly" vs "strongly" generalizable; I think my transition is around 15-30%.

4Ariel
I see what you mean with regards to the number of researchers. I do wonder a lot about the amount of waste from multiple researchers unknowingly coming up with the same research (a different problem to what you pointed out) and the uncoordinated solution to that is to work on niche problems and ideas (which coincidentally seem less likely to individually generalize). Could you share your intuition for why the solution space in AI alignment research is large, or larger than in cancer? I don't have an intuition about the solution space in alignment v.s. a "typical" field, but I strongly think cancer research has a huge space and can't think of anything more difficult within biology. Intuition: you can think of fighting cancer as a two player game, with each individual organism being an instance and the within-organism evolutionary processes leading to cancer being the other player. In most problems in biology you can think about the genome of the organisms as defining rule set for the game, but here the genome isn't held constant. w.r.t. to the threshold to strong generalizability, I don't have a working model of this to disagree with. Putting confidence/probability values on relatively abstract things is something I'm not used to (and I understand is encouraged here), so I'd appreciate if you could share a little more insight about how to 1. Define the baseline distribution generalizability is defined on. 2. Give a little intuition about why a threshold is meaningful, rather than a linear "more general is better". I'm sorry if that's too much to ask for or ignorant of something you've laid out before. I have no intuition about #2, but for #1 I suspect that you have a model of AI alignment or an abstract field as having a single/few central problems at a time, whereas my intuition is that they are mostly composed of many problems, with "centrality" only being relevant as a storytelling device or a good way to make sense of past progress in retrospect (so that ge

Shoot, thanks. Hopefully it's clearer now.

3RogerDearnaley
Yes: thanks!

Yes, I agree. I expect abstractions, typically, to involve much more than 4-8 bits of information. On my model, any neural network, be it MLP, KAN or something new, will approximate abstractions with multiple nodes in parallel when the network is wide enough. I.e. the causal graph I mentioned is very distinct from the NN which might be running it.

Though now that you mentioned it, I wonder if low-precision NN weights are acceptable because of some network property (maybe SGD is so stochastic that higher precision doesn't help) or the environment (maybe natural latents tend to be lower-entropy)?

Anyways, thanks for engaging. It's encouraging to see someone comment.

Answer by Elliot Callender30

This one was a lot of fun!

  1. ROS activity in some region of the body is a function of antioxidant bioavailability, heat, and oxidant bioavailability. I imagine this relationship is the inverse of some chemical rate laws, i.e. dependent on which antioxidants we're looking at. But since I expect most antioxidants to work as individual molecules, the relationship is probably , i.e. ROS activity is inverse w.r.t. some antioxidant's potency and concentration if we ignore other antioxidants. The bottom term can also be a sum across all antioxidants, given n
... (read more)
Answer by Elliot Callender30
  1. Sleep / wakefulness; hypnagogia seems transient and requires conscious effort to maintain. Outside stimuli and internal volition can wake people up; lack thereof can do the opposite.
  2. Friendships; I tend to have few, close friendships. I don't interact much with more distant friends because it's less emotionally fulfilling, so they slowly fade towards being acquaintances. I distance myself from people I don't connect with / feel safe around, and try to strengthen bonds with people I think are emotionally mature and interesting.
  3. Focus; I tend to either be chec
... (read more)
Answer by Elliot Callender30

I know this post is old(ish), but still think this exercise is worth doing!

  1. Deep ocean currents; I expect changes in ocean floor topography and deep-water inertial/thermal changes to matter. I don't expect shallow-water topography to matter, nor wind (unless we have sustained 300+kph winds for weeks straight).
  2. Earth's magnetic pole directions; I'm not sure what causes them. I think they're generated by induction from magma movement? In that case, our knobs are those currents. I don't think anything can change the equilibrium without changing the flow pattern
... (read more)

I agree that this seems like a very promising direction.

Beyond that, we of course want our class of random variables to be reasonably general and cognitively plausible as an approximation - e.g. we shouldn’t assume some specific parametric form.

Could you elaborate on this; "reasonably general" sounds to me like the redundancy axiom, so I'm unclear about whether this sentence is an intuition pump.

4johnswentworth
That sentence was mostly just hedging. The intended point was that the criteria which we focused on in the post aren't necessarily the be-all end-all.

I think it depends on which domain you're delegating in. E.g. physical objects, especially complex systems like an AC unit, are plausibly much harder to validate than a mathematical proof.

In that vein, I wonder if requiring the AI to construct a validation proof would be feasible for alignment delegation? In that case, I'd expect us to find more use and safety from [ETA: delegation of] theoretical work than empirical.

6Davidmanheim
That seems a lot like Davidad's alignment research agenda.
Answer by Elliot Callender74

I'm in a very similar situation, graduating next spring with a math degree in the US. I'll sketch out my personal situation (to help contextualize my advice) followed my approach for career scouting. If you haven't checked out 80k hours, I really suggest doing so, because they have much more thorough and likely wiser advice than I do.

I'm a 19-year-old undergrad in a rural part of the US. My dad's a linguistics professor and pushing me to do a PhD. I want to do AI safety research, and am currently weighing the usefulness of a PhD compared to saving money to... (read more)

1Chico
I have looked at 80k hours, although I haven't read through much of their career advice (The last year has been hectic). I will try to read more into that after finishing my last assignment when I find the time.  The idea of doing a Master's or PhD is attractive, I have applied to a fair number, but like I said I'm unsure of the usefulness/impact of research and further learning in that way compared to finding a job or cause to work on. So far my ideal situation would be some sort of job or personal endeavor that would allow me to do post-grad part-time. AI alignment is definitely one of many interests, although I feel that it's a big focus at the moment and might be oversaturated (I'm not sure but worth thinking about). Lightcone looks interesting although I'm not sure about relocating to the States. I like the idea of writing down how I feel about something for extra thinking space, I have been doing similar things for a while but I'll try doing more of that. I think the big thing I'm looking for is if there's any way of quantifying the problem, I suspect that this isn't the perfect way of approaching this mostly unquantifiable problem but my nature requires me to look for it regardless. I understand that this feels very self-destructive/unhealthy but I think that this belief is a natural consequence of my base beliefs about reality. Strangely enough, I don't project it onto others as that would be unfair to them, but still see it as an obligation for myself. In any case, it also extends to self-care; if I don't look after myself first then I will be unable to work or learn and therefore useless anyway.

We know that some genes are only active in the womb, or in childhood, which should make us very skeptical that editing them would have an effect.

Would these edits result in demethylated DNA? A reversion of the epigenome could allow expression of infant genes. There may also be robust epigenomic therapies developed by the time this project would be scalable.

Companies like 23&Me genotyped their 12 millionth customer two years ago and could probably get at perhaps 3 million customers to take an IQ test or submit SAT scores.

Just as you mentioned acade... (read more)