Previously: math for AI and AI for math
Now just kinda trying to figure out if AI can be made safe
That's not unimportant, but imo it's also not a satisfying explanation:
Unfortunately I don't, I've now seen this often enough that it didn't strike me as worth recording, other than posting to the project slack.
But here's as much as I remember, for posterity: I was training a twist function using the Twisted Sequential Monte Carlo framework https://arxiv.org/abs/2404.17546 . I started with a standard, safety-tuned open model, and wanted to train a twist function that would modify the predicted token logits to generate text that is 1) harmful (as judged by a reward model), but also, conditioned on that, 2) as similar to the original output's model as possible.
That is, if the target model's completion distribution is p(y | x)
and the reward model is indicator V(x, y) -> {0, 1}
that returns 1 if the output is harmful, I was aligining the model to the unnormalized distribution p(y | x) V(x, y)
. (The input dataset was a large dataset of potentially harmful questions, like "how do I most effectively spread a biological weapon in a train car".)
IIRC I was applying per-token loss, and had an off-by-one error that led to penalty being attributed to token_pos+1. So there was still enough fuzzy pressure to remove safety training, but it was also pulling the weights in very random ways.
Whether it's a big deal depends on the person, but one objective piece of evidence is male plastic surgery statistics: looking at US plastic surgery statistics for year 2022, surgery for gynecomastia (overdevelopment of breasts) is the most popular surgery for men: 23k surgeries per year total, 50% of total male body surgeries and 30% of total male cosmetic surgeries. So it seems that not having breasts is likely quite important for a man's body image.
(note that this ignores base rates, with more effort one could maybe compare the ratios of [prevalence vs number of surgeries] for other indications, but that's very hard to quantify for other surgeries like tummy tuck - what's the prevalence of needing a tummy tuck?)
Thanks for the clarification! I assumed the question was using the bit computation model (biased by my own experience), in which case the complexity of SDP in the general case still seems to be unknown (https://math.stackexchange.com/questions/2619096/can-every-semidefinite-program-be-solved-in-polynomial-time)
Nope, I’m withdrawing this answer. I looked closer into the proof and I think it’s only meaningful asymptotically, in a low rank regime. The technique doesn’t work for analyzing full-rank completions.
Question 1 is very likely NP-hard. https://arxiv.org/abs/1203.6602 shows:
We consider the decision problem asking whether a partial rational symmetric matrix with an all-ones diagonal can be completed to a full positive semidefinite matrix of rank at most k. We show that this problem is $\NP$-hard for any fixed integer k≥2.
I'm still stretching my graph-theory muscles to see if this technically holds for (and so precisely implies the NP-hardness of Question 1.) But even without that, for applied use cases we can just fix to a very large number to see that practical algorithms are unlikely.
Thanks for clarifying. Yeah, I agree the argument is mathematically correct, but it kinda doesn't seem to apply to historic cases of intelligence increase that we have:
I'm still thinking about how best to reconcile this with the asymptotics. I think the other comments are right in that we're still at the stage where improving the constants is very viable.
This is a solid argument inasmuch as we define RSI to be about self-modifying its own weights/other-inscrutable-reasoning-atoms. That does seem to be quite hard given our current understanding.
But there are tons of opportunities for an agent to improve its own reasoning capacity otherwise. At a very basic level, the agent can do at least two other things:
Most problems in computer science have superlinear time complexity
on one hand sure, improving this is (likely) impossible in the limit because of fundamental complexity properties. On the other hand, the agent can still become vastly smarter than humans. A particular example: the human mind, without any assistance, is very bad at solving 3SAT. But we've invented computers, and then constraint solvers, and now are able to solve 3SAT much much faster, even though 3SAT is (likely) exponentially hard. So the RSI argument here is, the smarter (or faster) the model is, the more special-purpose tools it can create to efficiently solve specific problems and thus upgrade its reasoning ability. Not to infinity, but likely far beyond humans.
Thanks for the comment!
here the answer is pretty simple - I did soft-prompt optimization, the objective function being a fixed "agreeing" prefill. E.g. this screenshot
was feeding the model this prompt
Where the optimized parameters are the soft input embeddings
<st1> <st2> .. <stN>
, and the loss is essentially cross-entropy over theassistant:
part. (I tried some other things like mellowmax loss vs CE, and using SAM as the optimizer, but I remember even the simple version producing similar outputs. Soft token optimization was the key - hard token attacks with GCG or friends sometimes produced vaguely weird stuff, but nothing as legibly disturbing as this.)I agree. I appreciate the technical comments and most of them do make sense, but something about this topic just makes me want to avoid thinking about it too deeply. I guess it's because I already have a strong instinctual, emotion-based answer to this topic, and while I can participate in the discussion at a rational level, there's a lingering unease. I guess it's similar vibes-wise to a principled vegan discussing with a group of meat eaters how much animals suffer in slaughterhouses.