All of Razied's Comments + Replies

Razied60

More insightful than what is conserved under the scaling symmetry of ReLU networks is what is not conserved: the gradient. Scaling  by  scales  by  and  by , which means that we can obtain arbitrarily large gradient norms by simply choosing small enough . And in general bad initializations can induce large imbalances in how quickly the parameters on either side of the neuron learn. 

Some time ago I tried training some networks while setting these symmetries to the values th... (read more)

4Lucius Bushnaq
See footnote 1.
Razied4-9

I suspect the expert judges would need to resort to known jailbreaking techniques to distinguish LLMs. A fair interesting test might be against expert-but-not-in-ML judges.

Razied2-4

Sorry to be blunt, but any distraction filter that can be disabled through the chrome extension menu is essentially worthless. Speaking from experience, for most people this will work for exactly 3 days until they find a website they really want to visit and just "temporarily" disable the extension in order to see it.

Razied20

For #5, I think the answer would be to make the AI produce the AI safety ideas which not only solve alignment, but also yield some aspect of capabilities growth along an axis that the big players care about, and in a way where the capabilities are not easily separable from the alignment. I can imagine this being the case if the AI safety idea somehow makes the AI much better at instruction-following using the spirit of the instruction (which is after all what we care about). The big players do care about having instruction-following AIs, and if the way to do that is to use the AI safety book, they will use it. 

2Donald Hobson
  So firstly, in this world capability is bottlenecked by chips. There isn't a runaway process of software improvements happening yet. And this means there probably aren't large easy capabilities software improvements lying around.  Now "making capability improvements that are actively tied to alignment somehow" sounds harder than making any capability improvement at all. And you don't have as much compute as the big players. So you probably don't find much. What kind of AI research would make it hard to create a misaligned AI anyway? A new more efficient matrix multiplication algorithm that only works when it's part of a CEV maximizing AI?  Likely somewhat true.  Perhaps. Don't underestimate sheer incompetence. Someone pressing the run button to test the code works so far, when they haven't programmed the alignment bit yet. Someone copying and pasting in an alignment function but forgetting to actually call the function anywhere. Misspelled variable names that are actually another variable. Nothing is idiot proof.   I mean presumably alignment is fairly complicated and it could all go badly wrong because of the equivalent of one malfunctioning o-ring.  Or what if someone finds a much more efficient approach that's harder to align. 
Razied20

Do you expect Lecun to have been assuming that the entire field of RL stops existing in order to focus on his specific vision?

4tailcalled
I'm not sure he has coherent expectations, but I'd expect his vibe is some combination of "RL doesn't currently work" and "fields generally implement safety standards".
Razied112

Very many things wrong with all of that:

  1.  RL algorithms don't minimize costs, but maximize expected reward, which can well be unbounded, so it's wrong to say that the ML field only minimizes cost. 
  2. LLMs minimize expected log probability of correct token, which is indeed bounded at zero from below, but achieving zero in that case means perfectly predicting every single token on the internet. 
  3. The boundedness of the thing you're minimizing is totally irrelevant, since maximizing  is exactly the same as maximizing  w
... (read more)
2tailcalled
Yann LeCun's proposals are based on cost-minimization.
Razied1616

The word "privilege" has been so tainted by its association with guilt that it's almost an infohazard to think you've got privilege at this point, it makes you lower your head in shame at having more than others, and brings about a self-flagellation sort of attitude. It elicits an instinct to lower yourself rather than bring others up. The proper reactions to all these things you've listed is gratitude to your circumstances and compassion towards those who don't have them. And certainly everyone should be very careful towards any instinct they have at publ... (read more)

3keltan
Hmmm, I think the original post was an interesting idea. I think your comment points to something related but different. Perhaps taboo words?
localdeity1411

I grew up knowing "privilege" to mean a special right that was granted to you based on your job/role (like free food for those who work at some restaurants) or perhaps granted by authorities due to good behavior (and would be taken away for misusing it).  Note also that the word itself, "privi"-"lege", means "private law": a law that applies to you in particular.

Rights and laws are social things, defined by how others treat you.  To say that your physical health is a privilege therefore seems like either a category error, or a claim that other pe... (read more)

Razied90

Weird side effect to beware for retinoids: they make dry eyes worse, and in my experience this can significantly decrease your quality of life, especially if it prevents you from sleeping well.

Razied20

Basically, this shows that every term in a standard Bayesian inference, including the prior ratio, can be re-cast as a likelihood term in a setting where you start off unsure about what words mean, and have a flat prior over which set of words is true.

If the possible meanings of your words are a continuous one-dimensional variable x, a flat prior over x will not be a flat prior if you change variables to y = f(y) for an arbitrary bijection f, and the construction would be sneaking in a specific choice of function f.

Say the words are utterances about the probability of a coin falling heads, why should the flat prior be over the probability p, instead of over the log-odds log(p/(1-p)) ?

2DanielFilan
In my post, I didn't require the distribution over meanings of words to be uniform. It could be any distribution you wanted - it just resulted in the prior ratio of "which utterance is true" being 1:1.
Razied20

Most of the weird stuff involving priors comes into being when you want posteriors over a continuous hypothesis space, where you get in trouble because reparametrizing your space changes the form of your prior, so a uniform "natural" prior is really a particular choice of parametrization. Using a discrete hypothesis space avoids big parts of the problem.

2DanielFilan
Why wouldn't this construction work over a continuous space?
2Richard_Kennaway
Only if there is a "natural" discretisation of the hypothesis space. It's fine for coin tosses and die rolls, but if the problem itself is continuous, different discretisations will give the same problems that different continuous parameterisations do. In general, when infinities naturally arise but cause problems, decreeing that everything must be finite does not solve those problems, and introduces problems of its own.
Razied20

Wait, why doesn't the entropy of your posterior distribution capture this effect? In the basic example where we get to see samples from a bernoulli process, the posterior is a beta distribution that gets ever sharper around the truth. If you compute the entropy of the posterior, you might say something like "I'm unlikely to change my mind about this, my posterior only has 0.2 bits to go until zero entropy". That's already a quantity which estimates how much future evidence will influence your beliefs. 

2Richard_Ngo
The thing that distinguishes the coin case from the wind case is how hard it is to gather additional information, not how much more information could be gathered in principle. In theory you could run all sorts of simulations that would give you informative data about an individual flip of the coin, it's just that it would be really hard to do so/very few people are able to do so. I don't think the entropy of the posterior captures this dynamic.
Razied40

Surely something like the expected variance of  would be a much simpler way of formalising this, no? The probability over time is just a stochastic process, and OP is expecting the variance of this process to be very high in the near future.

2Richard_Ngo
The variance over time depends on how you gather information in the future, making it less general. For example, I may literally never learn enough about meteorology to update my credence about the winds from 0.5. Nevertheless, there's still an important sense in which this credence is more fragile than my credence about coins, because I could update it. I guess you could define it as something like "the variance if you investigated it further". But defining what it means to investigate further seems about as complicated as defining the reference class of people you're trading against. Also variance doesn't give you the same directional information—e.g. OP would bet on doom at 2% or bet against it at 16%. Overall though, as I said above, I don't know a great way to formalize this, and would be very interested in attempts to do so.
Razied20

Unfortunately the entire complexity has just been pushed one level down into the definition of "simple". The L2 norm can't really be what we mean by simple, because simply scaling the weights in a layer by A, and the weights in the next layer by 1/A leaves the output of the network invariant, assuming ReLU activations, yet you can obtain arbitrarily high L2 norms by just choosing A high enough. 

1Decaeneus
Agreed with your example, and I think that just means that L2 norm is not a pure implementation of what we mean by "simple", in that it also induces some other preferences. In other words, it does other work too. Nevertheless, it would point us in the right direction frequently e.g. it will dislike networks whose parameters perform large offsetting operations, akin to mental frameworks or beliefs that require unecessarily and reducible artifice or intermediate steps. Worth keeping in mind that "simple" is not clearly defined in the general case (forget about machine learning). I'm sure lots has been written about this idea, including here.
Razied32

Unfortunately if OpenAI the company is destroyed, all that happens is that all of its employees get hired by Microsoft, they change the lettering on the office building, and sama's title changes from CEO to whatever high level manager positions he'll occupy within microsoft.

Razied20

Hmm, but here the set of possible world states would be the domain of the function we're optimising, not the function itself. Like, No-Free-Lunch states (from wikipedia):

Theorem 1: Given a finite set  and a finite set  of real numbers, assume that  is chosen at random according to uniform distribution on the set  of all possible functions from  to . For the problem of optimizing  over the set , then no algorithm performs better than blind search.

Here  is the set o... (read more)

2Seth Herd
Here it is: The No Free Lunch theorem for dummies. See particularly the second section: Sidenote: Why NFL has basically nothing to do with AGI and the first link to Yudkowsky's post on essentially the same thing. I think the thing about your descripton is that S -> V is not going to be chosen at random in our world. The no free lunch theorem states in essence (I'm pretty sure) that no classifier can both classify a big gray thing with tusks and big ears as both an elephant and not-an-elephant. That's fine, because the remainder of an AGI system can choose (by any other criteria) to make elephants either a goal or an anti-goal or neither. If the NFL theorem applied to general intelligences, it seems like humans couldn't love elephants at one time and hate them at a later time, with no major changes to their perceptual systems. It proves too much.
Answer by Razied20

Obviously LLMs memorize some things, the easy example is that the pretraining dataset of GPT-4 probably contained lots of cryptographically hashed strings which are impossible to infer from the overall patterns of language. Predicting those accurately absolutely requires memorization, there's literally no other way unless the LLM solves an NP-hard problem. Then there are in-between things like Barack Obama's age, which might be possible to infer from other language (a president is probably not 10 yrs old or 230), but within the plausible range, you also just need to memorize it. 

2[anonymous]
Where it gets interesting is when you leave the space of token strings the machine has seen, but you are somewhere in the input space "in between" strings it has seen. That's why this works at all and exhibits any intelligence. For example if it has seen a whole bunch of patterns like "A->B", and "C->D", if you give input "G" it will complete with "->F". Or for President ages, what if the president isn't real? https://chat.openai.com/share/3ccdc340-ada5-4471-b114-0b936d1396ad
Razied84

There is no optimization pressure from “evolution” at all. Evolution isn’t tending toward anything. Thinking otherwise is an illusion.

Can you think of any physical process at all where you'd say that there is in fact optimization pressure? Of course at the base layer it's all just quantum fields changing under unitary evolution with a given Hamiltonian, but you can still identify subparts of the system that are isomorphic with a process we'd call "optimization". Evolution doesn't have a single time-independent objective it's optimizing, but it does seem to me that it's basically doing optimization on a slowly time-changing objective.

1Neil
Fair enough. I certainly didn't try to mince words. My goal was to violently shave off any idea of "agency" my friend was giving to evolution. He was walking around satisfied with his explanation that evolution selects for the fittest and is therefore optimizing for the fittest.[1]  The point of the dialogue format was to point out that you can call it an optimization process, but when you taboo that word you figure out it's hard to pinpoint exactly what is being optimized for. If you're going to call something an optimization process, you'd better tell me exactly what is being optimized for. If you can't, you are probably using that word as a curiosity stopper or something.  I think we'll be able to pinpoint what evolution optimizes for, someday. [2] Gravity as a force optimizes for the creation of stars: enough so that loose clouds of hydrogen are pretty much guaranteed to form stars. You could say "gravity optimizes for the creation of stars from hydrogen clouds" and anticipate experience with seamless accuracy. Evolution is like this except it's so much more complex that in order to explain it as an optimization process you'll have to resort to the dreaded word "emergence".  I think there's also something to be said about reminding people from time to time that "optimization pressure" and "emergence" and are in the map, not the territory; the territory is a different beast. I think you could reasonably take on the "true" way of seeing things for an hour or two after reading this post, and then go back to your business believing in the heuristic that evolution is an optimization process (once you've finished with your partial transfiguration).  1. ^ Note the verb "optimized", which implies that something active is going on. 2. ^ In fact, most of the work has probably been done by Dawkins and others and there's a mountain of math out there that explains exactly what evolution is optimizing for. If that's the case, I definitely want to under
Razied7-8

Why would you want to take such a child and force them to ‘emotionally develop’ with dumber children their own age?

Because you primarily make friends in school with people in your grade, and if you skip too many grades, the physical difference between the gifted kid and other kids will prevent them from building a social circle based on physical play, and probably make any sort of dating much harder.

1Gesild Muka
I would argue if skipping grades was normalized physical differences wouldn't have a large impact on socialization (making friends, dating, etc.)
8lsusr
Personal counterfactual: I was smarter than my peers and didn't skip any grades. Result: I didn't physically play with or date the other students. Exceptions: I did play football and did Boy Scouts, but those were both after-school activities. Moreover, neither of them were strictly segregated by age. Football was weight-based, and Boy Scouts lumped everyone from 11 to 17 into the same troop. Putting students in the same math class based on age (ignoring intelligence) is like putting students on the same football team based on age (ignoring size).
2Nathan Helm-Burger
Yeah, splitting up the "grades" into subjects and letting smart kids take advanced math classes but still have, e.g. civics class with students their own age seems like a better option. You didn't even need to send them to a different physical space to let them move at their own pace in math. Just let them read a good math textbook while the teacher lectures.

The physical difference matters, but the mental difference tends to matter more.

https://files.eric.ed.gov/fulltext/EJ746290.pdf

Abstract: A 20-year longitudinal study has traced the academic, social, and emotional development of 60 young Australians with IQs of 160 and above. Significant differences have been noted in the young people’s educational status and direction, life satisfaction, social relationships, and self-esteem as a function of the degree of academic acceleration their schools permitted them in childhood and adolescence. The considerable majo

... (read more)
Answer by Razied43

Predicting the ratio at t=20s is hopeless. The only sort of thing you can predict is the variance in the ratio over time, like the ratio as a function of time is  , where  . Here the large number of atoms lets you predict  , but the exact number after 20 seconds is chaotic. To get an exact answer for how much initial perturbation still leads to a predictable state, you'd need to compute the lyapunov exponents of an interacting classical gas system, and I haven't been able to find a paper that does this wit... (read more)

https://www.sciencedirect.com/science/article/abs/pii/S1674200121001279

They find Lyapunov exponent of about 1 or 2 (where time is basically in units of time it takes for a particle at average velocity to cover the length of the box).

For room temp gas, this timescale is about 1/400 seconds. So the divergence after 20 seconds should increase by a factor of over e^8000 (until it hits the cieling of maximum possible divergence).

Since an Angstrom is only 10^-10 m, if you start with an Angstrom offset, the divergence reaches maximum by about a tenth of a second.

5habryka
The goal is not to predict the ratio, but to just predict which side will have more atoms (no matter how small the margin). It seems very likely to me that any such calculation would be extremely prohibitively expensive and would approximately require logical omniscience.  To clarify this, we are assuming that without random perturbation, you would get 100% accuracy in predicting which side of the system has more atoms at t=20s. The question is how much of that 100% accuracy you can recover with a very very small unknown perturbation.
Razied30

I'll try to say the point some other way: you define "goal-complete" in the following way:

By way of definition: An AI whose input is an arbitrary goal, which outputs actions to effectively steer the future toward that goal, is goal-complete.

Suppose you give me a specification of a goal as a function  from a state space to a binary output. Is the AI which just tries out uniformly random actions in perpetuity until it hits one of the goal states "goal-complete"? After all, no matter the goal specification this AI will eventually hit it, th... (read more)

2Liron
I agree that if a goal-complete AI steers the future very slowly, or very weakly - as by just trying every possible action one at a time - then at some point it becomes a degenerate case of the concept. (Applying the same level of pedantry to Turing-completeness, you could similarly ask if the simple Turing machine that enumerates all possible output-tape configurations one-by-one is a UTM.) The reason "goal-complete" (or "AGI") is a useful coinage, is that there's a large cluster in plausible-reality-space of goal-complete agents with a reasonable amount of goal-complete optimization power (e.g. humans, natural selection, and probably AI starting in a few years), and another large distinguishable cluster of non-goal-complete agents (e.g. the other animals, narrow AI).
Razied20

E.g. I claim humans are goal-complete General Intelligences because you can give us any goal-specification and we'll very often be able to steer the future closer toward it.

If you're thinking of "goals" as easily specified natural-language things, then I agree with you, but the point is that turing-completeness is a rigorously defined concept, and if you want to get the same level of rigour for "goal-completeness", then most goals will be of the form "atom 1 is a location x, atom 2 is at location y, ..." for all atoms in the universe. And when averaged across all such goals, literally just acting randomly performs as well as a human or a monkey trying their best to achieve the goal.

2Liron
Hmm it seems to me that you're just being pedantic about goal-completeness in a way that you aren't symmetrically being for Turing-completeness. You could point out that "most" Turing machines output tapes full of 10^100 1s and 0s in a near-random configuration, and every computing device on earth is equally hopeless at doing that.
Razied120

Goal-completeness doesn't make much sense as a rigorous concept because of No-Free-Lunch theorems in optimisation. A goal is essentially a specification of a function to optimise, and all optimisation algorithms perform equally well (or rather poorly) when averaged across all functions.

There is no system that can take in an arbitrary goal specification (which is, say, a subset of the state space of the universe) and achieve that goal on average better than any other such system. My stupid random action generator is equally as bad as the superintelligence w... (read more)

6Seth Herd
The No Free Lunch theorem is irrelevant in worlds like ours that are a subset of possible data structures (world arrangements). I'm surprised this isn't better understood. I think Steve Byrnes did a nice writeup of this logic. I can find the link if you like.
2Liron
Well, I've never met a monkey that has an "optimization algorithm" by your definition. I've only met humans who have such optimization algorithms. And that distinction is what I'm pointing at. Goal-completeness points to the same thing as what most people mean by "AGI". E.g. I claim humans are goal-complete General Intelligences because you can give us any goal-specification and we'll very often be able to steer the future closer toward it. Currently, no other known organism or software program has this property to the degree that humans do. GPT-4 has it for an unprecedentedly large domain, by virtue of giving satisfying answers to a large fraction of arbitrary natural-language prompts.
Razied914

Zvi, you continue to be literally the best news aggregator on the planet for the stuff that I actually care about. Really, thanks a lot for doing this, it's incredibly valuable to me.

Razied31

Wouldn't lowering igf-1 also lead to really shity quality of life from lower muscle mass and much longer recovery times from injury?

2Hide
My first thought as well. IGF-1 exists for a reason. Growth is universally necessary for development, repair and function.
Razied5028

The proteins themselves are primarily covalent, but a quick google search says that the forces in the lipid layer surrounding cells are primarily non-covalent, and the forces between cells seem also non-covalent. Aren't those forces the ones we should be worrying about? 

It seems like Eliezer is saying "the human body is a sand-castle, what if we made it a pure crystal block?", and you're responding with "but individual grains of sand are very strong!"

3Nathan Young
Sure, but that does suggest that Yudkowsky could adjust his language a bit, right?

I mean... what if we did make it a pure crystal block? What would that do to the energy requirements for movement? Hydrogen bonds are pretty weak, but a cycle of "form-then-break" is pretty cheap. Covalent bonds are strong, but forming and breaking them repeatedly is energetically expensive.

1ThomasPilgrim
The structural proteins in the extracellular matrix and connective tissue (namely collagen and elastin) tend to have covalent crosslinks. So I I'm really not sure it's accurate to say that hydrogen bonds and van der Waals forces are what's holding the cells together.
3mabramov
That doesn't seem like the right analogy. The bonds are forced to fold over themselves because electrons repel each other and don't want to touch. So the natural structures are mostly tetrahedral structures. Think of the vertices of a tetrahedron having edges that shoot towards and meet at the centre and you will see that these form 109° angles. When you imagine a bunch of these connected, you will see that they all start folding over themselves and will need to take up the same space which, is not possible because the electrons will repel. So you get distortions and all kinds of stuff to push them away and then it's all complicated by a bunch of weak forces. The primary thing giving structure is this long string of covalent bonds. Also, "forces in the lipid layer surrounding cells" are not proteins
Razied20

But perhaps the bigger reason is that I find SIA intuitively extremely obvious. It’s just what you get when you apply Bayesian reasoning to the fact that you exist.

Correct, except for the fact that you're failing to consider the possibility that you might not exist at all...

My entire uncertainty in anthropic reasoning is bound up in the degree to which an "observer" is at all a coherent concept.

2Ape in the coat
Actually, it's the other way around. SIA always assumes that one may not have existed at all. This is the source of the Bayesian update and this itself may be a problem.  Basically, it requires assuming that all existing humans were randomly sampled from a finite set of immaterial souls - a pretty extraordinary claim about the way our universe works, without any evidence to support it. 
Razied1712

And my guess is that is how Hamas see and bill themselves.

And your guess would be completely, hopelessly wrong. There is an actual document called "The Covenant of Hamas" written in 1988 and updated in 2017, which you can read here, it starts with

Praise be to Allah, the Lord of all worlds. May the peace and blessings of Allah be upon Muhammad, the Master of Messengers and the Leader of the mujahidin, and upon his household and all his companions.

... so, uh, not a good start for the "not religious" thing. It continues:

1. The Islamic Resistance Movement “Ham

... (read more)
Razied31

It is important that Gazans won't feel like their culture is being erased.

 

A new education curriculum is developed which fuses western education, progressive values and Muslim tradition while discouraging political violence.

These two things are incompatible. Their culture is the entire problem. To get a sense of the sheer vastness of the gap, consider the fact that Arabs read on average 6 pages per year. It would take a superintelligence to somehow convince the palestinians to embrace western thought and values while not feeling like their culture is ... (read more)

1mrfox
From your link: "A general lack of educational opportunities in poor Arab countries can also add to these facts. Research for the Arab League region estimates that about 100 million people ? almost one in three - struggle to read and write." I'm at best unsure how much cultute is the Problem. Even if "not reading" was a cutural pillar for arabs, that can be changed without subverting everything else.
Razied20

Oh, true! I was going to reply that since probability is just a function of a physical system, and the physical system is continuous, then probability is continuous... but if you change an integer variable in C from 35 to 5343 or whatever, there's no real sense in which the variable goes through all intermediate values, even if the laws of physics are continuous.

Razied126

If he's ever attended an event which started out with less than a 28% chance of orgy, which then went on to have an orgy, then that statement is false by the Intermediate Value Theorem, since there would have been an instant in time where the probability of the event crossed 28%.

gbear6051510

That's only true if the probability is a continuous function - perhaps the probability instantaneously went from below 28% to above 28%.

Razied2314

The most basic rationalist precept is to not forcibly impose your values onto another mind.

What? How does that make any sense at all? The most basic precept of rationality is to take actions which achieve future world states that rank highly under your preference ordering. Being less wrong, more right, being bayesian, saving the world, not imposing your values on others, etc. are all deductions that follow from that most basic principle: Act and Think Such That You Win.

6Adam Kaufman
I find it useful to distinguish between epistemic and instrumental rationality. You're talking about instrumental rationality – and it could be instrumentally useful to convince someone of your beliefs, to teach them to think clearly, or to actively mislead them.  Epistemic rationality, on the other hand, means trying to have true beliefs, and in this case it's better to teach someone to fish than to force them to accept your fish.
Razied82

Wait, do lesswrongers not know about semaglutide and tirzepatide yet? Why would anyone do something as extreme as bariatric surgery when tirzepatide patients lose pretty much the same amount of weight after a year as with the surgery?

3lc
Two things: * Semaglutide only usually allows you to lose up to 20% of your excess bodyweight. For people who are 100, 200 pounds overweight, it will generally not be enough. * Sometimes semaglutide doesn't work at all. * There's less information about it. I've heard lots of anecdotal reports of semaglutide inducing tolerance and losing its effects over time. If this means spending 2k/month forever to maintain the same weight then that's pretty bad.
Razied51

But if you are right that you only respond to a limited set of story types, do you therefore aspire to opening yourself to different ones in future, or is your conclusion that you just want to stick to films with 'man becomes strong' character arcs?

Not especially, for the same reason that I don't plan on starting to eat 90% dark chocolate to learn to like it, even if other people like it (and I can even appreciate that it has a few health benefits). I certainly am not saying that only movies that appeal to me be made, I'm happy that Barbie exists and that ... (read more)

1Rosencrantz
Part of the point is that the standards we desire for ourselves may be contradictory and thus unachievable (e.g. Barbie's physical proportions). So it's not necessarily 'lower your standards', but 'seek more coherent, balanced standards'.  I also think you can enjoy the message-for-the-character without needing it for you but anyway, I get where you're personally coming from and appreciate your level of frankness about it! 
Razied110

I watched Barbie and absolutely hated it. Though it did provide some value to me after I spent some time thinking about why precisely I hated it. Barbie really showed me the difference between the archetypal story that appeals to males and the female equivalent, and how much just hitting that archetypal story is enough to make a movie enjoyable for either men or women.

The plot of the basic male-appealing story is "Man is weak. Man works hard with clear goal. Man becomes strong". I think men feel this basic archetypal story much more strongly than women, so... (read more)

7Rosencrantz
I suppose you may have correctly analysed your reason for not liking the movie. But if you are right that you only respond to a limited set of story types, do you therefore aspire to opening yourself to different ones in future, or is your conclusion that you just want to stick to films with 'man becomes strong' character arcs? I personally loved Barbie (man here!), and think it was hilarious, charming and very adroit politically. I also think that much of the moral messaging is pretty universal – Greta Gerwig obviously thinks so: when she says: "I think equally men have held themselves to just outrageous standards that no one can meet. And they have their own set of contradictions where they’re walking a tightrope. I think that’s something that’s universal." Is it possible that that message does strike some kind of chord with you but you don't want to hear it? (I guess I find 'absolutely hated' to be incredibly strong language for a film made with obvious skill and wit and that I think has no right to be as good as it is.)
Razied20

I'm fairly sure that there's architectures where each layer is a linear function of the concatenated activations of all previous layers, though I can't seem to find it right now. If you add possible sparsity to that, then I think you get a fully general DAG.

Razied40

Their paper for the sample preparation (here) has a trademark sign next to the "LK-99" name, which suggests they've trademarked it... strongly suggesting that the authors actually believe in their stuff.

3jeff8765
It also seems a patent was filed for this material in 2021 and was granted earlier this year prior to publication.
Razied149

There are a whole bunch of ways that trying to optimise for unpredictability is not a good idea:

  1. Most often technical discussions are not just exposition dumps, they're a part of the creative process itself. Me telling you an idea is an essential part of my coming up with the idea. I essentially don't know where I'm going before I get there, so it's impossible for me to optimise for unpredictability on your end.
  2. This ignores a whoooole bunch of status-effects and other goals of human conversation. The point of conversation is not solely to transmit informati
... (read more)
6dkl9
1 and 2 are absolutely correct, but for specific subsets. Outside such subsets, this optimisation still applies. 3 is correct sometimes as reversed advice. I see your point in 3 often (usually implicit). My post reverses that in response to it sometimes going too far. It seems I went too far. Hence the expanded original:
Razied40

I think you might want to look at the litterature on "sparse neural networks", which is the right search term for what you mean here.

2Richard_Kennaway
I don't think "sparse neural networks" fit the bill. All the references I've turned up for the phrase talk about the usual sort of what I've been calling layered NNs, but where most of the parameters are zero. This leaves intact the layer structure. To express more precisely the sort of connectivity I'm talking about, for any NN, construct the following directed graph. There is one node for every neuron, and an arc from each neuron A to each neuron B whose output depends directly on an output value of A. For the NNs as described in e.g. Andrej Karpathy's lectures (which I'm currently going through), this graph is a DAG. Furthermore, it is a DAG having the property of layeredness, which I define thus: A DAG is layered if every node A can be assigned an integer label L(A), such that for every edge from A to B, L(B) = L(A)+1. A layer is the set of all the nodes having a given label. The sparse NNs I've found in the literature are all layered. A "full" (i.e. not sparse) NN would also satisfy the converse of the above definition, i.e. L(B) = L(A)+1 would imply an edge from A to B. The simplest example of a non-layered DAG is one with three nodes A, B, and C, with edges from A to B, A to C, and B to C. If you tried to structure this into layers, you would either find an edge between two nodes in the same layer, or an edge that skips a layer. To cover non-DAG NNs also, I'd call one layered if in the above definition, "L(B) = L(A)+1" is replaced by "L(B) = L(A) ± 1". (ETA: This is equivalent to the graph being bipartite: the nodes can be divided into two sets such that every edge goes from a node in one set to a node in the other.) It could be called approximately layered if most edges satisfy the condition. Are there any not-even-approximately-layered NNs in the literature?
Razied4-6

I'm really confused about how anybody thinks they can "license" these models. They're obviously not works of authorship.

I'm confused why you're confused, if I write a computer program that generates an artifact that is useful to other people, obviously the artifact should be considered a part of the program itself, and therefore subject to licensing just like the generating program. If I write a program to procedurally generate interesting minecraft maps, should I not be able to license the maps, just because there's one extra step of authorship between me and them?

4jbash
If it generates them totally at random, then no. They have no author. But even in that case, if you do it in a traditional way you will at least have personally made more decisions about what the output looks like than somebody who trains a model. The whole point of deep learning is that you don't make decisions about the weights themselves. There's no "I'll put a 4 here" step.
Razied60

The word "curiosity" has a fairly well-defined meaning in the Reinforcement Learning literature (see for instance this paper). There are vast numbers of papers that try to come up with ways to give an agent intrinsic rewards that map onto the human understanding of "curiosity", and almost all of them are some form of "go towards states you haven't seen before". The predictable consequence of prioritising states you haven't seen before is that you will want to change the state of the universe very very quickly.

5mishka
Novelty is important. Going towards states you have not seen before is important. This will be a part of the new system, that's for sure. But this team is under no obligation to follow whatever current consensus might be (if there is a consensus). Whatever is the state of the field, it can't claim a monopoly on how words "curiosity" or "novelty" are interpreted, what are the good ways to maximize them... How one constrains going through a subset of all those novel states by aesthetics, by the need to take time and enjoy ("exploit") those new states, and by safety considerations (so, by predicting whether the novel state will be useful and not detrimental)... All this will be on the table... Some of the people on this team are known for making radical breakthroughs in machine learning and for founding new subfields in machine learning. They are not going to blindly copy the approaches from the existing literature (although they will take existing literature into account).
Razied20

Not too sure about the downvotes either, but I'm curious how the last sentence misses the point? Are you aware of a formal definition of "interesting" or "curiosity" that isn't based on novelty-seeking? 

2Tapatakt
I think for all definitions of "curiosity" that make sense (that aren't like "we just use this word to refer to something completely unrelated to what people usually understand by it") maximally curious AI kills us, so it doesn't matter how curiosity is defined in RL literature.
Razied1312

According to reports xAI will seek to create a "maximally curious" AI, and this also seems to be the main new idea how to solve safety, with Musk explaining: "If it tried to understand the true nature of the universe, that's actually the best thing that I can come up with from an AI safety standpoint," ... "I think it is going to be pro-humanity from the standpoint that humanity is just much more interesting than not-humanity."

Is Musk just way less intelligent than I thought? He still seems to have no clue at all about the actual safety problem. Anyone thi... (read more)

2Tapatakt
I think last sentence kinda misses the point, but in general I agree. Why all this downvotes?
Razied7-2

A theory of the popularity of anime.

Much like there have been ten thousand reskins of Harry Potter I’ve been waiting for more central examples of English-language cultural products to take that story archetype and just run with it. There is clearly a demand.

Well then Rejoice! The entire genre of Progression Fantasy is what you desire, and you need only browse the Best Of RoyalRoad to see lots of english-language stories that scratch that particular itch. In fact, I find these english stories immensely superior to anything in anime or manga. 

A particul... (read more)

Razied207

Overall, a headline that seems counterproductive and needlessly divisive.

Probably the understatement of the decade, this article is literally an "order" from Official Authority to stop talking about what I believe is literally the most important thing in the world. I guess this is not literally the headline that would maximally make me lose respect for Nature... but it's pretty close. 

This article is a pure appeal to authority. It contains no arguments at all, it only exists as a social signal that Respectable Scientists should steer away from talk of... (read more)

3Noosphere89
Yep, that's the biggest issue I have with my own side of the debate on AI risk, in that quite often, they don't even try to state why it isn't a risk, and instead appeal to social authority, and while social authority is evidence, it's too easy to filter that evidence a lot to be useful. To be frank, I don't blame a lot of the AI risk people for not being convinced that we aren't doomed, even though reality doesn't grade on a curve, the unsoundness of the current arguments against doom don't help, and it is in fact bad that my side keeps doing this.
Razied20

That's not a math or physics paper, and it includes a bit more "handholding" in the form of an explicit database than would really make me update. The style of scientific papers is obviously very easy to copy for current LLMs, what I'm trying to get at is that if LLMs can start to make genuinely novel contributions at a slightly below-human level and learn from the mediocre article they write, pure volume of papers can make up for quality.

Razied106
  • "This has been killing people!"
  • "Yes, but it might kill all people!"
  • "Yes, but it's killing people!"
  • "Of course, sure, whatever, it's killing people, but it might kill all people!"


But this isn't the actual back-and-forth, the third point should be "no it won't, you're distracting from the people currently being killed!". This is all a game to subtly beg the question. If AI is an existential threat, all current mundane threats like misinformation, job loss, AI bias, etc. are rounding errors to the total harm, the only situation where you'd talk about them is i... (read more)

4Vladimir_Nesov
It's possible to consider relatively irrelevant things, such as everything in ordinary human experience, even when there is an apocalypse on the horizon. The implied contextualizing norm asks for inability to consider them, or at least increases the cost.
5the gears to ascension
Sure, I agree, the asteroid is going to kill us all. But it would be courteous to acknowledge that it's going to hit a poor area first, and they'll die a few minutes earlier. Also, uh, all of us are going to die, I think that's the core thing! we should save the poor area, and also all the other areas!
Razied20

I don't think we need superhuman capability here for stuff to get crazy, pure volume of papers could substitute for that. If you can write a mediocre but logically correct paper with $50 of compute instead of with $10k of graduate student salary, that accelerates the pace of progress by a factor of 200, which seems enough for me to enable a whole bunch of other advances which will feed into AI research and make the models even better.

2JBlack
So you're now strongly expecting to die in less than 6 months? (Assuming that the tweet is not completely false)
Razied20

If we get to that point of AI capabilities, we will likely be able to make 50 years of scientific progress in a matter of months for domains which are not too constrained by physical experimentation (just run more compute for LLMs), and I'd expect AI safety to be one of those. So either we die quickly thereafter, or we've solved AI safety. Getting LLMs to do scientific progress basically telescopes the future.

2Donald Hobson
Possible alternatives.  1. AI can make papers as good as the average scientist, but wow is it slow. Total AI paper output is less than total average scientist output, even with all available compute thrown at it.  2. AI can write papers as good as the Average scientist. But a lot of progress is driven by the most insightful 1% of scientists. So we get ever more mediocre incremental papers without any revolutionary new paradigms.  3. AI can make papers as good as the average scientist. For AI safety reasons, this AI is kept rather locked down and not run much. Any results are not trusted in the slightest.  4. AI can make papers as good as the average scientist. Most of the peer review and journal process is also AI automated. This leads to a goodhearting loop. All the big players are trying to get papers "published" by the million. Almost none of these papers will ever be read by a human. There may be good AI safety ideas somewhere in that giant pile of research. But good luck finding them in the massive piles of superficially plausible rubbish. If making a good paper becomes 100x easier, but making a rubbish paper becomes a million times easier, and telling the difference becomes 2x easier, the whole system get's buried in mountains of junk papers.  5. AI's can do and have done AI safety research. There are now some rather long and technical books that present all the answers. Capabilities is now a question of scaling up chip production. (Which has slow engineering bottlenecks) We aren't safe yet. When someone has enough chips, will they use that AI safety book or ignore it? What goal will they align their AI to?
3JBlack
Are you assuming that there will be a sudden jump in AI scientific research capability from subhuman to strongly superhuman? It is one possibility, sure. Another is that the first AIs capable of writing research papers won't be superhumanly good at it, and won't advance research very far or even in a useful direction. It seems to me quite likely that this state of affairs will persist for at least six months. Do you give the latter scenario less than 0.01 probability? That seems extremely confident to me.
Razied20

Fair point, "non-trivial" is too subjective, the intuition that I meant to convey was that if we get to the point where LLMs can do the sort of pure-thinking research in math and physics at a level where the papers build on top of one another in a coherent way, then I'd expect us to be close to the end. 

Said another way, if theoretical physicists and mathematicians get automated, then we ought to be fairly close to the end. If in addition to that the physical research itself gets automated, such that LLMs write their own code to do experiments (or run the robotic arms that manipulate real stuff) and publish the results, then we're *really* close to the end. 

Razied33

If the question is ‘what’s one experiment that would drop your p(doom) to under 1%?’ then I can’t think of such an experiment that would provide that many bits of data, without also being one where getting the good news seems absurd or being super dangerous.

Not quite an experiment, but to give an explicit test: if we get to the point where an AI can write non-trivial scientific papers in physics and math, and we then aren't all dead within 6 months, I'll be convinced that p(doom) < 0.01, and that something was very deeply wrong with my model of the world.

3JBlack
If that evidence would update you that far, then your space of doom hypotheses seems far too narrow. There is so much that we don't know about strong AI. A failure to be rapidly killed only seems to rule out some of the highest-risk hypotheses, while still leaving plenty of hypotheses in which doom is still highly likely but slower.
1cwillu
“Non-trivial” is a pretty soft word to include in this sort of prediction, in my opinion. I think I'd disagree if you had said “purely AI-written paper resolves an open millennium prize problem”, but as written I'm saying to myself “hrm, I don't know how to engage with this in a way that will actually pin down the prediction”. I think it's well enough established that long form internally coherent content is within the capabilities of a sufficiently large language model.  I think the bottleneck on it being scary (or rather, it being not long before The End) is the LLM being responsible for the inputs to the research.
Load More