All of Rebecca's Comments + Replies

Rebecca20

He may have bought it originally for protection?

Rebecca10

What are the illegal things that would be needed?

5AnthonyC
The specifics of what I'm thinking of vary a lot between jurisdictions, and some of them aren't necessarily strictly illegal so much as "Relevant authorities might cause you a lot of problems even if you haven't broken any laws." But roughly speaking, I'm thinking about the umbrella of everything that kids are no longer allowed to do that increase demands on parents compared to past generations, plus all the rules and policies that collectively make childcare very expensive, and make you need to live in an expensive town to have good public schools. Those are the first categories that come to mind for me.
Rebecca40

How do we know money wouldn’t work? Surely most of the reasons people don’t have more children could be solved with more resources (eg surrogacy, childcare, private school fees, a larger house).

4AnthonyC
I think it's more a matter of Not Enough Dakka plus making it illegal to do those things in what should be reasonable ways. I agree there are economic (and regulatory) interventions that could make an enormous difference, but for various reasons I don't think any government is currently willing and able to implement them at scale. A crisis needs to be a lot more acute to motivate that scale of change.
Rebecca128

Being homeless sucks, it’s pretty legitimate to want to avoid that

-1Valentin2026
I agree, they have a really bad life, but Eliezer seems to talk here about those who work 60 hours/week to ensure their kids will go to a good school. Slightly different problem.  And on homeless people, there are different cases. In some UBI indeed will help. But, unfortunately, in many cases the person has mental health problems or addiction, and simply giving them money may not help. 
Rebecca58

I’ve found use of the term catastrophe/catastrophic in discussions of SB 1047 makes it harder for me to think about the issue. The scale of the harms captured by SB 1047 has a much much lower floor than what EAs/AIS people usually term catastrophic risk, like $0.5bn+ vs $100bn+. My view on the necessity of pre-harm enforcement, to take the lens of the Anthropic letter, is very different in each case. Similarly, while the Anthropic letter talks about the the bill as focused on catastrophic risk, it also talks about “skeptics of catastrophic risk” - surely ... (read more)

Rebecca30

Perhaps when you share the post with friends you could quote some of the bits focused on progressive concerns?

Rebecca20

a dramatic hardware shift like that is likely going to mean a significant portion of progress up until that shift in topics like interpretability and alignment may be going out the window.

Why is this the case?

6kromem
It's still early to tell, as the specific characteristics of a photonic or optoelectronic neural network are still formulating in the developing literature. For example, in my favorite work of the year so far, the researchers found they could use sound waves to reconfigure an optical neural network as the sound waves effectively preserved a memory of previous photon states as they propagated: https://www.nature.com/articles/s41467-024-47053-6 In particular, this approach is a big step forward for bidirectional ONN, which addresses what I think is the biggest current flaw in modern transformers - their unidirectionality. I discussed this more in a collection of thoughts on directionality impact on data here: https://www.lesswrong.com/posts/bmsmiYhTm7QJHa2oF/looking-beyond-everett-in-multiversal-views-of-llms If you have bidirectionality where previously you didn't, it's not a reach to expect that the way in which data might encode in the network, as well as how the vector space is represented, might not be the same. And thus, that mechanistic interpretability gains may get a bit of a reset. And this is just one of many possible ways it may change by the time the tech finalizes. The field of photonics, particularly for neural networks, is really coming along nicely. There may yet be future advances (I think this is very likely given the pace to date), and advantages the medium offers that electronics haven't. It's hard to predict exactly what's going to happen when two different fields which have each had unexpected and significant gains over the past 5 years collide, but it's generally safe to say that it will at very least result in other unexpected things too.
Rebecca10

The weights could be stolen as soon as the model is trained though

Rebecca119

unless the nondisparagement provision was mutual

This could be true for most cases though

Rebecca10

That seems like a valuable argument. It might be worth updating the wording under premise 2 to clarifying this? To me it reads as saying that the configuration, rather than the aim, of OpenAI was the major red flag.

Rebecca92

My impression is that post-board drama, they’ve de-emphasised the non-profit messaging. Also in a more recent interview Sam said basically ‘well I guess it turns out the board can’t fire me’ and that in the long term there should be democratic governance of the company. So I don’t think it’s true that #8-10 are (still) being pushed simultaneously with the others.

I also haven’t seen anything that struck me as communicating #3 or #11, though I agree it would be in OpenAI’s interest to say those things. Can you say more about where you are seeing that?

Rebecca10

So the argument is that Open Phil should only give large sums of money to (democratic) governments? That seems too overpowered for the OpenAI case.

Rebecca20

In that case OP’s argument would be saying that donors shouldn’t give large sums of money to any sort of group of people, which is a much bolder claim

(I'm the OP)

I'm not trying to say "it's bad to give large sums of money to any group because humans have a tendency to to seek power." 

I'm saying "you should be exceptionally cautious about giving large sums of money to a group of humans with the stated goal of constructing an AGI."

You need to weight any reassurances they give you against two observations:

  1. The commonly observed pattern of individual humans or organisations seeking power (and/or wealth) at the expense of the wider community. 
  2. The strong likelihood that there will be an opportunity f
... (read more)
Rebecca-10

I was more focused on the ‘company’ part. To my knowledge there is no such thing as a non-profit company?

Rebecca236

Noting that while Sam describes the provision as being about “about potential equity cancellation”, the actual wording says ‘shall be cancelled’ not ‘may be cancelled’, as per this tweet from Kelsey Piper: https://x.com/KelseyTuoc/status/1791584341669396560

Rebecca30

Instances in history in which private companies (or any individual humans) have intentionally turned down huge profits and power are the exception, not the rule.

OpenAI wasn’t a private company (ie for-profit) at the time of the OP grant though.

1Stephen Fowler
This does not feel super cruxy as the the power incentive still remains. 
2dr_s
Aren't these different things? Private yes, for profit no. It was private because it's not like it was run by the US government.
Rebecca10

Is that not what Altman is referring to when he talks about vested equity? My understanding was employees had no other form of equity besides PPUs, in which case he’s talking non-misleadingly about the non-narrow case of vested PPUs, ie the thing people were alarmed about, right?

2James Payor
It may be that talking about "vested equity" is avoiding some lie that would occur if he made the same claim about the PPUs. If he did mean to include the PPUs as "vested equity" presumably he or a spokesperson could clarify, but I somehow doubt they will.
2Linch
OpenAI has something called PPUs ("Profit Participation Units") which in theory is supposed to act like RSUs albeit with a capped profit and no voting rights, but in practice is entirely a new legal invention and we don't really know how it works.
Rebecca80

Did OpenAI have the for-profit element at that time?

8Buck
No. E.g. see here
Rebecca1-5

Sure, but you weren’t providing reasons to not believe the argument, or reasons why your interpretation is at least as implausible

Rebecca10

Zvi has already addressed this - arguing that if (D) was equivalent to ‘has a similar cost to >=$500m in harm’, then there would be no need for (B) and (C) detailing specific harms, you could just have a version of (D) that mentions the $500m, indicating that that’s not a sufficient condition. I find that fairly persuasive, though it would be good to hear a lawyer’s perspective

1Jiro
"This very clearly does not" apply to X and "I have an argument that it doesn't apply to X" are not the same thing. (And it wouldn't be hard for a court to make some excuse like "these specific harms have to be $500m, and other harms 'of similar severity' means either worse things with less than $500m damage or less bad things with more than $500m damage". That would explain the need to detail specific harms while putting no practical restriction on what the law covers, since the court can claim that anything is a worse harm. Always assume that laws of this type are interpreted by an autistic, malicious, genie.)
5localdeity
I'd guess modesty / non-self-promotion norms, and/or avoiding the appearance of a conflict of interest.  "Hey, everyone, listen to this important and wise thing Bob said!  (BTW Bob says that I'm great, and he is so right.)"

I think calling this a strategic meaning is not that helpful. I would say the issue is that “isolated” is underspecified. It’s not like there was a fully fleshed out account that was then backtracked on, it’s more like: what was the isolation? were they isolated from literally everyone who wasn’t Kat, Emerson or Drew, or were they isolated /pushed to isolate more than is healthy from people they didn’t need to have their ‘career face’ on for? We now know the latter was meant, but either was plausible.

0Jiro
I gave a quote from Ben's post above. She specified families, romantic partners, and locals. Maybe this doesn't count as fully fleshed, but the new interpretation certainly backtracks on these.
2tailcalled
It's unhelpful when it comes to just explaining the meaning of "isolation" as that is better communicated by something like Ozy's description, or in various other ways. However, I think at least in theory it can be helpful for understanding where to focus one's attention in the case of community drama like this, including for other questions than just the "isolation" one. Maybe not in practice because I haven't gone in depth for explaining what I mean by "strategic" though.

This quote doesn’t say anything about the board member/s being people who are researching AI safety though - it’s Nathan’s friends who are in AI safety research not the board members.

I agree that based on this quote, it could have very well been just a subset of the board. But I believe Nathan’s wife works for CEA (and he’s previously MCed an EAG), and Tasha is (or was?) on the board of EVF US, and so idk, if it’s Tasha he spoke to and the “multiple people” was just her and Helen, I would have expected a rather different description of events/vibe. E.g. so... (read more)

I think you might be misremembering the podcast? Nathan said that he was assured that the board as a whole was serious about safety, but I don’t remember the specific board member being recommended as someone researching AI safety (or otherwise more pro safety than the rest of the board). I went back through the transcript to check and couldn’t find any reference to what you’ve said.

“ And ultimately, in the end, basically everybody said, “What you should do is go talk to somebody on the OpenAI board. Don’t blow it up. You don’t need to go outside of the ch... (read more)

2gwern
I was not referring to the podcast (which I haven't actually read yet because from the intro it seems wildly out of date and from a long time ago) but to Labenz's original Twitter thread turned into a Substack post. I think you misinterpret what he is saying in that transcript because it is loose and extemporaneous "they're committed" could just as easily refer to "are serious people on the board" who have "been chosen" for that (implying that there are other members of the board not chosen for that); and that is what he says in the written down post:

They weren’t the only non employee board members though - that’s what I meant by the part about not being concerned about safety, that I took it to rule out both Toner and McCauley.

(Although it for some other reason you were only looking at Toner and McCauley, then no, I would say the person going around speaking to OAI employees is_less_ likely to be out of the loop on GPT-4’s capabilities)

8gwern
The other ones are unlikely. Shivon Zilis & Reid Hoffman had left by this point; Will Hurd might or might not still be on the board at this point but wouldn't be described nor recommended by Labenz's acquaintance as researching AI safety, as that does not describe Hurd or D'Angelo; Brockman, Altman, and Sutskever are right out (Sutskever researches AI safety but Superalignment was a year away); by process of elimination, over 2023, the only board members he could have been plausibly contacting would be Toner and McCauley, and while Toner weakly made more sense before, now McCauley does. (The description of them not having used the model unfortunately does not distinguish either one - none of the writings connected to them sound like they have all that much hands-on experience and would be eagerly prompt-engineering away at GPT-4-base the moment they got access. And I agree that this is a big mistake, but it is, even more unfortunately, and extremely common one - I remain shocked that Altman had apparently never actually used GPT-3 before he basically bet the company on it. There is a widespread attitude, even among those bullish about the economics, that GPT-3 or GPT-4 are just 'tools', which are mere 'stochastic parrots', and have no puzzling internal dynamics or complexities. I have been criticizing this from the start, but the problem is, 'sampling can show the presence of knowledge and not the absence', so if you don't think there's anything interesting there, your prompts are a mirror which reflect only your low expectations; and the safety tuning makes it worse by hiding most of the agency & anomalies, often in ways that look like good things. For example, the rhyming poetry ought to alarm everyone who sees it, because of what it implies underneath - but it doesn't. This is why descriptions of Sydney or GPT-4-base are helpful: they are warning shots from the shoggoth behind the friendly tool-AI ChatGPT UI mask.)

Why do you think McCauley is likely to be the board member Labenz spoke to? I had inferred that it was someone not particularly concerned about safety given that Labenz reported them saying they could be easily request access to the model if they’d wanted to (and hadn’t). I took the point of the anecdote to be ‘here was a board member not concerned about safety’.

3gwern
Because there is not currently any evidence that Toner was going around talking to a bunch of people, whereas this says McCauley was doing so. If I have to guess "did Labenz talk to the person who was talking to a bunch of people in OA, or did he talk to the person who was as far as I know not talking to a bunch of people in OA?", I am going to guess the former.

Could you link to some examples of “ OAers being angry on Twitter today, and using profanity & bluster and having oddly strong opinions about how it is important to refer to roon as @tszzl and never as @roon”? I don’t have a twitter account so can’t search myself

In the latter case it is the 3rd party driving the article, airing the accusations in a public forum, and deciding how they are framed, rather than Avery.

If, without the 3rd party, Avery would have written an essentially identical article then the differences aren’t relevant. But in the more likely case where Avery is properly a “source” for an article for which the 3rd party is counterfactually responsible, then the 3rd party also bears more responsibility for the effect of the article on Pat’s reputation etc. Fortunately, the 3rd party, not being anonymous, can be practically judged for their choices in writing the article, in the final accounting.

And these situations are different again from someone posting under their real name but referring to sources who agreed to be sources on the condition of anonymity

2jefftk
Yes, but I'm not sure in an important way? What do you see as the relevant differences between a case where Avery as 'Alex' writes a post about Pat, and an otherwise similar case where Avery acts as a source for a third party writing about Pat who agrees to use 'Alex' for them?
Rebecca128

I strongly think that much or even most of the commentary could have been discarded in favour of more evidence

I’m not sure, I haven’t been using Manifold for very long, sorry!

Why would attraction ruin the friendship?

You can also rate how someone resolved their market afterwards, which I assume does something

7River
I did not know this. How long has this been around?  Still strikes me as a really bad idea to ignore the norms that actual financial markets have developed over centuries of experience, but I am curious if this will actually solve the problem of judges biased by having a position in their own markets.

Ah I see. I took ‘recreational’, given NL’s context, to mean something like ‘ADHD medication taken recreationally’.

Thanks for checking! Have now figured out the issue, the thing I described was happening when Google docs opened in safari (which I knew), but I’ve now gotten it to open in the app proper.

Yeah Ben should have said illicit not illegal, because they are illegal to bring across the border except if you have a valid prescription, even if the place you purchased them didn’t require a prescription. But I wouldn’t consider it an unambiguous falsehood, like the following is mostly a sliding scale of frustrating ambiguity:

  1. ‘asked Alice to illegally bring Schedule II medication into the country’ [edit: entirely correct according to NL’s stating of the facts]
  2. ‘asked Alice to illegally bring Schedule II drugs into the country’ [some intermediate versi
... (read more)
3habryka
Just for the record, I think the original claim continues to be true, and in addition to prescription drugs, Alice was also asked to bring over recreational drugs that are more clearly illegal. I do think that if the facts were such that it was just ADHD meds, the way the claim was presented (especially given that it explicitly mentioned 'recreational drugs') would have been misleading.  However, I don't think it was just ADHD meds (this is somewhat hard to prove since as far as I know basically all requests for things like this were made in voice, but we might be able to find something).

FYI, when I click on some proportion (possibly 100%?) of these links to the Google doc (including the links in your comment here) it just takes me to the very start of Google doc, the beginning of the contents section, and I can’t always figure out which section to click on. Possibly a mobile issue with Google docs, but thought I should let you know 🙂

2KatWoods
Thanks for letting me know! Strange. It shouldn't be doing that. Usually if you wait a couple of seconds, it'll jump to the right section. It's working on both my mobile and laptop. If you try waiting a couple seconds and that doesn't work, let me know. Maybe DM me and then we can troubleshoot, then we can post the solution up when we figure it out. 

Does anyone know why it it Francesca vs Harvard and not Gino v Harvard?

2Linch
My guess is that it's because "Francesca" sounds more sympathetic as a name.

What is “a good track record with respect to aging processes” referring to?

3jmh
Very late response but that was just saying, from my cursory knowledge, that everyone seems to agree that naked mole-rats do not seem to display the same aging related problems other species do.
Rebecca2510

The assertion is that Sam sent the email reprimanding Helen to others at OpenAI, not to Helen herself, which is a fundamentally different move.

I can’t conceive of a situation in which the CEO of a non-profit trying to turn the other employees against the people responsible for that non-profit (ie the board) would be business-as-usual.

Rebecca6953

Most of this is interesting, and useful, speculation, but it reads as a reporting of facts…

I found this quite hard to parse fyi

Rebecca112

Sorry yeah I could have explained what I meant further. The way I see it:

‘X is the most effective way that I know of’ = X tops your ranking of the different ways, but could still be below a minimum threshold (e.g. X doesn’t have to even properly work, it could just be less ineffective than all the rest). So one could imagine someone saying “X is the most effective of all the options I found and it still doesn’t actually do the job!”

‘X is an effective way’ = ‘X works, and it works above a certain threshold’.

‘X is Y done right’ = ‘X works and is basically th... (read more)

This reads as some sort of confused motte and bailey. Are RSPs “an effective way” or “the most effective way… [you] know of”? These are different things, with each being stronger/weaker in different ways. Regardless, the title could still be made much more accurate to your beliefs, e.g. ~’RSPs are our (current) best bet on a pause’. ‘An effective way’ is definitely not “i.e … done right”, but “the most effective way… that I know of” is also not.

4evhub
I disagree? I think the plain English meaning of the title "RSPs are pauses done right" is precisely "RSPs are the right way to do pauses (that I know of)" which is exactly what I think and exactly what I am defending here. I honestly have no idea what else that title would mean.
Rebecca3-1

I understood the original comment to be making essentially the same point you’re making - that lying has a bad track record, where ‘lying has a bad track record of causing mistrust’ is a case of this. In what way do you see them as distinct reasons?

6Adam Zerner
I see them as distinct because what I'm saying is that lying generally tends to lead to bad outcomes (for both the liar and society at large) whereas mistrust specifically is just one component of the bad outcomes. Other components that come to my mind: * People don't end up with accurate information. * Expectations that people will cooperate (different from "tell you the truth") go down. * Expectations that people will do things because they are virtuous go down. But a big thing here is that it's difficult to know why exactly it will lead to bad outcomes. The gears are hard to model. However, I think there's solid evidence that it leads to bad outcomes.
Rebecca115

I understood NinetyThree to be talking about vegans lying about issues of health (as Elizabeth was also focusing on), not about the facts of animal suffering. If you agree with the arguments on the animal cruelty side and your uncertainty is focused on the health effects on you of a vegan diet vs your current one (which you have 1st hand data on), it doesn’t really matter what the meat industry is saying as that wasn’t a factor in the first place

3tailcalled
Maybe. I pattern-matched it this way because I had previously been discussing psychological sex differences with Ninety-Three on discord, where he adopted the HBD views on them due to a perception that psychologists were biased, but he wasn't interested in making arguments or in me doing followup studies to test it. So I assumed a similar thing was going on here with respect to eating animals.

Could you clarify who you are defining as carnists?

5tailcalled
I tend to think of ideology as a continuum, rather than a strict binary. Like people tend to have varying degrees of belief and trust in the sides of a conflict, and various unique factors influencing their views, and this leads to a lot of shades of nuance that can't really be captured with a binary carnist/not-carnist definition. But I think there are still some correlated beliefs where you could e.g. take their first principal component as an operationalization of carnism. Some beliefs that might go into this, many of which I have encountered from carnists: * "People should be allowed to freely choose whether they want to eat factory-farmed meat or not." * "Animals cannot suffer in any way that matters." * "One should take an evolutionary perspective and realize that factory farming is actually good for animals. After all, if not for humans putting a lot of effort into farming them, they wouldn't even exist at their current population levels." * "People who do enough good things out of their own charity deserve to eat animals without concerning themselves with the moral implications." * "People who design packaging for animal products ought to make it look aesthetically pleasing and comfortable." * "It is offensive and unreasonable for people to claim that meat-eating is a horribly harmful habit." * "Animals are made to be used by humans." * "Consuming animal products like meat or milk is healthier than being strictly vegan." One could make a defense of some of the statements. For instance Elizabeth has made a to-me convincing defense of the last statement. I don't think this is a bug in the definition of carnism, it just shows that some carnist beliefs can be good and true. One ought to be able to admit that ideology is real and matters while also being able to recognize that it's not a black-and-white issue.

They’re talking about technical research orgs/labs, not ancillary orgs/projects

Load More