All of Larks's Comments + Replies

Larks30

Thanks very much! Yeah, I agree political will seems like a big issue. But I also hear people saying that they don't know what to push for, so I wanted to try to offer a concrete example of a system that wasn't as destructive to any constituency's interests as e.g. a total pause.

Larks51

Thanks very much for your feedback, though I confess I'm not entirely sure where to go with it. My interpretation is you have basically two concerns:

  1. This policy doesn't really directly regulate algorithmic progress, e.g. if it happened on smaller amounts of compute.
  2. Algorithmic theft/leakage is easy.

The first one is true, as I alluded in the problems section. Part of my perspective here is coming from a place of skepticism about regulatory competence - I basically believe we can get regulators to control total compute usage, and to evaluate specific models ... (read more)

2Nathan Helm-Burger
I agree that I have no faith in current governments to implement and enforce policies that are more complex than things on the order of governance compute and chip export controls. I think the conclusion this points towards is that we need new forms of governance. Not to replace existing governments, but to complement them. Voluntary mutual inspection contracts with privacy-respecting technology using AI inspectors. Something of that sort. Here's some recent evidence of compute thresholds not being reliable: https://novasky-ai.github.io/posts/sky-t1/ Here's another self-link to some of my thoughts on this: https://www.lesswrong.com/posts/tdrK7r4QA3ifbt2Ty/is-ai-alignment-enough?commentId=An6L68WETg3zCQrHT
Larks21

You probably should have said 'yes' when asked if it was AI-written.

Larks40

This started happening in Hawaii, and to a lesser extent in Arizona. The resolution, apart from reducing net metering subsidies, has been to increased the fixed component of the bill (which pays for the grid connection) and reduce the variable component. My impression is this has been a reasonably effective solution, assuming people don't want to cut their connection entirely.

Larks20

I agree with you that basically anything in the stock market has much less counterparty risk than that. I disagree with basically all non-trading examples you give. 

It's not just the stock market, it's true for the bond market, the derivatives market, the commodities market... financial markets, a category which includes prediction markets, cannot function effectively with counterparty risk anything like 5%. 

My sense is around 1/20 Ubers don't show up, or if they show up, fail to do their job in some pretty obvious and clear way. 

If the Uber... (read more)

2habryka
Hmm, maybe I am just failing to model something here. Isn't really the only thing that happens when you have 5% randomly-distributed counterparty risk that you end up with like 5% spreads? That seems fine to me. To be clear, I don't feel very confident here, I just don't really understand why you can't just price in counterparty risk and then maybe end up with some bigger spreads (which I do agree is sad for prediction markets, but for most markets I don't mind the spread that much).
Larks71

In most domains except the most hardened part of the stock market counterparty risk is generally >5%.

This seems quite wrong to me:

  • High Yield Corporate Bond OAS spreads are <5% according to bloomberg, and most of that is economic risk, not "you will get screwed by a change of rules" risk.
  • Trades on US stock exchanges almost always succeed, many more 9s than just one.
  • If I buy a product in a box in a supermarket the contents of the box match the label >>95% of the time.
  • Banks make errors with depositor balances <<5% of the time.
  • Most employers
... (read more)
4habryka
Sorry, I wanted to say "except the most hardened parts of the world (like the stock market)". I agree with you that basically anything in the stock market has much less counterparty risk than that. I disagree with basically all non-trading examples you give.  My sense is around 1/20 Ubers don't show up, or if they show up, fail to do their job in some pretty obvious and clear way.  True for the most commoditized products. For anything else error rates seem to me to be around 5%. My guess is my overall Amazon error rate has been around 2%, which is lower, but not much lower (usually Amazon sent me something broken that previously was returned where they couldn't spot the error). I think that's false, at least the statistics on wage theft seemed quite substantial to me. I am kind of confused how to interpret these, but various different studies on Wikipedia suggest wage theft on-average to be around 5%-15% (higher among lower-income workers). I agree this is true for gas and water (and mostly true for electricity, though PG&E is terrible and Berkeley really has a lot of outages).  ---------------------------------------- Overall, I think 5% counterparty risk seems about right for most contracts I sign or business relationship I have. I agree that trading infrastructure is quite robust and in highly commoditized environments you get below that, but that's not the majority of my economic transactions.
Larks50

A bit dated but have you read Robin's 2007 paper on the subject?

Prediction markets are low volume speculative markets whose prices offer informative forecasts on particular policy topics. Observers worry that traders may attempt to mislead decision makers by manipulating prices. We adapt a Kyle-style market microstructure model to this case, adding a manipulator with an additional quadratic preference regarding the price. In this model, when other traders are uncertain about the manipulator’s target price, the mean target price has no effect on prices, and

... (read more)
1notfnofn
No, but it's exactly what I was looking for, and surprisingly concise. I'll see if I believe the inferences from the math involved when I take the time to go through it!
Larks20

Yes, sorry for being unclear. I meant to suggest that this argument implied 'accelerate agents and decelerate planners' could be the desirable piece of differential progress.

This post seems like it was quite influential. This is basically a trivial review to allow the post to be voted on.

Larks127

I agree in general, but think the force of this is weaker in this specific instance because NonLinear seems like a really small org. Most of the issues raised seem to be associated with in-person work and I would be surprised if NonLinear ever went above 10 in-person employees. So at most this seems like one order of magnitude in difference. Clearly the case is different for major corporations or orgs that directly interact with many more people. 

habryka*13-2

Note that one of the reasons why I cared about getting this report out was that Nonlinear was getting more influential as a middleman in the AI Safety funding ecosystem, through which they affected many people's lives and I think had influence beyond what a naive headcount would suggest. The Nonlinear network had many hundreds of applications. 

As a personal example, I also think Lightcone, given that its at the center of a bunch of funding stuff, and infrastructure work, should also be subject to greater scrutiny than specific individuals, given the number of individuals that are affected by our work. And we are about the same size as Nonlinear, I think.

Larks1915

I think there will be some degree to which clearly demonstrating that false accusations were made will ripple out into the social graph naturally (even with the anonymization), and will have consequences. I also think there are some ways to privately reach out to some smaller subset of people who might have a particularly good reason to know about this. 

If this is an acceptable resolution, why didn't you just let the problems with NonLinear ripply out into the social graph naturally?

3Lukas_Gloor
An organization gets applications from all kinds of people at once, whereas an individual can only ever work at one org. It's easier to discreetly contact most of the most relevant parties about some individual than it is to do the same with an organization. I also think it's fair to hold orgs that recruit within the EA or rationalist communities to slightly higher standards because they benefit directly from association with these communities. That said, I agree with habryka (and others) that 
habryka2412

I think that's a good question, and indeed I think that should be the default thing that happens!

In this case we decided to do something different because we received a lot of evidence that Nonlinear was actively suppressing negative information about them. As Ben's post states, the primary reason we got involved with this was that we heard Nonlinear was actively pressuring past employees to not say bad things about them, and many employees we talked to fely very scared of retribution if they told anyone about this, even privately, as long as it could some... (read more)

Larks134

If most firms have these clauses, one firm doesn't, and most people don't understand this, it seems possible that most people would end up with a less accurate impression of their relative merits than if all firms had been subject to equivalent evidence filtering effects.

In particular, it seems like this might matter for Wave if most of their hiring is from non-EA/LW people who are comparing them against random other normal companies.

Larks30

I would typically aim for mid-December, in time for the American charitable giving season.

Larks364

After having written an annual review of AI safety organisations for six years, I intend to stop this year. I'm sharing this in case someone else wanted to in my stead.

Reasons

  • It is very time consuming and I am busy.
  • I have a lot of conflicts of interests now.
  • The space is much better funded by large donors than when I started. As a small donor, it seems like you either donate to:
    • A large org that OP/FTX/etc. support, in which case funging is ~ total and you can probably just support any.
    • A large org than OP/FTX/etc. reject in which case there is a high chance
... (read more)
5robertzk
Thank you, Larks! Salute. FYI that I am at least one who has informally committed (see below) to take up this mantle. When would the next one typically be due? https://twitter.com/robertzzk/status/1564830647344136192?s=20&t=efkN2WLf5Sbure_zSdyWUw
4lc
🥲
gwern266

I will miss the annual updates. I didn't care about the question of who to donate to, but it was always good for catching up on research I missed in the flood, and a great starting point for beginners - I could just tell them to read a dozen papers from each annual review which sounded interesting and they'd have a great overview of how things were going.

Thank you for the work you've put in!

LarksΩ11268

Alignment research: 30

Could you share some breakdown for what these people work on? Does this include things like the 'anti-bias' prompt engineering?

Jacob_HiltonΩ8134

It includes the people working on the kinds of projects I listed under the first misconception. It does not include people working on things like the mitigation you linked to. OpenAI distinguishes internally between research staff (who do ML and policy research) and applied staff (who work on commercial activities), and my numbers count only the former.

Larks160

I would expect that to be the case for staff who truly support faculty. But many of them seem to be there to directly support students, rather than via faculty. The number of student mental health coordinators (and so on) you need doesn't scale with the number of faculty you have. The largest increase in this category is 'student services', which seems to be definitely of this nature.

Larks50

Thanks very much for writing this very diligent analysis.

I think you do a good job of analyzing the student/faculty ratio, but unless I have misread it seems like this is only about half the answer. 'Support' expenses rose by even more than 'Instruction', and the former category seems less linked to the diversity of courses offered than to things like the proliferation of Deans, student welfare initiatives, fancy buildings, etc.

2johnswentworth
I expect that the need for "support" roles scales roughly-linearly with the number of faculty. As I understand it, this is how most businesses ordinarily work, i.e. the need for support staff is roughly proportional to the number of object-level workers, at least once the organization passes ~100 people.
LarksΩ480

Is your argument about personnel overlap that one could do some sort of mixed effect regression, with location as the primary independent variable and controls for individual productivity? If so I'm so somewhat skeptical about the tractability: the sample size is not that big, the data seems messy, and I'm not sure it would capture necessarily the fundamental thing we care about. I'd be interested in the results if you wanted to give it a go though!

More importantly, I'm not sure this analysis would be that useful. Geography-based-priors only really seem us... (read more)

4Owain_Evans
I agree with most of this -- and my original comment should have been clearer. I'm wondering if the past five years of direct observations leads you to update the geography-based prior (which has been included in your alignment review for since 2018). How much do you expect the quality of alignment work to differ from a new organization based in the Bay vs somewhere else? (No need to answer: I realize this is probably a small consideration and I don't want to start an unproductive thread on this topic). 
LarksΩ3100
  • I prioritized posts by named organizations.
    • Diffractor does not list any institutional affiliations on his user page.
    • No institution I noticed listed the post/sequence on their 'research' page.
    • No institution I contacted mentioned the post/sequence.
  • No post in the sequence was that high in the list of 2021 Alignment Forum posts, sorted by karma.
  • Several other filtering methods also did not identify the post

However upon reflection it does seem to be MIRI-affiliated so perhaps should have been affiliated; if I have time I may review and edit it in later.

Vanessa KosoyΩ10190

Notice that in MIRI's summary of 2020 they wrote "From our perspective, our most interesting public work this year is Scott Garrabrant’s Cartesian frames model and Vanessa Kosoy’s work on infra-Bayesianism."

Larks40

13 years later: did anyone end up actually making such a book?

Larks50

The labels on the life satisfaction chart appear to be wrong; January 2021 comes before December 2020.

1DPiepgrass
I wonder what the correct labels would have been. Was "Joe Biden inaugurated" supposed to be the local minimum?
4Pattern
(From the link above, to 2020 AI Alignment Literature Review and Charity Comparison, by Larks.) Research Organisations * FHI: The Future of Humanity Institute * CHAI: The Center for Human-Aligned AI * MIRI: The Machine Intelligence Research Institute * GCRI: The Global Catastrophic Risks Institute * CSER: The Center for the Study of Existential Risk * OpenAI * Google Deepmind * BERI: The Berkeley Existential Risk Initiative * Ought * GPI: The Global Priorities Institute * CLR: The Center on Long Term Risk * CSET: The Center for Security and Emerging Technology * AI Impacts * Leverhulme Center for the Future of Intelligence * AI Safety camp * FLI: The Future of Life Institute * Convergence * Median Group * AI Pulse Other Research Capital Allocators * LTFF: Long-term future fund * OpenPhil: The Open Philanthropy Project * SFF: The Survival and Flourishing Fund Other Organisations * 80,000 Hours
Larks40

Well, with hemispherectomy, those problems are no more. Hemispherectomy is a procedure where half of the brain is removed. It has been performed multiple times without any apparent complications (example).

I was skeptical until I read the example. Now I am convinced!

3Mati_Roy
woops, I just realized the link was missing: https://www.webmd.com/brain/news/20191119/they-had-half-their-brains-removed-heres-what-happened-after
Larks30

It's hard to sell 1 million eggs for one price, and 1 million for another price.

Are you sure this is the case? It's common for B2B transactions to feature highly customised and secret pricing and discounts. And in this case they're not selling the same product from the customer's point of view: one buyer gets a million ethical eggs, while another gets a million ordinary (from their point of view) eggs.

1Pongo
Not sure at all! It still seems like the ordering is tricky. They don't know how many ethical eggs they've sold when selling towards the consumer. There's not a guarantee of future ethical eggs when buying the certificate. Maybe it works out OK, and they can sell 873,551 eggs at a regular price after that many certificates were bought, and the rest at the higher price. I know very little about how the food supply chain works
Larks20

Thanks for writing this; ordered.

Larks180

A teacher in year 9 walked up to a student who was talking, picked them up and threw them out of an (open) first floor window. 

Worth clarifying for US readers that 'first floor' in the UK would be 'second floor' in the US, because UK floor indexing starts at zero. So this event is much worse than it sounds.

3DirectedEvolution
I updated my OP with the link, thanks for sharing it!
Larks170

At the moment, the poor person and the rich person are both buying things. If the rich person buys more vaccine, that means they will buy less of the other things, so the poor person will be able to have more of them. So the question is about the ratios of how much the two guys care about the vaccine and how much they care about the other thing... and the answer is the rich guy will pay up for the vaccine when his vaccine:other ratio is higher than the other guys. This is the efficient allocation.

It might be the case that it is separately desirable to redi... (read more)

2scarcegreengrass
Are you saying there would be a causal link from the poor person's vaccine:other ratio to the rich person's purchasing decision? How does that work?

At the moment, the poor person and the rich person are both buying things. If the rich person buys more vaccine, that means they will buy less of the other things, so the poor person will be able to have more of them. So the question is about the ratios of how much the two guys care about the vaccine and how much they care about the other thing... and the answer is the rich guy will pay up for the vaccine when his vaccine:other ratio is higher than the other guys.

This is only true if the rich person is already spending as much money as possible, so an incr... (read more)

LarksΩ3100

Hey Daniel, thanks very much for the comment. In my database I have you down as class of 2020, hence out of scope for that analysis, which was class of 2018 only. I didn't include 2019 or 2020 classes because I figured it takes time to find your footing, do research, write it up etc., so absence of evidence would not be very strong evidence of absence. So please don't consider this as any reflection on you. Ironically I actually did review one of your papers in the above - this one - which I did indeed think was pretty relevant! (Cntrl-F 'Hendrycks' to find the paragraph in the article). Sorry if this was not clear from the text.

Larks120

Larks, excellent name choice for your AttackBot.

Thanks! I figured it was in the spirit of a DefectBot to defect linguistically as well, and there was a tiny chance someone might be doing naive string-matching. 

Larks80

You will have to wait for next time's obituary I'm afraid! I think Isusr should have a good grasp on the philosophical and ethical traditions I was attempting to channel with CooperateBot - while the insights are deep, I think the lengthy code is quite clear on the matter.

Larks50

I actually have no idea - I guess we are just two naturally very cooperative people!

Larks50

Cool competition! It makes me wish I had had more time to put into CooperateBot. At present I would say it instantiated a relatively naive view of cooperation, and could do much better if I invested more time considering the true nature of generosity. Looking at the obituary I suspect that CooperateBot may not last much longer.

3Tetraspace
How does your CooperateBot work (if you want to share?). Mine is OscillatingTwoThreeBot which IIRC cooperates in the dumbest possible way by outputting the fixed string "2323232323...".
3Vanilla_cabs
Can you tell us who is Insub and the story of your alliance with them?
Answer by Larks90

Holding constant the total amount of taxes you pay, it is better not to get a refund. This is the perspective you should take at the beginning of the year. 

Holding constant the amount of taxes you have already paid, it is better to get a refund. This is the perspective you should take at the end of the year. 

Larks70

I attempted to produce a rough estimate of this here (excerpted below):

... One (BERI funded!) study suggested that banning large gatherings reduced r0 by around 28%.
Unfortunately, protests seem in many ways ideal for spreading the disease. They involve a large number of people in a relatively small area for an extended period of time. Even protests which were advertised as being socially distanced often do not end up that way. While many people wear masks, photos of protests make clear that many do not, and those that are are often using cloth masks
... (read more)
Larks20

I still found this helpful as it allowed me to exit my directional Yang and Buttigieg positions with negative transaction cost.

Larks220

I would like to add that I think this is bad (and have the codes). We are trying to build social norms around not destroying the world; you are blithely defecting against that.

5jefftk
I'm not doing anything unilaterally. If I do anything at this point it will be after some sort of fair polling.
Larks150
This case is more complicated than the corporate cases because the powerful person (me) was getting merely the appearance of what she wanted (a genuine relationship with a compatible person), not the real thing. And because the exploited party was either me or Connor, not a third party like bank customers. No one thinks the Wells Fargo CEO was a victim the way I arguably was.

I think you have misunderstood the Wells Fargo case. These fake accounts generally didn't bring in any material revenue; they were just about internal 'number of new a... (read more)

Raemon100

Hmm. I'm not sure I fully understand the Wells Fargo case but I interpreted it as a concern between four parties:

1. The people who got fake accounts signed up for them.

2. The employees doing the fake signups

3. A middle management tier, which set quotas

4. A higher level management tier that (presumaby?) wanted middle management to actually be making money.

So, the people being defrauded are not the customers, but the higher management tier, basically. (But, also, this entire thing might just be a weird game that middle management tiers play with each ot... (read more)

4Elizabeth
Fixed, thank you for pointing this out.
Larks50

Is this very different from founding a pharmaceutical company?

3ryan_b
Yes - this fund requires pharmaceutical companies to generate the IP in the first place, and also to sell the successful drugs. A new pharmaceutical company will face the same risk profile as existing pharmaceutical companies; I would be very surprised if one could suddenly start investing according to the opposite pattern the others use. On the other hand, I don't see any reason why an existing pharmaceutical conglomerate could not employ this strategy or a similar one. They already have a huge amount of IP laying around undeveloped (it is from them a fund like this would acquire it) and other huge companies like General Electric have deliberately explored financial engineering as a corporate strategy. It failed in that case, but in this one we are just talking about supplementing the core strategy rather than replacing it.
Larks100

Critch wrote a related paper:

Existing multi-objective reinforcement learning (MORL) algorithms do not account for objectives that arise from players with differing beliefs.Concretely, consider two players with different beliefs and utility functions who may cooperate to build a machine that takes actions on their behalf. A representation is needed for how much the machine’s policy will prioritize each player’s interests over time. Assuming the players have reached common knowledge of their situation, this paper derives a recursion that any Pareto optimal

... (read more)
Load More