All of Pongo's Comments + Replies

Pongo20

The thought saver at the end of the heritability section asks you to remember some strategies for self control, but they've not been introduced yet

2spencerg
Good catch, thanks for pointing that out!
Pongo10

Presumably the welfare premium is reduced if the ethical egg providers can recoup some costs from a quality premium

2Archimedes
Sure, but the non-fungibility reintroduces the distribution and packaging complications. Part of the premise was to "sell these on the wholesale market as if they were totally normal eggs." The decoupling only works under fungibility assumptions, i.e., when the quality difference is not distinguishable or not important. Fortunately, this is close enough to true for many parts of the market and a partial solution is still progress.
Pongo10

Not sure at all! It still seems like the ordering is tricky. They don't know how many ethical eggs they've sold when selling towards the consumer. There's not a guarantee of future ethical eggs when buying the certificate.

Maybe it works out OK, and they can sell 873,551 eggs at a regular price after that many certificates were bought, and the rest at the higher price. I know very little about how the food supply chain works

Pongo10

IIUC, this exposes the high-welfare egg co to more risk. It's hard to sell 1 million eggs for one price, and 1 million for another price. So they probably have to choose to sell at the low welfare price constantly. But this means they build up a negative balance that they're hoping ethical consumers will buy them out of.

3Larks
Are you sure this is the case? It's common for B2B transactions to feature highly customised and secret pricing and discounts. And in this case they're not selling the same product from the customer's point of view: one buyer gets a million ethical eggs, while another gets a million ordinary (from their point of view) eggs.
Pongo90

Was this actually cross posted by EY, or by Rob or Ben? I prefer it being mentioned in the latter case

I posted this, and I'll make a note that I did so for any future Eliezer content where I hit the 'submit' button.

The causal process for this article looked like this:

  1. Eliezer wrote a lot of AI alignment content for Arbital. (This continues to be a good resource and I encourage you to browse it as-is, but it's pretty disorganized and a lot of it has notes like 'todo: fix X'.)
  2. Eliezer delegated to me the task of cleaning up and cross-posting his Arbital stuff. This isn't at the top of my priority list yet (partly because the content is already available on the
... (read more)
Pongo160

To add more color to the inadequate equilibrium: I didn’t want to hang out with people with a lot of risk, not because of how bad COVID would be for me, but because of how it would limit which community members would interact with me. But this also meant I was a community member who was causing other people to take less risk.

Pongo20

I didn’t mean to predict on this; I was just trying to see number of predictions on first one. Turns out that causes prediction on mobile

4jp
You can un-predict by clicking on the prediction block again.
Pongo30

Hoping, I guess, that the name was bad enough that others would call it an Uhlmann Filter

Pongo10

Oh, and I also notice that a social manoeuvring game (the game that governs who is admitted) is a task where performance is correlated with performance on (1) and (2)

Pongo140

First time I’ve seen a highlighted mod comment. I like it!

Pongo120

Most of the Inner Rings I've observed are primarily selected on (1) being able to skilfully violate the explicit local rules to get things done without degrading the structure the rules hold up and (2) being fun to be around, even for long periods and hard work.

Lewis acknowledges that Inner Rings aren't necessarily bad, and I think the above is a reason why.

1Pongo
Oh, and I also notice that a social manoeuvring game (the game that governs who is admitted) is a task where performance is correlated with performance on (1) and (2)
Pongo20

Making correct decisions is hard. Sharing more data tends to make them easier. Whether you'll thrive or fail in a job may well depend on parts you are inclined to hide. Also, though we may be unwilling to change many parts of ourselves, other times we are getting in our own way, and it can help to have more eyes on the part of the territory that's you

Pongo30

All uses of the second person "you" and "your" in this post are in fact me talking to myself

I wanted to upvote just for this note, but I decided it's not good to upvote things based on the first sentence or so. So I read the post, and it's good, so now I can upvote guilt-free!

4AnnaSalamon
Oh man; that article is excellent and I hadn't seen it. If anyone's wondering whether to click the link: highly recommend.
Pongo10

I would also be salty

Pongo30

I think this tag should cover, for example, auction mechanics. Auctions don't seem much like institutions to me

2Bird Concept
Maybe a tag-split would be in order. I think the actual technical, economic field of Mechanism Design gets discussed a bunch of LW. I'm not a huge fan of "Institution design" as a name since it's not actually an established name for a field (I think?), but it might have slightly different connotations. 
3Raemon
Well, similarly, I want a tag that's about Institution Design, including the parts of building institutions that are not mechanisms (like, culture and stuff). I'm a little salty because I originally created this tag and called it Institution Design, and then someone came and changed it. :P
Pongo50

Note also Odin was "Woden" in Old English

9gilch
In Old English Wednesday was "Wōdnesdæg", but yeah. Woden is the Anglo-Saxon version of Odin. Also Friday was "Frīġedæġ". It's not actually clear from the record if Fríge and Freyja are the same goddess or not, but they're so similar that it's a matter of some debate among scholars. The Norse pantheon apparently had no equivalent for Saturn, so Saturday kept the Roman name. The weekdays were named for the seven "naked-eye" planets known to Hellenistic astrology. (The Sun and Moon counted as planets in that system.) The seven planetary gods were said to watch over the Earth in hourly shifts in order of (geocentric) distance: Saturn, Jupiter, Mars, Sun, Venus, Mercury, Moon. Because a 24-hour day is 3 in arithmetic modulo 7, a different god opened each day of the week, and so the day was named for its opening god. Counting by 24's (or 3's) in this cycle ordered by distance gives us the familiar order of days of the week.
2lsusr
Yes. Fixed.
Pongo10

Should I be reading all the openings of transparent envelopes as actual openings, or are they sometimes looking at the sealed envelope and seeing what it contains (the burnings incline me towards the second interpretation, but I'm not sure)?

EDIT: Oh, I think I understand better now

Pongo10

Made a quick neural network (reaching about 70% accuracy), and checked all available scores.

 

 Its favorite result was: +2 Cha, +8 Wis. It would have like +10 Wis if it were possible.

For at least the top few results, it wanted to (a) apportion as much to Wis as possible, then (b) as much to Cha, then (c) as much to Con. So we have for (Wis, Cha, Con): 1. (8, 2, 0) 2. (8, 1, 1) 3. (8, 0, 2) 4. (7, 3, 0) 5. (7, 2, 1) ...

Pongo30

Seems like it's a lazy sort to me (with obvious wrinkles from the fact that the list can grow). It also seems to be a variant of drop sort (which is  via cheating) designed for repeated passes on the remaining list

Pongo40

Any favored resources on metta?

7Kaj_Sotala
TWIM has been a pretty popular metta variant recently. I like it as well, even though I always forget what some of the Rs were again. The linked Ron Crouch article also has good tips.
Pongo*80

If we solve the problem normally thought of as "misalignment", it seems like this scenario would now go well. If we solve the problem normally thought of as "misuse", it seems like this scenario would now go well. This argues for continuing to use these categories, as fixing the problems they name is still sufficient to solve problems that do not cleanly fit in one bucket or another

3Sam Clarke
This might be totally obvious, but I think it's worth pointing out that even if we "solve misalignment" - which I take to mean solving the technical problem of intent alignment - Bob could still chose to deploy a business strategising AI, in which case this failure mode would still occur. In fact, with a solution to intent alignment, it seems Bob would be more likely to do this, because his business strategising assistant will actually be trying to do what Bob wants (help his business suceed).
Pongo30

Sure!

I see people on twitter, for example, doing things like having GPT-3 provide autocomplete or suggestions while they're writing, or doing grunt work of producing web apps. Plausibly, figuring out how to get the most value out of future AI developments for improving productivity is important.

There's an issue that it's not very obvious exactly how to prepare for various AI tools in the future. One piece of work could be thinking more about how to flexibly prepare for AI tools with unknown capabilities, or predicting what the capabilities will be.

Other th... (read more)

Pongo20

Now, it's still possible that accumulation of slow-turnover senescent cells could cause the increased production rate of fast-turnover senescent cells.

Reminds me of this paper, in which they replaced the blood of old rats with a neutral solution (not the blood of young rats), and found large rejuvinative effects. IIRC, they attributed it to knocking the old rats out of some sort of "senescent equilibrium"

1jmh
The Conboys are looking to start human trials with they neutral blood replacement approach (https://newatlas.com/medical/diluted-blood-plasma-reverse-aging-in-mice/?utm_source=New+Atlas+Subscribers&utm_campaign=9db0c9efb9-EMAIL_CAMPAIGN_2020_06_16_01_29&utm_medium=email&utm_term=0_65b67362bd-9db0c9efb9-92444869).  I didn't realize they were tying their work to senescent cells though.
Answer by Pongo*80

If timelines are short, where does the remaining value live? Some fairly Babble-ish ideas:

  • Alignment-by-default
    • Both outer alignment and inner by default
      • With full alignment by default, there's nothing to do, I think! One could be an accelerationist, but the reduction in suffering and lives lost now doesn't seem large enough for the cost in probability of aligned AI
      • Possibly value could be lost if values aren't sufficiently cosmopolitan? One could try and promote cosmopolitan values
    • Inner alignment by default
      • Focus on tools for getting good estimates of human va
... (read more)
3Daniel Kokotajlo
Thanks! I found this the most helpful of the answers so far. I'd be interested to hear more about leveraging AI tools for direct work; can you say more?
Pongo30

Even more so, I would love to see your unjustifiable stab-in-the-dark intuitions as to where the center of all this is

Curious why this in particular (not trying to take umbrage with wanting this info; I agree that there’s a lot of useful data here. Would be a thing I’d also want to ask for, but wouldn’t have prioritised)

Good question. I'm not sure if this will make sense, but: this is somehow the sort of place where I would expect peoples' stab-in-the-dark faculties ("blindsight", some of us call it at CFAR) to have some shot at seeing the center of the thing, and where by contrast I would expect that trying to analyze it with explicit concepts that we already know how to point to would... find us some very interesting details, but nonetheless risk having us miss the "main point," whatever that is.

Differently put: "what is up with institutional cultures lately?" is a que... (read more)

Pongo20

Seems like you’re missing an end to the paragraph that starts “Related argument”

2AnnaSalamon
Thanks! Fixed.
Pongo00

I liked your example of being uncertain of your probabilities. I note that if you are trying to make an even money bet with a friend (as this is a simple Schelling point), you should never Kelly bet if you have discounted rate of 2/3 or less of your naïve probabilities.

The maximum bet for  is when  is 1, which is  which crosses below 0 at 

Pongo60

In the pie chart in the Teams section, you can see "CooperateBot [Larks]" and "CooperateBot [Insub]"

Pongo10

Yeah, that's what my parenthetical was supposed to address

(particularly because there is large interpersonal variation in the strength of hedging a given qualifier is supposed to convey)

Perhaps you are able to get more reliable information out of such statements than I am.

Answer by Pongo20

I like qualifiers that give information on the person's epistemic state or, even better, process. For example:

  • "Technological progress happens exponentially (70%)"
  • "My rough intuition is that technological progress happens exponentially"
  • "Having spent a year looking into the history of technological progress, I concluded it happens exponentially: [here are specific data sources, or links, or some sort of pointer]"

Given that I don't start thinking that anyone can report directly the state of the world (rather than their beliefs and understanding of it), "From ... (read more)

2Adam Zerner
When interpreted literally, I agree. But "from what I understand" connotes a sense that you're not super confident eg. because you're talking about something that you're not an expert on.
Pongo120

You can steer a bit away from catastrophe today. Tomorrow you will be able to do less. After years and decades go by, you will have to be miraculously lucky or good to do something that helps. At some point, it's not the kind of "miraculous" you hope for, it's the kind you don't bother to model.

Today you are blind, and are trying to shape outcomes you can't see. Tomorrow you will know more, and be able to do more. After years and decades, you might know enough about the task you are trying to accomplish to really help. Hopefully the task you find yourself faced with is the kind you can solve in time.

Pongo10

I think you're right. I think inline comments are a good personal workflow when engaging with a post (when there's a post I want to properly understand, I copy it to a google doc and comment it), but not for communicating the engagement

Pongo40

My understanding is the first thing is what you get with UDASSA and the second thing would be what you get is if you think the Solomonoff prior is useful for predicting your universe for some other reason (ie not because you think the likelihood of finding yourself in some situation covaries with the Solomonoff prior's weight on that situation)

Pongo300

He treated it like a game, even though he was given the ability to destroy a non-trivial communal resource.

I want to resist this a bit. I actually got more value out of the "blown up" frontpage then I would have from a normal frontpage that day. A bit of "cool to see what the LW devs prepared for this case", a bit of "cool, something changed!" and some excitement about learning something.

7Lukas Finnveden
Yeah, I'm not sure if the net effect of the blown-up frontpage is positive or negative for me, but it's definitely dominated by enjoyability/learning from posts about it, rather than the inability to see the frontpage for a day. (Very similar to how the value of a game is dominated by the value of the time spent thinking about it, while playing.) I didn't even try to access the frontpage on the relevant day, but I don't think it would have been much of an inconvenience, anyway; I could have just gone to the EA forum or my RSS feed for similar content (or to greaterwrong for the same content, if that was still up).
Pongo50

That’s a very broad definition of ‘long haul’ on duration and on severity, and I’m guessing this is a large underestimate of the number of actual cases in the United Kingdom

If the definition is broad, shouldn't it be an overestimate?

3Zvi
The overestimate I was thinking of was the number of Covid cases period, not the number of long haul cases. Wording could be improved, I've edited original. 
Pongo10

My attempt to summarize the alignment concern here. Does this seem a reasonable gloss?

It seems plausible that competitive models will not be transparent or introspectable. If you can't see how the model is making decisions, you can't tell how it will generalize, and so you don't get very good safety guarantees. Or to put it another way, if you can't interact with the way the model is thinking, then you can't give a rich enough reward signal to guide it to the region of model space that you want

1Mark Xu
This seems not false but it also seems like you're emphasizing the wrong bits, e.g. I don't think we quite need the model to be transparent/"see how it's making decisions" to know how it will generalize. 
Pongo10

Most importantly, the success of the scheme relies on the correctness of the prior over helper models (or else the helper could just be another copy of GPT-Klingon)

I'm not sure I understand this. My understanding of the worry: what if there's some equilibrium where the model gives wrong explanations of meanings, but I can't tell using just the model to give me meanings.

But it seems to me that having the human in the loop doing prediction helps a lot, even with the same prior. Like, if the meanings are wrong, then the user will just not predict the correct word. But maybe this is not enough corrective data?

Pongo80

As someone who didn't receive the codes, but read the email on Honoring Petrov Day, I also got the sense it wasn't too serious. The thing that would most give me pause is "a resource thousands of people view every day".

I'm not sure I can say exactly what seems lighthearted about the email to me. Perhaps I just assumed it would be, and so read it that way. If I were to pick a few concrete things, I would say the phrase "with our honor intact" seems like a joke, and also "the opportunity to not destroy LessWrong" seems like a silly phrase (kind of similar to... (read more)

Pongo40

Suppose phishing attacks do have an 80%+ success rate. I have been the target of phishing attempts 10s of times, and never fallen for it (and I imagine this is not unusual on LW). This suggests the average LWer should not expect to fall victim to a phishing attempt with 80% probability even if that is the global average

Pongo40

This summary was helpful for me, thanks! I was sad cos I could tell there was something I wanted to know from the post but couldn't quite get it

In a Stag Hunt, the hunters can punish defection and reward cooperation

This seems wrong. I think the argument goes "the essential difference between a one-off Prisoner's Dilemma and an IPD is that players can punish and reward each other in-band (by future behavior). In the real world, they can also reward and punish out-of-band (in other games). Both these forces help create another equilibrium where people cooper... (read more)

Pongo20

I think it's probably true that the Litany of Gendlin is irrecoverably false, but I feel drawn to apologia anyway.

I think the central point of the litany is its equivocation between "you can stand what is true (because, whether you know it or not, you already are standing what is true)" and "you can stand to know what is true".

When someone thinks, "I can't have wasted my time on this startup. If I have I'll just die", they must really mean "If I find out I have I'll just die". Otherwise presumably they can conclude from their continued aliveness that they ... (read more)

Pongo30

I wonder why it seems like it suggests dispassion to you, but to me it suggests grace in the presence of pain. The grace for me I think comes from the outward- and upward-reaching (to me) "to be interacted with" and "to be lived", and grace with acknowledgement of pain comes from "they are already enduring it"

Pongo10

Wondering if these weekly talks should be listed in the Community Events section?

3habryka
Yeah, we have an open PR that adds online events to the Community section and the navigation menu on the left. Currently all events need a physical location, which is obviously pretty dumb during a global pandemic where most events are online, but it obviously made sense in the pre-pandemic world, so we've been encouraging a number of online meetup organizers to post them as normal posts instead. 
Pongo10

I like this claim about the nature of communities. One way people can Really Try in a community is by taking stands against the way the community does things while remaining part of the community. I can’t think of any good solutions for encouraging this without assuming closed membership (or other cures worse than the disease)

Load More