All of rank-biserial's Comments + Replies

"That sounds really interesting. If a company working on the problem was paying a lot, I would consider jumping ship."

The Chinese stated preferences here closely track Western revealed preferences. Americans are more likely to dismiss AI risk post-hoc in order to justify making more money, whereas it seems that Chinese people are less likely to sacrifice their epistemic integrity in order to feel like a Good Guy, Hire people, and pay them money!

3Tomás B.
Pay Terry Tao his 10 million dollars!

So that's how you draft scissor statements >:)

2Vanessa Kosoy
The point of reference was salaries of academics in the US, across all ranks.

My understanding here is that while this is true, it will discourage the 5%, who will just go work for FAANG and donate money to someone worse (or someone overwhelmed with work), simultaneously losing any chance at a meaningful job. The point being that yes, it's good to donate, but if everyone donates (since that is the default rat race route), noone will do the important work.

No! If everyone donates, there will be enough money to pay direct workers high salaries. I know this goes contra to the image of the selfless, noble Effective Altruist, but if you want shit to get done you should pay people lots of money to do it.

5mruwnik
i.e. make it so EA is an attractive alternative to tech, thereby solving both problems at once?

A human life, with all its joys and all its pains, adding up over the course of decades, is worth far more than your brain’s feelings of comfort or discomfort with a plan.

In case you're wondering why Rationalists Don't Win, it's because they're too busy lying to themselves about what their true preferences are.

Ok, sick. I largely agree with you btw (about the hamster wheel being corrosive). If I came off as agressive, fyi, I liked the spirit of your post a lot, and I strong-upvoted it.

1sudo
Cool!

Yes, selfish agents want to not get turned into paperclips. But they have other goals too. You can prefer alignment be solved, while not wanting to dedicate your mind, body, and soul to waging a jihad against it. Where can Charlie effectively donate, say, 10% of his salary to best mitigate x-risk? Not MIRI (according to MIRI).

You don't need that many resources.

True, if we're talking solely about alignment. If we're talking about the larger space of, as you put them, "maximize objective function"-type goals, then there's plenty of demand for resources. Let's say I wanna do (actually effective) longevity research. Since the competition for grant money is (like most things) Goodharted and broken, and because I don't have enough biology credentials, I'm gonna need to self-fund in order to buy lab materials and grad student slaves.

1sudo
There is no universal advice that I can give.  The problem is that people are assuming that wheeling is correct without checking that it is. I'm not proposing developing an allergic reaction to colleges or something.

If you have EA-like goals, you have a "maximize objective function"-type goal. It's in the same shape as "become as rich as possible" or "make the world as horrible as possible." Basically, the conventional path is highly highly unlikely to get you all the way there. In this case, you probably want to get into the

  1. Get skills+resources
  2. Use skills+resources to do impact
  3. Repeat

Loop.

I'm in a similar situation to yours. (I'm currently in the "Bob" stage of the Alice -> Bob -> Charlie pipeline.) How do you propose I, and those like us, go about d... (read more)

1sudo
My Solution -- Optimize skill-building over resource-collection. You don't need that many resources. Ask:  1. What is the skill I'm most interested in building right now? 2. What's the best way to get this skill? A couple of things: 1. Open source libraries are free tutoring 2. Most alignment-relevant skills can be effectively self-taught 3. Projects are learning opportunities that demand mastery 4. Tutoring is cheaper and possibly more effective than college tuition

One of the problems here is that, as of right now, there isn't much of a middle path between "Stay at Google and do nothing" and "quit your job to do alignment work full-time". Then there's the issue of status-seeking vs. altruism as a case of revealed vs. stated preferences. If there was a way to make $750k a year and save the world, people would be all over that. I, personally, would be all over that.

But there isn't. If we use johnswentworth as an optimistic case, those who would go into independent AI alignment work full-time would make about $90k per y... (read more)

2sudo
If you think that we have a fair shot at stopping AI apocalypse, and that AGI is a short-term risk, then it is absolutely rational to optimize for solving AI safety. This is true even if you are entirely selfish. Also, this essay is about advice given to ambitious people. It's not about individual people choosing unambitious paths (wheeling). Charlie is a sad example of what can happen to you. I'm not complaining about him.

I endorse Rohin Shah's response to that post.

You might think "well, obviously the superintelligent AI system is going to care about things, maybe it's technically an assumption but surely that's a fine assumption". I think on balance I agree, but it doesn't seem nearly so obvious to me, and seems to depend on how exactly the agent is built. For example, it's plausible to me that superintelligent expert systems would not be accurately described as "caring about things", and I don't think it was a priori obvious that expert systems wouldn't lead to AGI. Si

... (read more)
3Rob Bensinger
This seems like a very different position from the one you just gave: I took you to be saying, 'You can retroactively fit a utility function to any sequence of actions, so we gain no predictive power by thinking in terms of utility functions or coherence theorems at all. People worry about paperclippers not because there are coherence pressures pushing optimizers toward paperclipper-style behavior, but because paperclippers are a vivid story that sticks in your head.'

Even humans have a huge range of intellectual capacity, and someone who is good at math may not be good at say, writing a novel. So the idea of "general intelligence" is pretty weak from the outset, and it's certainly not a binary value that you either have or have not.

https://en.wikipedia.org/wiki/G_factor_(psychometrics)

Answer by rank-biserial70

I find point no. 4 weak.

  1. Unaligned AGI will try to do something horrible to humans (not out of maliciousness, necessarily, we could just be collateral damage), and will not display sufficiently convergent behavior to have anything resembling our values.

I worry that when people reason about utility functions, they're relying upon the availability heuristic. When people try to picture "a random utility function", they're heavily biased in favor of the kind of utility functions they're familiar with, like paperclip-maximization, prediction error minimiza... (read more)

5Rob Bensinger
Coherence arguments imply a force for goal-directed behavior.

https://en.wikipedia.org/wiki/Brain_stimulation_reward

https://doi.org/10.1126/science.140.3565.394

https://sci-hub.hkvisa.net/10.1126/science.140.3565.394

In our earliest work with a single lever it was noted that while the subject would lever-press at a steady rate for stimulation to various brain sites, the current could be turned off entirely and he would continue lever-pressing at the same rate (for as many as 2000 responses) until told to stop.

It is of interest that the introduction of an attractive tray of food produced no break in responding, alt

... (read more)

It's to push our understanding of some crazy niche field and push it further beyond. The widespread recognition is nice, I'm sure, but I don't think most PhD graduates' goals are this.

Insufficiently Hansonpilled.

1Sweetgum
I randomly decided to google “hansonpilled” today to see if anyone had coined the term, congratulations on being one of two results.
2ljh2
Sorry, what does "hansonpilled" mean? Does Robin Hanson have some insight on this as well?

It's time for me to shill for Big Pharma again.

I spent years trying to come up with a mental strategy that would reliably generate willpower, all to no avail. To my knowledge, there is one (1) way to cut through this Gordian knot, and that way is amphetamines.

  • It is easy to get a psychiatrist to prescribe you amphetamines.[1]
  • If you request generic drugs over brand-name, they're pretty inexpensive.
  • Amphetamines greatly reduce one's appetite; Adderall was originally sold as a diet pill. For the first time in my life, I could empathize with those people who
... (read more)

Hamming distance assumes that the strings you're comparing are the same length, but "Yavin" is shorter than "Yarvin". Levenshtein distance is the smallest number of insertions, deletions, or character substitutions required to get from string A to string B, while Hamming distance only counts char-substitutions.

2Pattern
Yavin Yvain I thought the av switch is more natural than the other. I wouldn't call it a common typo, but... I can see why if you count a switch as the same as adding a letter, then they'd be the same.

Epistemic status: Shaky

That's not how it works. People bred chihuahuas from wolves, and the blue whale evolved from a goat-sized land mammal. If the additive genes that control spinal length just upregulate some pre-existing developmental process, it's not too much of a leap to assume that the additive genes that control intelligence just turn up a similar developmental knob. That would imply that intelligence-increasing mutations are common (being less complex), and that intelligence could be selected for like any other trait.

1Yair Halberstadt
How many generations did it take for both? I would estimate about 2000 for dogs, and many more for whales. That's plenty of time for new mutations to occur.

Within humans, intelligence-related genes behave additively. That is, the predicted IQ of a child of a 95 IQ mother and a 115 IQ father follows a normal distribution centered at 105. Since "general intelligence" exists to some degree in octopuses, we're amplifying an existing trait, rather than creating one de novo. If intelligence-related genes behave additively in octopuses as well (and we should assume that they do), then breeding octopi for intelligence should be as straightforward as breeding cows for milk.

1Yair Halberstadt
I would agree for small increases in intelligence, but what happens when we use up all the available genes for intelligence that already exist in octopuses, and need to wait for new ones to mutate?

Well, "Yarvin" is closer in Levenshtein distance, plus you're acting coy, so I'm updating towards "Yarvin" 😎

2Pattern
Don't they have the same Hamming distance?
2Raemon
I think he just literally means both

Some Yavin guy posted some good stuff. Then I found out he had his own blog. That was pretty good.

Is "Yavin" supposed to be "Yarvin" or "Yvain"? Both are quite plausible.

4Shmi
Imagine a cross between the two.
5Gordon Seidoh Worley
Yes ☺️

This is a Moldbug argument, but:

Markets aren't totally efficient, but they are the most adequate part of our civilization. For-profit corporations usually follow the same general structure, so the best form of governance looks like a joint-stock company.

3Gitdes
More like a statement, rather than an argument. 
0TAG
Efficient at what? There's no such thing as just being optimal...you have to optimise something.
2jmh
While I don't fully agree with the claim here (the comment or in the book) the best argument along these lines is probably David Friedman and "The Machinery of Freedom".
  • The "electrode in the reward center" setup has been proven to work in humans, whereas jhanas may not tranfer over Neuralink.
  • Deep brain stimulation is FDA-approved in humans, meaning less (though nonzero) regulatory fuckery will be required.
  • Happiness is not pleasure; wanting is not liking. We are after reinforcement.
2Logan Riggs
Could you link the proven part? Jhana's seem much healthier, though I'm pretty confused imagining your setup so I don't have much confidence. Say it works and gets past the problems of generalizing reward (eg the brain only rewards for specific parts of research and not others) and ignoring downward spiral effects of people hacking themselves, then we hopefully have people who look forward to doing certain parts of research.  If you model humans as multi-agents, it's making a certain type of agent (the "do research" one) have a stronger say in what actions get done. This is not as robust as getting all the agents to agree and not fight each other. I believe jhana gets part of that done because some sub-agents are pursuing the feeling of happiness and you can get that any time.

Judging by the fox domestication experiment, intelligence could emerge way sooner than 50 generations. The fox experiment used a population size on the order of 1e2, and selected the top 20% most docile foxes to act as the breeding pool for the next generation. You could accelerate this process by increasing the population size, while selecting a smaller fraction of intelligent octopuses.

1Yair Halberstadt
Docility seems much easier to breed than intelligence, because it's about reducing the foxes ability to do things (act violently), rather than create new capabilities. I.e. a random mutation has a high chance of increasing docility, but a very low chance of increasing intelligence.
Answer by rank-biserial30

Phantoms in the Brain by V.S. Ramachandran is a fun read, and will teach you a thing or two about neuropsychology.

1Yair Halberstadt
Really enjoyed that book when I read it about 10 years ago!
Answer by rank-biserial10

Code by Simon Singh is a nice, low-effort intro to cryptography.

[This comment is no longer endorsed by its author]Reply
1Yair Halberstadt
Interesting. I wasn't impressed by his Fermat's last theorem book which was basically purely a biography of various mathematicians and contained very little actual information. Is this one better? As it happens I found Applied Cryptography extremely readable (albeit out of date), so in this case I'm happy to recommend a textbook.
  1. Eco-terrorism, animal rights terrorism, educational terrorism (!), and many other forms of terrorism are morally justified and should be practiced by advocates of those causes.

Ah yes, the old "let's bomb DeepMind" approach to AI safety.

2rogersbacon
I hope it goes without saying that this is a heresy and not something I actually believe. A recent article in the Journal of Controversial Ideas makes the case for animal-rights terrorism.  "There is widespread agreement that coercive force may be used to prevent people from seriously and wrongfully harming others. But what about when those others are non-human animals? Some militant animal rights activists endorse the use of violent coercion against those who would otherwise harm animals. In the philosophical literature on animal ethics, however, theirs is a stance that enjoys little direct support. I contend that such coercion is nevertheless prima facie morally permissible."

Boo

(I'm booing all the terrorisms)

  1. The virtually universal practice of assigning a permanent name at birth (which only exists so that governments can more easily tax us) has caused a species-wide shift towards a more narcissistic and egotistical mode of being.

Neither of us are using our government-assigned names right now. Has the internet-culture norm of creating pseudonymous usernames countered this shift?

1rogersbacon
Interesting. I guess in some ways yes because it's giving people access to another form of identity but it's also kind of orthogonal in that the identity is only used in virtual environment and it's pseudonymous. The argument in this heresy is that we being less attached to our names IRL would cause a shift of some kind in cognition/consciousness. 
  1. Money is, in fact, the root of all evil and we will never be free until all monetary systems are abolished.

Money allows people to express their preferences more than they otherwise could. People's revealed preferences are invariably more evil than their stated preferences. Money reveals evil, but it does not create it.

3Measure
Revealing+enabling suppressed evil causes harm.
  1. We will never stop fighting about race. The only way out of this racial hellscape is to simply breed race away by tweaking dating app algorithms to increase the number of interracial matings.

First, this will never work, because people have strong racial preferences. Second, people will still fight about racial features even if everyone is mixed. Just look at Brazil.

1rogersbacon
Being a little tongue-in-cheek with this one, but I think recent US history shows racial preferences are more malleable than we might think. Will there be a tipping point when everyone is either mixed or has a close relative that's mixed where it will seem a little more silly to argue about race? I don't know about Brazil and would be curious to hear more like Ben Pace. 
3Ben Pace
For the less-travelled like me, can someone explain how people fight about racial features in Brazil given more people being mixed?

I'm glad Demis is somewhat reasonable. How tf can we pull the fire alarm? A petition?

[This comment is no longer endorsed by its author]Reply
5P.
Did you change your mind about Demis being reasonable or a petition being a good idea? And why? I’m considering sending him an email.
Answer by rank-biserial10

Eliezer replied to a comment of mine recently, coming out in favor of going down the human augmentation path. I also think genetically engineered Von Neumann babies are too far off to be realistic.

If we can really crack human motivation, I expect possible productivity gains of maybe one or two OOM.

Picture a pair of reseachers, one of whom controls an electrode wired to the pleasure centers of the other. Imagine they have free access to methamphetamine and LSD.

You don't need to be a genius to make this happen.

Great! If I recall correctly, you wanted genetically optimized kids to be gestated and trained.

I suspect that akrasia is a much bigger problem than most people think, and to be truly effective, one must outsource part of their reward function. There could be massive gains.

What do you think about the setup I outlined, where a pair of reseachers exist such that one controls an electrode embedded in the other's reward center? Think Focus from Vinge's A Deepness In The Sky.

1Logan Riggs
The better version then reward hacking I can think of is inducing a state of jhana (basically a pleasure button) in alignment researchers. For example, use neuro-link to get the brain-process of ~1000 people going through the jhanas at multiple time-steps, average them in a meaningful way, induce those brainwaves in other people.  The effect is people being satiated with the feeling of happiness (like being satiated with food/water), and are more effective as a result.
6Marvin
I think memetically 'optimized' kids (and adults?) might be an interesting alternative to explore. That is, more scalable and better education for the 'consequentialists' (I have no clue how to teach people that are not 'consequentialist', hopefully someone else can teach those) may get human thought-enhancement results earlier, and available to more people. There has been some work in this space and some successes, but I think that in general, the "memetics experts" and the "education experts" haven't been cooperating properly as much as they should. I think it would seem dignified to me to try bridging this gap. If this is indeed dignified, then that would be good, because I'm currently in the early stages of a project trying to bridge this gap.
8jayterwahl
(I predict that would help with AI safety, in that it would swiftly provide useful examples of reward hacking and misaligned incentives)

Fine. What do you think about the human-augmentation cluster of strategies? I recall you thought along very similar lines circa ~2001.

I don't think we'll have time, but I'd favor getting started anyways.  Seems a bit more dignified.

I'm not well calibrated on sub 1% probabilities. Yeah, the odds are low.

There are other classes of Hail Mary. Picture a pair of reseachers, one of whom controls an electrode wired to the pleasure centers of the other. Imagine they have free access to methamphetamine and LSD. I don't think research output is anywhere near where it could be.

So - just to be very clear here - the plan is that you do the bad thing, and then almost certainly everybody dies anyways even if that works?

I think at that level you want to exhale, step back, and not injure the reputations of the people who are gathering resources, finding out what they can, and watching closely for the first signs of a positive miracle.  The surviving worlds aren't the ones with unethical plans that seem like they couldn't possibly work even on the open face of things; the real surviving worlds are only injured by people who imagine that throwing away their ethics surely means they must be buying something positive.

One of Eliezer's points is that most people's judgements about adding 1e-5 odds (I assume you mean log odds and not additive probability?) are wrong, and even systematically have the wrong sign. 

Shmi150

The post talks about how most people are unable to evaluate these odds accurately, and that an indicator of you thinking you found a loophole actually being a sign that you are one of those people.

These are all strategies to buy time, so that alignment efforts may have more exposure to miracle-risk.

6Eliezer Yudkowsky
And what do you think are the chances that those strategies work, or that the world lives after you hypothetically buy three or six more years that way?

I can imagine a plausible scenario in which WW3 is a great thing, because both sides brick each other's datacenters and bomb each other's semiconductor fabs. Also, all the tech talent will be spent trying to hack the other side and will not be spent training bigger and bigger language models.

I imagine that WW3 would be an incredibly strong pressure, akin to WW2, which causes governments to finally sit up and take notice of AI.

And then spend several trillion dollars running Manhattan Project Two: Manhattan Harder, racing each other to be the first to get AI. 

And then we die even faster, and instead of being converted into paperclips, we're converted into tiny American/Chinese flags

-1jbash
That only gives you a brief delay on a timeline which could, depending on the horizons you adopt, be billions of years long. If you really wanted to reduce s-risk in an absolute sense, you'd have to try to sterilize the planet, not set back semiconductor manufacturing by a decade. This, I think, is a project which should give one pause.
  • Coordination (cartelization) so that AI capabilities are not a race to the bottom

  • Coordination to indefinitely halt semiconductor supply chains

  • Coordination to shun and sanction those who research AI capabilities (compare: coordination against embyronic human gene editing)

  • Coordination to deliberately turn Moore's Law back a few years (yes, I'm serious)

9Eliezer Yudkowsky
And do you think if you try that, you'll succeed, and that the world will then be saved?
-13Muad'Dib

Did you read the OP post? The post identifies dignity with reductions in existential risk, and it talks a bunch about the 'let's violate ethical injunctions willy-nilly' strategy

Yeah, most of them do. I have some hope for the strategy-cluster that uses widespread propaganda[1] as a coordination mechanism.

Given the whole "brilliant elites" thing, and the fecundity of rationalist memes among such people, I think it's possible to shift the world to a better Nash equilibrium.


  1. Making more rationalists is all well and good, but let's not shy away from no holds barred memetic warfare. ↩︎

1jbash
Coordination on what, exactly?

Is it not obvious to you that this constitutes dying with less dignity, or is it obvious but you disagree that death with dignity is the correct way to go?

What about Hail Mary strategies that were previously discarded due to being too risky? I can think of a couple off the top of my head. A cornered rat should always fight.

Never even THINK ABOUT trying a hail mary if it also comes with an increased chance of s-risk. I'd much rather just die.

Do they perchance have significant downsides if they fail?  Just wildly guessing, here.  I'm a lot more cheerful about Hail Mary strategies that don't explode when the passes fail, and take out the timelines that still had hope in them after all.

Possible modifications:

  • Allow withdrawing funds at any point. People like to feel in control, and this setup eliminates the need for the fundraiser to have a deadline.

  • For longer-term projects, let funds accumulate interest at slightly below market rates. The difference is pocketed by the crowdfunding platform. Customers get to feel financially smart while also contributing to a cool idea.

3Yoav Ravid
Yep, it's important to allow the canceling/changing of pledges, Kickstarter already allows that until the last 24 hours of the campaign where you can only cancel/change your pledge if it doesn't make the project go below funding level. How do you do refunds without a deadline though? Can you explain the second Idea a bit more? I don't think I understand it.
Load More