All of Patodesu's Comments + Replies

If a 5-star system of voting were to be implemented, the UI of voting could continue being the same, and the weights of previous votes could be used but as if they had in between 1 stars increments: strong downvote, downvote, no vote, upvote, strong upvote.

And a middle (3 stars) vote could be added.

I know that people don't think of both ways of voting as equivalents, and a regular "upvote" could reduce the score of a comment/ post. 

But they are similar enough, and the UI would be much simpler and not discourage people from voting.

Patodesu30

Cool, so MIRI is focusing on public passive support, PauseAI and others in active public support.

Now, can an org focus on the lobbying of pausing/ stopping (or redlines for killswitches) then?

I'm interested in what do people think are the best ways of doing advocacy in a way that gives more weight to the risks than the (supposed) benefits.

Talking about all the risks? Focusing on the expert polls instead of the arguments?

Some people post about AI Safety in the EA Forum without crossposting here

When they say stopping I think they refer to stopping it forever, instead of slowing down, regulating and even pausing development.

Which I think is something pretty much everyone agrees on.

I think there's two different misalignments that you're talking about. So you can say there's actually two different problems that are not recieving enough attention.

One is obvious and is between different people. 

And the other is inside every person. The conflict between different preferences, the not knowing what they are and how to aggregate them to know what we actually want.

Human empowerment is a really narrow target too

I'm kinda new here, so where all this EAF fear comes from?

Even if you think S-risks from AGI are 70 times less likely than X-risks, you should think how many times worse would it be. For me would be several orders of magnitude worse.

Can non Reinforcement Learning systems (SL or UL) become AGI/ Superintelligence and take over the world? If so, can you give an example?

3Lone Pine
They can if researchers (intentionally or accidentally) turn the SL/UL system into a goal based agent. For example, imagine a SayCan-like system which uses a language model to create plans, and then a robotic system to execute those plans. I'm personally not sure how likely this is to happen by accident, but I think this is very likely to happen intentionally anyway.