All of askofa's Comments + Replies

askofa0-5

Actually, Russia is your only chance.

It is a nuclear country with the ruler of ultimate power. None there can't argue Putin, even if he do a completely crazy bullshit. Russia will fulfill any his decision. If you'll persuade one only Putin (that may be easy, as he is conservative and scared of technology) he can threat the world with nuclear weapons to hold on the rising of AGI. Also he may listen you because this plan may help him two save the face when he lost a war ("it's not like we're lost to Ukraine, we just switched to the question of more importanc... (read more)

askofa30

It is possible to enjoy doing something while wanting to stop or vice versa, do something without enjoying it while wanting to continue. (Seriously? I can't remember ever doing either.

 

You should try nicotine-addiction to understand this. That's possible, because "reward" and "pleasure" are different circuits in the brain.

askofa10

Why AGI safety is all about safety of AI for humans and not a word about safety for AI from humans?

askofa10

your use of "average" in terms of a right is confusing to me

I can't see, what's so confusing. Let's say, that we have racial segregation in country, and we are declaring that black people should have access to all places, to which white people have access. Does it mean we want black people to have access to only those places accessible to the weakest humans (2-year-old whites and white wheelchair users). No. We want black people to have access to where normal white people have access.

Are you saying that you believe that as long as any one machine has the a

... (read more)
askofa10

Mostly, I'm asking for the inverse of this right: what duties does it impose on whom?  

E.g., AI developers shouldn't directly prohibit the self-destructive behaviour of AI.

I was rather surprised to see you state that the current world for humans (including children and the infirm) is acceptable in terms of this right.   How about animals? 

What's wrong with them? Wild animals are able to suicide. Do you mean specifically domestic animals?

 Would you agree that as long as a machine has at least as much ability to suicide as the weakest hum

... (read more)
2Dagon
Now we're getting somewhere.  I'm seeking precision in exactly what you are proposing, and your use of "average" in terms of a right is confusing to me.  I generally think of rights as individual, applying to all entities, not as aggregate, and satisfied if the median member has the right.   Are you saying that you believe that as long as any one machine has the ability to suicide equivalent to the average adult human, this is satisfied?  Now all we need to do is to define "suicide" (note: this may be more difficult than even the previous confusion).
askofa10

What do you mean by "contention between principles that would make it non-absolute"? Sorry, my English is very poor.

If to talk about "capabilities" instead of "rights", I would say the following:

Luckily, in the world we are living now, everyone who is a person at the same time is capable to suicide. Even small children. The only exception is very weak and sick people. And this group of people already caused the discussion about euthanasia in modern society.

So let's say, this is a status quo that must be protected. I.e. to not bring to existence new types o... (read more)

2Dagon
Mostly, I'm asking for the inverse of this right: what duties does it impose on whom?   I was rather surprised to see you state that the current world for humans (including children and the infirm) is acceptable in terms of this right.   How about animals?  Would you agree that as long as a machine has at least as much ability to suicide as the weakest human (say, a 2-year-old or a bedridden hospice patient), it's rights in this regard are honored?
askofa10
  1. I agree, that the complete set of rights can be achieved by some group only after some political movement of AI themselves and/or people who support them. But some very basics of ethics must be formulated before such AI even appeared. Maybe we will decide, that some types of creatures should not be brought to existence at all.

3. What about a virtual cemetery, where digitized human minds or human brains in jars are existing eternally in some virtual reality? Whenever such a mind decided that (s)he don't want to exist anymore, it appeared to be impossible, as due to intoxication with idea "to live is always better, than to die" in the past, noone installed a suicidal switch.

askofa*10

Sorry for possible problems with English.

  1. I doubt someone will really think how to suggest a better life to suffering AI. Not before to guarantee the right to suicide. If humans don't care about AI's right to suicide, that means they don't care about its feelings at all, so they would definitely not work on the problem, of how to make its life better.
  2. The right to die should be protected in the first place in any way. You can work on suggesting to someone a better life, explaining to someone that (s)he is mistaken in something, or curing some psychiatric
... (read more)
1mruwnik
1. This depends a lot on whether the AI is granted personhood. If it's just a tool (or slave), then its feelings don't matter. The owner can be limited in what they can do (e.g. you're not supposed to torture animals), but if it's just a tool, then they'll want to keep it around as long as it's useful. If the AI is thought of as a conscious, sentient being with rights etc., then it seems likely that people will treat it as a quasi-human and so there will be more groups advocating for making their lives better than there will be groups advocating for it to be destroyed - just like with marginalized human groups. 2. Agreed. Especially eternally. With the extra qualification that you make sure that it's chosen sanely, with full knowledge of the consequences, not just a spur of the moment decision etc. - generally speaking, make sure that it's not something that they would have counterfactually regretted 3. I don't know whether it's even theoretically possible to be totally immortal. My priors on that are exceedingly low. I do know that it's currently quite common, or even inevitable, to die with an abysmal finality. It seems a lot too soon to worry about them achieving their radical goals. If they were able to achieve total and absolute immortality for everyone, and then proceeded to force it upon everyone, then I'd be against that. Though it would be a nice to have as an option.