Comment author: JoshuaZ 18 December 2014 01:51:26AM *  0 points [-]

Please explain how say a trolley problem fits into your framework.

Comment author: L29Ah 18 December 2014 03:59:44AM 0 points [-]

Please explain how say a trolley problem fits into your framework.

The correct choice is to check out who do you want to be killed and saved more, and what are, for instance, the social consequences of your actions. I don't understand your question, it seems.

Comment author: JoshuaZ 17 December 2014 08:10:47PM *  1 point [-]

Not at all. If I do something that doesn't accomplish my goals that's generally labeled as something like "stupid." If I decide that I want to kill lots of people, the problem with that is ethical even if my goals are fulfilled by it. Most intuitions don't see these as the same thing.

Comment author: L29Ah 17 December 2014 08:38:31PM 0 points [-]

How does this contradicts my notion of ethics? You will surely use what you know about the ethical properties of manslaughter to reach the goal and save yourself from the troubles, like manipulating the public opinion in your favor via, for instance, imitation the target people attacking you. Or even consider if the goal is worthy at all.

Comment author: JoshuaZ 17 December 2014 06:19:01PM 1 point [-]

Because ethics is essentially simplified applied modeling of other beings.

This seems like a very non-standard notion of what constitutes ethics. Can you expand on this captures the usual intuitions about what the concerns of ethics are?

Comment author: L29Ah 17 December 2014 07:44:46PM 0 points [-]

This seems like a very non-standard notion of what constitutes ethics. Can you expand on this captures the usual intuitions about what the concerns of ethics are?

The concerns of ethics for a given agent is to facilitate one to interact with others effectively, no?

Comment author: JoshuaZ 17 December 2014 01:45:58PM 0 points [-]

It does not as the other person is parseable as multiple ones as well

That's not obvious. What if one entity is parseable in such a way and another one isn't?

the corresponding ethics will be constructed from special cases of the entity's behaviour, like it was done before.

Why?

I still don't get how the anthropic principle cares about the labels we assign to stuff.

Right. They shouldn't. So situations like this one may be useful intuition pumps.

Comment author: L29Ah 17 December 2014 02:18:48PM 0 points [-]

That's not obvious. What if one entity is parseable in such a way and another one isn't?

Every human produces lots of different kinds of behaviour, so it can be modeled as a pack of specialized agents.

Why?

Because ethics is essentially simplified applied modeling of other beings.

Comment author: JoshuaZ 17 December 2014 03:28:38AM 0 points [-]

Well, if we put our dual Manfred's in one trolley car, and one person in another, then the ethics might care.

More substantially, once uploads start being a thing, the ethics of these situations will matter.

The other contexts where these issues matter is in anthropics, expectations and trying to understand what the implications of Many-Worlds are. In this case, making the separation completely classical may be helpful: when one cannot understand a complicated situation, looking at a simpler one can help.

Comment author: L29Ah 17 December 2014 08:35:56AM 0 points [-]

It does not as the other person is parseable as multiple ones as well.

Uploading is not a thing atm, and once it is viable, the corresponding ethics will be constructed from special cases of the entity's behaviour, like it was done before.

I still don't get how the anthropic principle cares about the labels we assign to stuff.

Comment author: L29Ah 16 December 2014 01:10:16PM *  1 point [-]

How many people am I?

Does it make any difference?

Comment author: Manfred 15 December 2014 11:53:57PM 1 point [-]

The key property of me in this case is the anthropic one - 'my' existence allows me to infer things about causes of my existence.

Comment author: L29Ah 16 December 2014 07:50:05AM 0 points [-]

It does not as you don't obtain any world properties that 'your' existence should reflect with such a definition.

Comment author: L29Ah 21 November 2014 07:42:36AM *  1 point [-]

Video games are nicely described by <https://en.wikipedia.org/wiki/Reinforcement#Schedules>. I'd try make the disgusting part of the gaming, like the resources spent per utility achieved and the methods they are capturing my attention, the most sailent when i go over my decision three to do something.

Regarding "learning to want", it is a matter of constructing and applying a model of your motivation (like i did ↑, but subjectively tailored).

// hey wtf lw.c doesn't respect rfc1738

In response to ...
Comment author: shminux 12 November 2014 03:09:22AM *  4 points [-]

OP or at least the link in it should be removed promptly to not provide the troll with any free SEO.

In response to comment by shminux on ...
Comment author: L29Ah 21 November 2014 07:18:59AM 0 points [-]
Comment author: L29Ah 17 November 2014 04:34:56AM 2 points [-]

How would you build any other skill or habit? I don't really understand how the answer to your specific question would be different.

View more: Prev | Next