Comment author: [deleted] 06 June 2015 10:15:42PM 1 point [-]

I don't think it's an active waste of time to explore the research that can be done with things like AIXI models. I do, however, think that, for instance, flaws of AIXI-like models should be taken as flaws of AIXI-like models, rather than generalized to all possible AI designs.

So for example, some people (on this site and elsewhere) have said we shouldn't presume that a real AGI or real FAI will necessarily use VNM utility theory to make decisions. For various reasons, I think that exploring that idea-space is a good idea, in that relaxing the VNM utility and rationality assumptions can both take us closer to how real, actually-existing minds work, and to how we normatively want an artificial agent to behave.

Comment author: query 06 June 2015 10:56:32PM 0 points [-]

Modulo nitpicking, agreed on both points.

Comment author: [deleted] 05 June 2015 05:48:41PM 0 points [-]

Sometimes toy models are helpful and some times they are distractions that lead nowhere or embody a mistaken preconception. I see you as claiming these models are distractions, not that no model is possible. Accurate?

I very much favor bottom-up modelling based on real evidence rather than mathematical models that come out looking neat by imposing our preconceptions on the problem a priori.

The classes U and F above, should something like that ever come to pass, need not be AIXI-like (nor need they involve utility functions).

Right. Which is precisely why I don't like when we attempt to do FAI research under the assumption of AIXI-like-ness.

Comment author: query 05 June 2015 06:12:14PM *  0 points [-]

I very much favor bottom-up modelling based on real evidence rather than mathematical models that come out looking neat by imposing our preconceptions on the problem a priori.

(edit: I think I might understand after-all; it sounds like you're claiming AIXI-like things are unlikely to be useful since they're based mostly on preconceptions that are likely false?)

I don't think I understand what you mean here. Everyone favors modeling based on real evidence as opposed to fake evidence, and everyone favors avoiding the import of false preconceptions. It sounds like you prefer more constructive approaches?

Right. Which is precisely why I don't like when we attempt to do FAI research under the assumption of AIXI-like-ness.

I agree if you're saying that we shouldn't assume AIXI-like-ness to define the field. I disagree if you're saying it's a waste for people to explore that idea space though: it seems ripe to me.

Comment author: AeroRails 05 June 2015 04:32:01PM 1 point [-]

That sounds plausible, but how do you start to reason about such models of computation if they haven't even been properly defined yet?

Comment author: query 05 June 2015 05:15:36PM 0 points [-]

Formally, you don't. Informally, you might try approximate definitions and see how they fail to capture elements of reality, or you might try and find analogies to other situations that have been modeled well and try to capture similar structure. Mathematicians et al usually don't start new fields of inquiry from a set of definitions, they start from an intuition grounded in reality and previously discovered mathematics and iterate until the field takes shape. Although I'm not a physicist, the possibly incorrect story I've heard is that Feynman path integrals are a great example of this.

Comment author: query 05 June 2015 04:48:31PM 0 points [-]

Oh yes, it sounds like I did misunderstand you. I thought you were saying you didn't understand how such a thing could happen in principle, not that you were skeptical of the currently popular models. The classes U and F above, should something like that ever come to pass, need not be AIXI-like (nor need they involve utility functions).

I think I'm hearing that you're very skeptical about the validity of current toy mathematical models. I think it's common for people to motte and bailey between the mathematics and the phenomena they're hoping to model, and it's an easy mistake for most people to make. In a good discussion, you should separate out the "math wank" (which I like to just call math) from the transfer of that wank to reality that you hope to model.

Sometimes toy models are helpful and some times they are distractions that lead nowhere or embody a mistaken preconception. I see you as claiming these models are distractions, not that no model is possible. Accurate?

Comment author: query 05 June 2015 05:04:36PM *  0 points [-]

I offered the transform as an example how things can mathematically factor, so like I said, that may not be what the solution looks like. My feeling is that it's too soon to throw out anything that might look like that pattern though.

Comment author: [deleted] 05 June 2015 04:32:03PM *  2 points [-]

Then, once you actually construct an arbitrary GAI, you already know how to transform it into an FAI.

Frankly, I don't trust this claim for a second, because important components of the Friendliness problem are being completely shunted aside. For one thing, in order for this to even start making sense, you have to be able to specify a computable utility function for the AGI agent in the first place. The current models being used for this "mathematical" research don't have any such thing, ie: AIXI specifies reward as a real-valued percept rather than a function over its world-model.

The problem is not the need for large amounts of computing power (ie: the problem is not specifying the right behavior and then "scaling it down" or "approximating" a "tractable example from the class"). The problem is not being able to specify what the agent values in detail. No amount of math wank about "approximation" and "candidate class of formal models U" is going to solve the basic problem of having to change the structure away from AIXI in the first place.

I really ought to apologize for use of the term "math wank", but this really is the exact opposite approach to how one constructs correct programs. What you don't do to produce a correct computer program, knowing its specification, is try to specify a procedure that will, given an incomplete infinity of time, somehow transform an arbitrary program from some class of programs into the one you want. What you do is write the single exact program you want, correct-by-construction, and prove formally (model checking, dependent types, whatever you please) that it exactly obeys its specification.

If you are wondering where the specification for an FAI comes from, well, that's precisely the primary research problem to solve! But it won't get solved by trying to write a function that takes as input an arbitrary instance or approximation of AIXI and returns that same instance of AIXI "transformed" to use a Friendly utility function.

Comment author: query 05 June 2015 04:48:31PM 0 points [-]

Oh yes, it sounds like I did misunderstand you. I thought you were saying you didn't understand how such a thing could happen in principle, not that you were skeptical of the currently popular models. The classes U and F above, should something like that ever come to pass, need not be AIXI-like (nor need they involve utility functions).

I think I'm hearing that you're very skeptical about the validity of current toy mathematical models. I think it's common for people to motte and bailey between the mathematics and the phenomena they're hoping to model, and it's an easy mistake for most people to make. In a good discussion, you should separate out the "math wank" (which I like to just call math) from the transfer of that wank to reality that you hope to model.

Sometimes toy models are helpful and some times they are distractions that lead nowhere or embody a mistaken preconception. I see you as claiming these models are distractions, not that no model is possible. Accurate?

Comment author: [deleted] 04 June 2015 03:16:01PM 4 points [-]

I honestly don't understand how on Earth it would even be possible to understand FAI without understanding AGI on a general level. On some level, what you need isn't a team of Sufficiently Advanced Geniuses who can figure out Friendliness while simultaneously minimizing their own understanding of AGI, but old-fashioned cooperation among the teams who are likely to become able to build AGI, with the shared goal of not building any agent that would defy its creators' intentions.

(You can note that the creators' intentions might be, so to speak, "evil", but an agent that faithfully follows the "evil" intentions of an "evil" sort of human operator is already far Friendlier in kind than a paperclip maximizer -- it was just given the wrong operator.)

Comment author: query 05 June 2015 04:16:48PM *  0 points [-]

A mathematical model of what this might look like: you might have a candidate class of formal models U that you think of as "all GAI" such that you know of no "reasonably computable"(which you might hope to define) member of the class (corresponding to an implementable GAI). Maybe you can find a subclass F in U that you think models Friendly AI. You can reason about these classes without knowing any examples of reasonably computable members of either. Perhaps you could even give an algorithm for taking an arbitrary example in U and transforming it via reasonable computation into an example of F. Then, once you actually construct an arbitrary GAI, you already know how to transform it into an FAI.

So the problem may be factorable such that you can solve a later part before solving the first part.

So, I'd agree it might be hard to understand F without understanding U as a class of objects. And lets leave aside how you would find and become certain of such definitions. If you could, though, you might hope that you can define them and work with them without ever constructing an example. Patterns not far off from this occur in mathematical practice, for example families of graphs with certain properties known to exist via probabilistic methods, but with no constructed examples.

Does that help, or did I misunderstand somewhere?

(edit: I don't claim an eventual solution would fit the above description, this is just I hope a sufficient example that such things are mathematically possible)

Comment author: [deleted] 28 May 2015 03:54:15PM *  3 points [-]

Reading comments like this make me feel far better about my relative lack of social life. The things people who have it must go through... I think would rather be confined to my family (thankfully I am married to an intelligent woman) than to have to bite my tongue and not tell idiots that they are idi... well, at least telling them that they are wrong.

Is it a useful model that the enjoyment of having larger social circles comes at the price of frequent tongue-biting and being polite when you feel like doing a dramatic facepalm?

Comment author: query 29 May 2015 05:38:00PM 1 point [-]

I've enlarged my social circles, or the set of circles I can comfortably move in, and didn't end up on that model. I think I originally felt that way a lot, and I worked on the "feeling like doing a dramatic facepalm" by reflecting on it in light of my values. When dramatic face palms aren't going to accomplish anything good, I examine why I have that feeling and usually I find out it's because my political brain is engaged even when this isn't a situation where I'll get good outcomes from political-style conversation. You can potentially change your feelings so that other responses that you value become natural.

A warning though, it took me a long time to learn how to do this in a manner that didn't make me feel conflicted. There's always the danger of "I was nice to the person even after they said X" priming you in a bad way. You could also do even worse than before in terms of your values, simply because you're inexperienced with the non-facepalm response and do it badly. I think it was worth it for me, but depending on your situation it might be dangerous to mess with.

Comment author: Lumifer 27 May 2015 07:58:51PM 2 points [-]

Maybe the other person is a jerk and is on an obnoxious power trip at your expense. If you don't get suspicious and (internally) angry you're just setting yourself up as a victim.

Generic advice doesn't apply everywhere. A default "nod and slowly back away" response isn't bad but is not always useful.

Comment author: query 27 May 2015 08:12:34PM 0 points [-]

Agreed on the 2nd paragraph.

Optimally, you'd be have an understanding of the options available, how you work internally, and how other people respond so you could choose the appropriate level of anger, etc. Thus it's better to explore suggestions and see how they work than to naively apply them in all situations.

Comment author: lululu 27 May 2015 03:59:40PM 10 points [-]

Here is my favorite method -

Situation: Someone says something totally cuckoo crazy but they are someone I have to cooperate with in order to complete a task or who I have to maintain a good social relationship because we share friends or because they are otherwise cool. Also, the person is not convincible (I hang with hippies, this happens a lot).

Solution: a conspiratory shrug followed by "ehh... who can say, really" or "eh... the world is a strange place" or, if the statement is totally super crazy, just "ehh..."

Examples:

"I switched to a gluten free diet and I think my energy centers really cleared up!" shrug "ehh... the world is a strange place"

"Why do people say its strange that I named my baby Glutenball? They just don't understand how gluten symbolizes the glue that holds us all together!" shrug "ehh... who can say really?"

"9/11 was a conspiracy by the gluten lobby!" shrug "ehh..."

Comment author: query 27 May 2015 07:57:37PM *  4 points [-]

Seconded! Another phrase (whose delivery might be hard to convey in text) is "Look, I dunno, but anyways..."

Maybe the big idea is to come across as not expressing much interest in the claim, instead of opposing the claim? I think most people are happy to move on with the conversation when they get a "move on" signal, and we exchange these signals all the time.

I also like that this is an honest way to think about: I really am not interested in what I expect will happen with that conversation (even if I am interested in the details of countering your claim.)

Comment author: Lumifer 27 May 2015 07:41:29PM -1 points [-]

So it's just unrolling the basic "don't be an asshole, be polite instead" advice?

Comment author: query 27 May 2015 07:48:17PM 1 point [-]

I don't know what you mean, but I think I see a lot of people "being polite" but failing at one of these when it would be really useful for them.

For example, you can be polite while internally becoming more suspicious and angry at the other person (#3 and #4) which starts coming out in body language and the direction of conversation. Eventually you politely end the conversation in a bad mood and thinking the other person is a jerk, when you could've accomplished a lot more with a different internal response.

View more: Prev | Next