Comment author: RichardKennaway 22 March 2013 08:06:19AM *  1 point [-]

It's been tried, and our computational capabilities fell woefully short of succeeding.

Is that because we don't have enough brute force, or because we don't know what calculation to apply it to?

I would be unsurprised to learn that calculating the folding state having global minimum energy was NP-complete; but for that reason I would be surprised to learn that nature solves that problem, rather than finding a local minimum.

I don't have a background in biology, but my impression from Wikipedia is that the tension between Anfinsen's dogma and Levinthal's paradox is yet unresolved.

Comment author: Polymeron 17 April 2013 07:16:04AM *  1 point [-]

The two are not in conflict.

A-la Levinthal's paradox, I can say that throwing a marble down a conical hollow at different angles and force can have literally trillions of possible trajectories; a-la Anfinsen's dogma, that should not stop me from predicting that it will end up at the bottom of the cone; but I'd need to know the shape of the cone (or, more specifically, its point's location) to determine exactly where that is - so being able to make the prediction once I know this is of no assistance for predicting the end position with a different, unknown cone.

Similarly, Eliezer is able to predict that a grandmaster chess player would be able to bring a board to a winning position against himself, even though he has no idea what moves that would entail or which of the many trillions of possible move sets the game would be comprised of.

Problems like this cannot be solved on brute force alone; you need to use attractors and heuristics to get where you want to get.

So yes, obviously nature stumbled into certain stable configurations which propelled it forward, rather than solve the problem and start designing away. But even if we can never have enough computing power to model each and every atom in each and every configuration, we might still get a good enough understanding of the general laws for designing proteins almost from scratch.

Comment author: Polymeron 09 August 2012 02:12:41AM *  7 points [-]

When I was studying under Amotz Zahavi (originator of the handicap principle theory, which is what you're actually discussing), he used to make the exact same points. In fact, he used to say that "no communication is reliable unless it has a cost".

Having this outlook on life in the past 5 years made a lot of things seem very different - small questions like why some people don't use seatbelts and brag about it, or why men on dates leave big tips; but also bigger questions like advertizing, how hierarchical relationships really work, etc.

Also explained a lot about possible origins of (some) altruistic behaviors; Zahavi's favorite examples were from the research he and his wife conducted, wherein they observed small groups of social birds (forgot the species, sorry) where altrusitic behavior is common. And it turns out, it's the dominant birds who behave altrusitically, rather than exploit their weaker brethren - but doing so as a show of strength. My own favorite example is when a lower-status male caught a worm and tried feeding it to the alpha male. The latter proceeded to beat him up, take the morsel, and force-feed it back to the weaker male.

Best course I ever took :)

Comment author: Polymeron 12 June 2012 05:23:40PM *  1 point [-]

These questions seem decidedly UNfair to me.

No, they don't depend on the agent's decision-making algorithm; just on another agent's specific decision-making algorithm skewing results against an agent with an identical algorithm and letting all others reap the benefits of an otherwise non-advantageous situation.

So, a couple of things:

  1. While I have not mathematically formulated this, I suspect that absolutely any decision theory can have a similar scenario constructed for it, using another agent / simulation with that specific decision theory as the basis for payoff. Go ahead and prove me wrong by supplying one where that's not the case...

  2. It would be far more interesting to see a TDT-defeating question that doesn't have "TDT" (or taboo versions) as part of its phrasing. In general, questions of how a decision theory fares when agents can scan your algorithm and decide to discriminate against that algorithm specifically, are not interesting - because they are losing propositions in any case. When another agent has such profound understanding of how you tick and malice towards that algorithm, you have already lost.

Comment author: [deleted] 24 July 2011 11:40:07PM *  10 points [-]

I think you are doing this wrong, usually by the time I have convincing human-like androids I'm pretty near turning the Moon into computronium. You should focus on buying more processing power at this stage, perhaps set up a few hard to detect deep sea domes.

In response to comment by [deleted] on Help me transition to human society!
Comment author: Polymeron 01 June 2012 06:37:02AM *  2 points [-]

For a while now, I've been meaning to check out the code for this and heavily revise it to include things like data storage space, physical manufacturing capabilities, non-immediately-lethal discovery by humans (so you detected my base in another dimension? Why should I care, again?), and additional modes of winning. All of which I will get around to soon enough.

But, I'll tell you this. Now when I revise it, I am going to add a game mode where your score is in direct proportion to the amount of office equipment in the universe, with the smallest allowed being a functional paperclip. I am dead serious about this.

Comment author: TheOtherDave 24 May 2012 01:13:22PM 0 points [-]

(nods) Likely.

And I agree that humans might be able to detect attempts at deception in a system at that stage of its development. I'm not vastly confident of it, though.

Comment author: Polymeron 26 May 2012 06:01:19AM 0 points [-]

I have likewise adjusted down my confidence that this would be as easy or as inevitable as I previously anticipated. Thus I would no longer say I am "vastly confident" in it, either.

Still good to have this buffer between making an AI and total global catastrophe, though!

Comment author: Bugmaster 23 May 2012 09:13:22PM 2 points [-]

It is quite conceivable that all the necessary information is already out there, but we haven't been able to connect all the dots just yet.

Wouldn't this mean that creating fully functional self-replicating nanotechnology is just a matter of performing some thorough interdisciplinary studies (or meta-studies or whatever they are called) ? My impression was that there are currently several well-understood -- yet unresolved -- problems that prevent nanofactories from becoming a reality, though I could be wrong.

Comment author: Polymeron 24 May 2012 08:57:11AM 0 points [-]

The way I see it, there's no evidence that these problems require additional experimentation to resolve, rather than find an obscure piece of experimentation that has already taken place and whose relevance may not be immediately obvious.

Sure, that more experimentation is needed is probable; but by no means certain.

Comment author: TheOtherDave 23 May 2012 07:03:35PM 4 points [-]

depending on how seriously you take it, such analysis might still take place, hard and prolonged though it may be).

Yup, agreed that it might.
And agreed that it might succeed, if it does take place.

One can have "detectors" in place set to find specific behaviors, but these would have assumptions that could easily fail. Detectors that would still be useful would be macro ones - where it tries to access and how - but these would provide only limited insight into the AI's thought process.

Agreed on all counts.

Re: what the AI knows... I'm not sure how to move forward here. Perhaps what's necessary is a step backwards.

If I've understood you correctly, you consider "having a conversation" to encompass exchanges such as:
A: "What day is it?"
B: "Na ni noo na"

If that's true, then sure, I agree that the minimal set of information about humans required to do that is zero; hell, I can do that with the rain.
And I agree that a system that's capable of doing that (e.g., the rain) is sufficiently unlikely to be capable of effective deception that the hypothesis isn't even worthy of consideration.
I also suggest that we stop using the phrase "having a conversation" at all, because it does not convey anything meaningful.

Having said that... for my own part, I initially understood you to be talking about a system capable of exchanges like: A: "What day is it?"
B: "Day seventeen."
A: "Why do you say that?"
B: "Because I've learned that 'a day' refers to a particular cycle of activity in the lab, and I have observed seventeen such cycles."

A system capable of doing that, I maintain, already knows enough about humans that I expect it to be capable of deception. (The specific questions and answers don't matter to my point, I can choose others if you prefer.)

Comment author: Polymeron 24 May 2012 08:55:04AM 0 points [-]

My point was that the AI is likely to start performing social experiments well before it is capable of even that conversation you depicted. It wouldn't know how much it doesn't know about humans.

Comment author: semianonymous 25 April 2012 12:26:07PM 2 points [-]

It is not accusation or insult. It is the case though that the people in question (Luke, Eliezer) need to assume the possibility that people they are talking to are more intelligent than they are - something that is clearly more probable than not given available evidence - and they seem not to.

Comment author: Polymeron 23 May 2012 07:00:29PM 0 points [-]

I don't see how that would be relevant to the issue at hand, and thus, why they "need to assume [this] possibility". Whether they assume the people they talk to can be more intelligent than them or not, so long as they engage them on an even intellectual ground (e.g. trading civil letters of argumentation), is simply irrelevant.

Comment author: TheOtherDave 23 May 2012 04:39:47PM 1 point [-]

Yes, I agree that there's no guarantee that humans would behave as you describe.
Indeed, I don't find it likely.
But, sure, they might.

=== I agree that a stack trace can exist outside the AI's zone of control. What I was expressing skepticism about was that a system with even approximately human-level intelligence necessarily supports a stack trace that supports the kind of analysis you envision performing in the first place, without reference to intentional countermeasures.

By way of analogy: I can perform a structural integrity analysis on a bar of metal to determine whether it can support a given weight, but performing an equivalent analysis on a complicated structure comprising millions of bars of metal connected in a variety of arrangements via a variety of connectors using the same techniques is not necessarily possible.

But, sure, it might be.

======

I'd love to see a plausible line of thought by which the AI would try and succeed in deceiving humans at the get-go point without exposing its intentions.

Well, one place to start is with an understanding of the difference between "the minimal set of information about humans required to have a conversation with one at all" (my phrase) and "the most basic knowledge about humans" (your phrase). What do you imagine the latter to encompass, and how do you imagine the AI obtained this knowledge?

Comment author: Polymeron 23 May 2012 05:05:16PM *  0 points [-]

What I was expressing skepticism about was that a system with even approximately human-level intelligence necessarily supports a stack trace that supports the kind of analysis you envision performing in the first place, without reference to intentional countermeasures.

Ah, that does clarify it. I agree, analyzing the AI's thought process would likely be difficult, maybe impossible! I guess I was being a bit hyperbolic in my earlier "crack it open" remarks (though depending on how seriously you take it, such analysis might still take place, hard and prolonged though it may be).

One can have "detectors" in place set to find specific behaviors, but these would have assumptions that could easily fail. Detectors that would still be useful would be macro ones - where it tries to access and how - but these would provide only limited insight into the AI's thought process.

[...]the difference between "the minimal set of information about humans required to have a conversation with one at all" (my phrase) and "the most basic knowledge about humans" (your phrase). What do you imagine the latter to encompass, and how do you imagine the AI obtained this knowledge?

I actually perceive your phrase to be a subset of my own; I am making the (reasonable, I think) assumption that humans will attempt to communicate with the budding AI. Say, in a lab environment. It would acquire its initial data from this interaction.

I think both these sets of knowledge depend a lot on how the AI is built. For instance, a "babbling" AI - one that is given an innate capability of stringing words together onto a screen, and the drive to do so - would initially say a lot of gibberish and would (presumably) get more coherent as it gets a better grip on its environment. In such a scenario, the minimal set of information about humans required to have a conversation is zero; it would be having conversations before it even knows what it is saying. (This could actually make detection of deception harder down the line, because such attempts can be written off as "quirks" or AI mistakes)

Now, I'll take your phrase and twist it just a bit: The minimal set of knowledge the AI needs in order to try deceiving humans. That would be the knowledge that humans can be modeled as having beliefs (which drive behavior) and these can be altered by the AI's actions, at least to some degree. Now, assuming this information isn't hard-coded, it doesn't seem likely that is all an AI would know about us; it should be able to see some patterns at least to our communications with it. However, I don't see how such information would be useful for deception purposes before extensive experimentation.

(Is the fact that the operator communicates with me between 9am and 5pm an intrinsic property of the operator? For all I know, that is a law of nature...)

Comment author: Bugmaster 20 May 2012 08:57:24PM 4 points [-]

Yes, I understand what "exponential complexity" means :-)

It sounds, then, like you're on the side of kalla724 and myself (and against my Devil's Advocate persona): the AI would not be able to develop nanotechnology (or any other world-shattering technology) without performing physical experiments out in meatspace. It could do so in theory, but in practice, the computational requirements are too high.

But this puts severe constraints on the speed with which the AI's intelligence explosion could occur. Once it hits the limits of existing technology, it will have to take a long slog through empirical science, at human-grade speeds.

Comment author: Polymeron 23 May 2012 04:35:32PM 1 point [-]

Actually, I don't know that this means it has to perform physical experiments in order to develop nanotechnology. It is quite conceivable that all the necessary information is already out there, but we haven't been able to connect all the dots just yet.

At some point the AI hits a wall in the knowledge it can gain without physical experiments, but there's no good way to know how far ahead that wall is.

View more: Prev | Next