All of Tiiba3's Comments + Replies

Tiiba320

"""it is quite likely that the universe may not be able to support vast orders of magnitude of intelligence"""

Why?

Tiiba300

A question about Andrew Ng, who was mentined in this thread. Is that his real name?

Tiiba300

So, like, every bacterium is its own species?

Tiiba300

There was a blank header? I didn't notice.

Tiiba300

"""What about from aligning the dots along the lines of the image?"""

Wouldn't you need to find them first?

Tiiba300

So who was the congressman?

Tiiba300

Grant: group selection does happen, but only very slowly. Natural selection works when its units are destroyed, and tribes go extinct pretty rarely compared to individuals.

Merely being poor does not make a selection unit unfit, as far as evolution is concerned. It has to disappear.

Tiiba3120

Well, a picture of a zebra is real.

And you'll probably agree that the merely real is, in some ways, in need of improvement, which is the whole point of transhumanism.

-7[anonymous]
Tiiba370

"""Obert: "A duty is something you must do whether you want to or not." """

Obey gravity. It's your duty!

--Obert

Tiiba300

I know that random behavior requires choices. The machine IS choosing - but because all choices are equal, the result of "max(actionList)" is implementation-dependent. "Shut down OS" is in that list, too, but "make no choice whatsoever" simply doesn't belong there.

Tiiba300

Let's say I have a utlity function and a finite map from actions to utilities. (Actions are things like moving a muscle or writing a bit to memory, so there's a finite number.)

One day, the utility of all actions becomes the same. What do I do? Well, unlike Asimov's robots, I won't self-destructively try to do everything at once. I'll just pick an action randomly.

The result is that I move in random ways and mumble gibberish. Althogh this is perfectly voluntary, it bears an uncanny resemblance to a seizure.

Regardless of what else is in a machine with such a utility function, it will never surpass the standard of intelligence set by jellyfish.

Tiiba3170

"we could imagine that "sexiness" starts by eating an Admirer"

Harsh, but fair.

Tiiba300

Julian, I think the box you're not opening is Pandora's box.

Tiiba300

Virge is mixing up instrumental and terminal values. No biscuit.

Tiiba310

An AI could screw us up just by giving bad advice. We'll be likely to trust it, because it's smart and we're too lazy to think. A modern GPS receiver can make you drive into a lake. An evil AI could ruin companies, start wars, or create an evil robot without lifting a finger.

Besides, it's more fun to create FAI and let it do what it wants than to build Skynet and then try to confine it forever. You'll still have only one chance to test it, whenever you decide to do that.

Tiiba310

I seem to be unable to view the referenced comment.

Hmm, no replies after all this time?

Tiiba370

A thought occured to me: people who are offended by the idea that a mere machine can think simply might not be imagining the right machine. They imagine maybe a hundred neurons, each extending 10-15 synapses to the others. And then they can't make head or tail of even that, because it's already too big. Scope insensitivity, in other words.

Tiiba320

"So I can imagine another math in which 2+2=5 is not obviously false, but needs a long proof and complicated equations..."

So, from the fact that another mind might take a long time to understand integer operations, you conclude that it has "another math"? And what does that mean for algorithms?

If an intelligence is general, it will be able to, in time, understand any concept that can be understood by any other general or narrow intelligence. And then use it to create an algorithm. Or be conquered.

1bigjeff5
Think of binary arithmetic verses decimal arithmetic verses hexadecimal arithmetic. Certain things in each of these arithmetics are extremely easy compared to the other two. For example, in binary, multiplying by 2 is absurdly easy, but multiplying by 10 is much harder. Multiplying by 16 is actually slightly easier than 10, as there are some cool tricks that apply between the two sets. In decimal, multiplying by 10 is never hard, no matter how big the number. Multiplying by 2 can be hard if the number is big enough, but it's still pretty easy. Multiplying by 16 takes some mental gymnastics right from the get-go (well, for most people anyway). You see the pattern, so I won't do hex. Basic floating point arithmetic is quite easy in decimal, but doing this in binary is significantly more difficult and often results in non-terminating numbers, or even non-real numbers akin to dividing one by three or pi in decimal. 10.06 might look nice and clean in decimal, but it's a nightmare in binary. The net result in computer science is that you have to be very, very careful with binary rounding errors, since almost every floating point calculation is going to require rounding for most numbers. And that's just starting with a different number of digits on your hands. Imagine if you looked at the world in a completely different way than we do, what would math look like? The physics wouldn't change, but perhaps calculus is as easy as addition is to us, but subtraction requires 8 years of schooling to wrap your head around. What if Martians could follow the movements of electrons, but couldn't tell that their fingers, thumb, and palm were the same thing as their hand? What would their math look like then?
Tiiba310

Latanius:

"Tiiba: an algorithm is a model in our mind to describe the similarities of those physical systems implementing it. Our mathematics is the way we understand the world... I don't think the Martians with four visual cortexes would have the same math, or would be capable of understanding the same algorithms... So algorithms aren't fundamental, either."

One or more of us is confused. Are you saying that a Martian wiith four visual cortices would be able to compress any file? Would add two and two and get five?

They can try, sure, but it won't work.

Tiiba3-10

Please delete my post. I see that Tom said that already.

Tiiba320

mtraven:

"That suggests that the algorithm itself is not a physical thing, but something else. And those something elses have very little to do with the laws of physics."

An algorithm can exist even without physics. It's math.

Tiiba330

I'm pretty confused by this discussion. People toss out terms like reductionist or anti-reductionist, and I can't even tell what they disagree about.

Here's what I know:

1) There are quarks and electrons, maybe some strings too. Nobody seems to dispute the quarks and electrons, at least. There are also clusters of particles.

2) Everything above that level is an abstraction that only exists in our heads. Yeah, those atoms really are near each other, but the only thing that makes them a "computer" is that we use them for computing. Same applies to bra... (read more)