Less Wrong is a community blog devoted to refining the art of human rationality. Please visit our About page for more information.

In response to That Alien Message
Comment author: Mader_Levap 25 July 2016 07:32:57PM *  -1 points [-]

"I don't trust my ability to set limits on the abilities of Bayesian superintelligences."

Limits? I can think up few on the spot already.

Environment: CPU power, RAM capacity etc. I don't think even you guys claim something as blatant as "AI can break laws of physics when convenient".


  • Win this kind of situation in chess. Sure, AI would not allow occurence of that situation in first place during game, but that's not my point.

  • Make human understand AI. Note: uplifting does not count, since human then ceases to be human. As a practice, try teaching your cat Kant's philosophy.

  • Make AI understand itself fully and correctly. This one actually works on all levels. Can YOU understand yourself? Are you even theoretically capable of that? Hint: no.

  • Related: survive actual self-modification, especially without any external help. Transhumanist fantasy says AIs will do it all the time. Reality is that any self-preserving AI will be as eager to preform self-modification as you to get randomized extreme form of lobotomy (transhumanist version of Russian roulette, except with all bullets in every gun except one in gazilion).

I guess some people are so used to think about AI as magic omnipotent technogods they don't even notice it. Sad.

Comment author: Mader_Levap 25 July 2016 06:16:55PM -1 points [-]

I never seen anyone bragging about defeating strawmans so much. Hell, in one place he explicitly said about "Soul Swap World" that he made up on spot to happily destroy.

And I still do not know what I am supposed to think about personal identity. I happen to think ME is generated by brain. Brain that works so well it can generate mind despite all of those changes in atoms meticulously described by Yudkowsky.

Comment author: AlexM 16 July 2011 08:12:20PM *  15 points [-]

Why can't modern Nazis disavow ancient Nazi practice in favor of some true essence that makes sense in modern terms?

One can argue that holocaust denial is an attempt to bring nazism closer to modern ethical values. Real, authentic Nazis were proud of their achievement and would be outraged by thought that their successors would call them a lie.

Why not start your search for the true essence in Lord of the Rings

Some people do :-P

Comment author: Mader_Levap 18 September 2015 11:37:47AM -1 points [-]

Real, authentic Nazis were proud of their achievement

Not publicly. Holocaust denial exists since it (mass murdering of certain groups of humans) make them look bad. Of course, it is Insane Troll Logic, but I do not think anyone expects sane logic from Nazis.

Comment author: capybaralet 26 January 2015 06:02:25AM 4 points [-]

Which transhumanist ideas are "not even wrong"?

And do you mean simply 'not well specified enough'? Or more like 'unfalsifiable'?

You also seem to be implying that scientists cannot discuss topics outside of their field, or even outside its current reach.

My philosophy on language is that people can generally discuss anything. For any words that we have heard (and indeed, many we haven't), we have some clues as to their meaning, e.g. based on the context in which they've been used and similarity to other words.

Also, would you consider being cautious an inherently good thing?

Finally, from my experience as a Masters student in AI, many people are happy to give opinions on transhumanism, it's just that many of those opinions are negative.

Comment author: Mader_Levap 18 September 2015 10:22:53AM 1 point [-]

"Which transhumanist ideas are "not even wrong"?"

Technological Singularity, for example (as defined in Wikipedia). In my view, it is just atheistic version of Rapture or The End Of World As We Know It endemic in various cults and equally likely.

Reason for that is that recursive self-improvement is not possible, since it requires perfect self-knowledge and self-understanding. In reality, AI will be black box to itself, like our brains are black boxes to ourself.

More precisely, my claim is that any mind on any level of complexity is insuficient to understand itself. It is possible for more advanced mind to understand simpler mind, but it obviously does not help very much in context of direct self-improvement.

AI with any self-preservation instincts would be as likely to willingly preform direct self-modification to its mind as you to get stabbed by icepick through eyesocket.

So any AI improvement would have to be done old way. Slow way. No fast takeoff. No intelligence explosion. No Singularity.