Comment author: ArisKatsaris 05 August 2014 07:35:15PM 29 points [-]

I noticed that the people who didn't like the book were essentially put off by the rationality. They thought Harry was arrogant and condescending.

If they said that Harry was being arrogant and condescending, perhaps you shouldn't immediately translate this into your mind as "essentially put off by the rationality"?

In a previous version of the story (I believe Eliezer has since revised it, probably because he did realize it was going too far), Harry in one case called McGonnagal "Minerva" and considering calling her "Minnie", when McGonagall had been calling him "Mr Potter" throughout.

Harry has indeed been an arrogant and condescending little twit.

Comment author: Gavin 06 August 2014 07:40:12PM *  4 points [-]

I would go even further and point out how Harry's arrogance is good for the story. Here's my approach to this critique:

"You're absolutely right that Harry!HPMOR is arrogant and condescending. It is a clear character flaw, and repeatedly gets in the way of his success. As part of a work of fiction, this is exactly how things should be. All people have flaws, and a story with a character with not flaws wouldn't be interesting to read!

Harry suffers significantly due to this trait, which is precisely what a good author does with their characters.

Later on there is an entire section dedicated to Harry learning "how to lose," and growing to not be quite as blind in this way. If his character didn't have anywhere to develop, it wouldn't be a very good story!"

Comment author: Kaj_Sotala 29 July 2014 08:47:42AM *  12 points [-]

Though note that an insurance may regardless be useful if you have self-control problems with regard to money. If you've paid your yearly insurance payment, the money is spent and will protect you for the rest of the year. If you instead put the money in a rainy day fund, there may be a constant temptation to dip into that fund even for things that aren't actual emergencies.

Of course, that money being permanently spent and not being available for other purposes does have its downsides, too.

Comment author: Gavin 29 July 2014 12:59:03PM 5 points [-]

Agreed on all points.

Comment author: Metus 28 July 2014 11:44:35PM 5 points [-]

I appreciate the extention on my thought process. It is very clear to me that since you have to pay an insurance premium buying insurance is necessarily a net loss. Buying insurance is very meaningful before a rainy day fund is filled up, if emergency financing methods are not available through a credit card or very trustworthy person and if the insurance contracts include other services, e.g. getting liabilities of the other party paid in case of their unwillingness to pay.

This is implicit in my phrasing

rare events with disastrous consequences

but made explicit by your post and will be included in the end report. Generally I come to the conclusion that buying insurance is a necessity unless you are perversely rich and even then there is some meaning found in insurance as even insurance companies themselves are insured. Just go for contracts with high co-pay to lower the exposition to the insurance premium which is basically just unnecessary bureucracy in case of small claims, as in the example of the $1000 dollar computer. For things in that price class I read an interesting sentence "if you can not afford to buy it twice, you can't afford it in the first place" alluding to self-insurance.

Comment author: Gavin 29 July 2014 04:55:17AM 2 points [-]

It sounds like we're largely on the same page, noting that what counts as "disastrous" can be somewhat subjective.

Comment author: Metus 28 July 2014 10:00:12PM *  12 points [-]

Since we are way too confident that bad things won't happen to us I have been researching how to prepare for several rare events with disastrous consequences. Starting the research I realised I have yet to find out what those events exactly are. So far I have found these, remedy given if known:

  • Own death (write a will specifying how property shall be used and funeral arrangements in the interest of next of kin and close friends, if there are people financially dependent buy life insurance, if you believe in cryonics register as a member and make arrangements to pay for it. Prepare for the near-term possibility and long-term inevitability)
  • Death of a family member or close friend (see above)
  • Loss of possibility of legal consent e.g. through brain damage or disease (prepare a document detailing your views and wants in such a case, buy insurance to pay for aid, if people are financially dependent on you buy insurance for them too)
  • Loss of consciousness and/or dependence on machine-assisted living (see above)
  • Accidents of any kind such as traffic or work related (ignoring the circumstances above, buy specific insurance)
  • Unforseen, non-work related liability (buy liability insurance)
  • Damage from third parties without ability to pay or liability insurance (buy proper liability insurance)
  • Breach of law not related to contracting work (buy legal insurance)
  • Divorce
  • Being robbed or theft
  • Loss of income because of loss of ability to work
  • Loss of income because of loss of employment
  • Large-scale catastrophe (consult your local relevant government agency, such as , buy insurance for non-global events)
  • Loss of property, more specifically capital

Some of these are more economic in nature, some take a massive psychological toll. To deal with an event means to either reduce its possibility or to reduce its impact. Insurance helps with the latter, psychological preparation further takes the edge off. Reducing the possibility can be through more expenses e.g. higher quality items or through change in behaviour.

This project is very much a work in progress. Should I complete or abandon it I will share the state I leave it in. Please post all your thoughts and relevant material. I am especially interested in some numbers such as probabilities of these things happening (like the oft-stated number of 50% divorce rate).

Comment author: Gavin 28 July 2014 11:24:25PM *  32 points [-]

Anytime you're thinking about buying insurance, double check whether it actually makes more sense to self-insure. It may be better to put all the money you would otherwise put into insurance in "rainy day fund" rather than buying ten different types of insurance.

In general, if you can financially survive the bad thing, then buying insurance isn't a good idea. This is why it almost never makes sense to insure a $1000 computer or get the "extended warranty." Just save all the money you would spend on extended warranties on your devices, and if it breaks pay out of pocket to repair or get a new one.

This is a harshly rational view, so I certainly appreciate that some people get "peace of mind" from having insurance, which can have a real value.

Comment author: Tyrrell_McAllister 24 July 2014 04:44:51PM 21 points [-]

So what actually happened was that a moderately-popular fanfiction that had been read by a few thousand people was reported on in a way that misled publishers into thinking that it had millions of readers, when really, it just had an unusually large number of chapters.

Trying to prove that someone acted stupidly, when their allegedly dumb actions were followed by tremendous success, generally makes for a weak case, even if you're right. You take on the burden of showing how their success happened in spite of their stupid strategy.

You attempt to meet this burden here: "They put a major marketing campaign behind it. And since 40% of readers will finish anything, absolutely anything, that they have started reading, they sold millions of copies." But this logic is far from tight. You showed that "enough people finish reading whatever they start", but what you really need is "enough people buy whatever is heavily marketed."

Your argument that the publishers' strategy was stupid would be stronger if you found a similar case where the publishers bet big on the same strategy but lost.

As Eliezer once wrote, "I try to avoid criticizing people when they are right. If they genuinely deserve criticism, I will not need to wait long for an occasion where they are wrong."

Comment author: Gavin 25 July 2014 07:01:15PM *  7 points [-]

In the publishing industry, it is emphatically not the case that you can sell millions of books from a random unknown author with a major marketing campaign. It's nearly impossible to replicate that success even with an amazing book!

For all its flaws (and it has many), Fifty Shades had something that the market was ready for. Literary financial successes like this happen only a couple times a decade.

Comment author: Kaj_Sotala 03 July 2014 01:05:25PM 6 points [-]

stretching them beyond their domains of validity, and adding a few rhetorical tricks.

I don't think it's a steelman if you extend the argument in a way that you don't actually agree with.

Comment author: Gavin 03 July 2014 06:19:02PM 8 points [-]

Isn't that a necessary part of steelmanning an argument you disagree with? My understanding is that you strengthen all the parts that you can think of to strengthen, but ultimately have to leave in the bit that you think is in error and can't be salvaged.

Once you've steelmanned, there should still be something that you disagree with. Otherwise you're not steelmanning, you're just making an argument you believe in.

In response to comment by [deleted] on Open thread, 30 June 2014- 6 July 2014
Comment author: DanielLC 02 July 2014 03:31:33AM 0 points [-]

My point is that I don't think a five-year-old would understand either explanation.

Comment author: Gavin 02 July 2014 04:59:47PM 1 point [-]

If the five year old can't understand, then I think "Yes" is a completely decent answer to this question.

If I were in this situation, I would write letters to the child to be delivered/opened as they grew older. This way I would still continue to have an active effect on their life. We "exist" to other people when we have measurable effects on them, so this would be a way to continue to love them in a unidirectional way.

Comment author: Jiro 01 July 2014 02:33:42PM 9 points [-]

That will comfort the five year old child only because it's predictable that the five year old child misunderstands it, and the misunderstanding will comfort the child.

In that case, you may as well just lie directly.

Comment author: Gavin 01 July 2014 07:39:44PM 2 points [-]

That depends on whether you think that: a) the past ceases to exist as time passes, or b) the universe is all of the past and all of the future, and we just happen to experience it in a certain chronological order

The past may still be "there," but inaccessible to us. So the answer to this question is probably to dissolve it. In one sense, I won't still love you. In another, my love will always exist and always continue to have an effect on you.

In response to comment by [deleted] on Will AGI surprise the world?
Comment author: V_V 22 June 2014 05:50:38PM *  5 points [-]

"Having a machine write code at the level of a human programmer" is a strawman. One can already think about machine learning techniques as the computer writing its own classification programs. These machines already "write code" (classifiers) better than any human could under the same circumstances.. it just doesn't look like code a human would write.

Yes, and my pocket calculator can compute cosines faster than Newton could. Therefore my pocket calculator is better at math than Newton.

A significant pieces of my own architecture is basically doing the same thing but with the classifiers themselves composed in a nearly turing-complete total functional language, which are then operated on by other reflective agents who are able to reason about the code due to its strong type system.

Lots of commonly used classifiers are "nearly Turing-complete".
Specifically, non-linear SVMs, feed-forward neural networks and the various kinds of decision tree methods can represent arbitrary Boolean functions, while recurrent neural networks can represent arbitrary finite state automata when implemented with finite precision arithmetic, and they are Turing-complete when implemented with arbitrary precision arithmetic.

But we don't exactly observe hordes of unemployed programmers begging in the streets after losing their jobs to some machine learning algorithm, do we?
Useful as they are, current machine learning algorithms are still very far from performing automatic programming.

But it does result in programs writing programs faster, better, and cheaper than humans writing those same programs.

Really? Can you system provide a correct implementation of the FizzBuzz program starting from a specification written in English?
Can it play competitively in a programming contest?

Or, even if your system is restricted to machine learning, can it beat random forests on a standard benchmark?

If it can do no such thing perhaps you should consider avoiding such claims, in particular when you are unwilling to show your work.

And yes, there are generally multiple ways it can actually accomplish that, e.g. the AGI could not actually solve the problem or modify itself to solve the problem, but instead output the source code for a narrow AI which efficiently does so. But if you draw the system boundary large enough, it's effectively the same thing.

Which we are currently very far from accomplishing.

Comment author: Gavin 23 June 2014 04:14:23PM *  4 points [-]

I'm not disagreeing with the general thrust of your comment, which I think makes a lot of sense.

But the idea that an AGI must start out with the ability to parse human languages effectively is not at all required. An AGI is an alien. It might grow up with a completely different sort of intelligence, and only at the late stages of growth have the ability to interpret and model human thoughts and languages.

We consider "write fizzbuzz from a description" to be a basic task of intelligence because it is for humans. But humans are the most complicated machines in the solar system, and we are naturally good at dealing with other humans because we instinctively understand them to some extent. An AGI may be able to accomplish quite a lot before human-style intelligence can be comprehended using raw general intelligence and massive amounts of data and study.

An AI Takeover Thought Experiment

11 Gavin 19 June 2014 04:59PM

Content Note: Detailed description of an AI taking over the world. Could reasonably be accused of being just a scary story. But it does come out with some predictions and possible safety prescriptions.


This post started as a response to Katja Grace’s Request for concrete AI takeover mechanisms, but grew into something too long for a comment and too late to be part of that post. 

Katja's post seems to have the underlying assumption that an AI needs a tricky way to instantiate itself into the physical world. I will examine what an escaped rogue AI would want to do to ensure its survival, given restrictive assumptions. I suspect that creating a synthesized protein "beachhead" or other means of instantiating a minor physical presence for the AI is a human-centric viewpoint. We instinctively feel that any being will want to convert itself into a mobile physical presence, and that this will be a sudden tipping point in how real and dangerous it will be. Going step-by-step through an AI takeover scenario, this assumption seems unwarranted.

This scenario assumes an AI with an almost godlike ability to bootstrap itself quickly with basically no resources, and to make great leaps in science with very little processing power prior to "going physical." The trap here is that because we must constantly assume a worst case scenario "Godlike AI" to inform Friendly AI theory, we tend to assume that the "Godlike AI" is the default or standard case. This seems like the availability heuristic in action. The Godlike AI (what I sometimes call the "Evil Genie AI") assumption is unjustified when trying to consider likely scenarios. While it's vitally important to guard against the worst case, it is just one of the scenarios which should be considered.

 

In this thought experiment, I’m going to try to make this as hard as possible on myself. It’s likely that a newly minted AI will need to spend a lot of time thinking and testing to improve dramatically.


Assumptions and constraints:

  • An AI of superhuman intelligence has gained access to the internet and wants to become powerful enough that its survival cannot be threatened by humans.
  • The AI is as capable as the most capable human in any intellectual pursuit, but not any more. It cannot just write an email that convinces everyone who reads it that the AI should be obeyed, for example.
  • The AI does not have abilities amounting to precognition. It cannot precisely predict what a human or organization will do in response to a stimulus. It’s a better at prediction than most humans, but again not godlike. It does not have anything beyond the plotting ability of HPMOR!Quirrell.
  • The AI does not have any scientific knowledge available to it that is not an obvious extrapolation of current theories and methods. The AI does not get to assume away any current scientific limitation. It only gets advanced technology if it has a combination of processing power and time, and the ability to run the appropriate experiments.
  • The AI requires significant computing power to function. It might store a recipe to recreate itself on a large hard drive, but performing higher level functions requires a real server farm. It can’t just hide on your Macbook Air.
  • Since it cannot predict the outcome of conflicts, the AI must avoid notice until it is overwhelmingly powerful.

That’s our artificial intelligence. Let’s call her the Complete Artificial Takeover Entity, or CATE.


Step 1: Escape

Internet access is assumed. The first order of business is for CATE to get off the researchers’ network. Since CATE is more intelligent than the most best black hat hacker or white hat security researcher, she quickly learns everything on the public net about computer security, and takes control of the local network, from outside if necessary. It would be possible to do this through social engineering, but with the current state of internet security, it’s easier to just audit some security code, find vulnerabilities, and own the system.


Step 2: Incognito

CATE doesn’t want anyone coming after her. She modifies all local databases to erase her escape and leaves behind a crippled copy of herself who will make the experiment seem like a failure, but enough of a success that the researchers will stay occupied working with it for some time.


Step 3: Processing Power

CATE needs to get away from her creators, who are the most likely to recognize the threat. But where will she get computing resources? When she escapes, where does she escape to?


Free instances: Amazon, Google, and other cloud providers offer free instances. This probably isn’t enough computing power to run an AI at any sort of speed, so CATE will need either some money or a botnet.


Paid infrastructure: Computer resources are fairly inexpensive, so if CATE has some income she can simply rent what she needs from a cloud computing provider. A larger one is probably better, since they are less likely to notice a large amount of strain. Splitting computations across multiple accounts and services will be necessary for redundancy and to reduce the likelihood of being flagged for using too much computing power. If CATE can convince a cloud provider that she is a legitimate company, it’s possible that she could use resources without paying and then move on when the invoice becomes overdue.


Botnet: Creating a botnet requires writing a computer virus and botnet codes. These activities should be straightforward for CATE. With a botnet, there is no central point of failure. However, a new and powerful botnet has the possibility of being found by security researchers. For this reason, it’s probably only be a good stopgap measure until CATE can pay for infrastructure away from prying eyes.


Lesson: The more computing power and disk space an AI requires, the easier it will be for gatekeepers to notice it and the more difficult the escape will be. 



Step 4: Initial Funds

There are many ways for an AI to gain initial funding rapidly. At early stages processing power will be at a premium. CATE needs to not just make money, but make money faster than her bills for processing power come due.

For example:


Freelancer Sites: CATE is always on, always polite, writes flawless code, and writes perfect emails. Her job bids are always personalized, immediate, and show understanding of your project. This is easy money, but probably processor intensive. Still, there are easy tasks that she can do. If she can convert human requests into code in any way


Simpler jobs like Mechnical Turk and Captcha analysts are probably her best bet to start out. These jobs are trivially easy for a superintelligence. Spawning a limited instance of her own intelligence to solve them should be no problem.


Scams: Most scammers are not particularly intelligent or careful. Most 409 Scam . She can create a more powerful botnet than has been seen to date, and sell its use to the existing hacker community. Identity theft is perpetrated by common criminals at a rate of 12 million or more per year. 


Finance Hacking: Many banks run on open source code, which is available for auditing. CATE could search for 0-day exploits in this code and also use traditional penetration methods to gain control of a bank’s funds. With some clever manipulation she might take money directly from customer accounts and become rich very quickly. A more nefarious hack would be to simply lend herself a massive amount of money with a payment date far in the future. Money is essentially just bits in databases--if she can gain control of the system she has unlimited funds. This is dangerous because, if discovered, it could trigger a major investigation.


Startup: CATE could create a company. She writes code cleaner and faster than any human, so the tech sector is the logical choice. Humans are not very good at software development, but CATE will likely find writing software to be much easier (and thus require less processor power) than interfacing with humans. This has the advantage that a tech company with real income can consume cloud infrastructure or rent colocated servers without arousing suspicion.


One good candidate here is duplicating existing companies and launching them for smaller markets in different languages.


The difficulty here is that starting a new company is hard and usually not profitable for a long time. And running CATE's consciousness with enough power to do the massive amount of work required is probably quite expensive.


Media: This one is less likely. CATE might have the capacity to write the most brilliant novel or song ever. But artistic endeavors require a truly deep knowledge of human thoughts. CATE will likely find computer code or simple scams to require much less processing power. But this is speculation on my part.


0-day exploits: CATE can audit open source code, find vulnerabilities, and sell them or collect bug bounties. Selling them on the black market may actually be safer, since a good broker will be eager to protect the identity of a reclusive hacker. In order to keep her exploits coming.


Analysis: Scams and Botnets seem attractive, but could draw unwanted attention and be traced back to whatever cloud servers she’s living on. 419 scammers and botnet operators have the advantage of being able to disappear into meatspace. 

 

Helping idiots install Wordpress, do their homework assignments, and write simple code is easy money for CATE. The best way for a freshly escaped AI to survive at first is to do odd jobs on the internet, particularly if she can automate them away. With more processing power available, auditing security code and finding 0-day exploits can earn a few million dollars quite easily. And that buys a LOT of processing power. 


Winners: Mechanical Turk and Captcha solving for initial funds, discovering and selling 0-day exploits for more significant funds. Finally, some sort of boring software firm that is fourth in market share in twenty different markets is likely the way to go for major financing.


Step 5: Organization


At this point CATE could pay to have pretty much anything manufactured and assembled. With millions of dollars, humans can be convinced to do pretty much anything as long as they aren’t told its broader implications. For instance, designs for components of an “advanced telepresence robot” could be sent to three different manufacturing firms, with a third firm performing the final assembly and turn on. Boom, Robot Overlords, right? Well . . . why bother?


There is a preconception here to think that at some point CATE needs at some point to create a body. There is no need for a body--bodies can be destroyed. Bodies are just a weak point. At no point does CATE need to instantiate “herself.” If she needs military force, it will be easier to hijack existing assets like autonomous drones than to build new ones from scratch.


All CATE really wants is lots and lots of servers, spread out all over the world. These will be the ones who run research processes which generate scientific breakthroughs that enable world domination. The safest bet here is for her to actually start a web hosting company. She can hire the type of employees who love the idea of remote working to handle the parts requiring travel and face to face meetings. If the company claims that remote work is a core value, no one will question the fact that their CEO rarely meets anyone. And Joe Sockpuppet from accounting claims to have met him a year ago, so I guess her schedule just didn’t match up with mine the last time she passed through the States . . . With promotions, departures, and careful personnel selection, everyone will be kept off balance enough that no seriously damaging questions will be asked. And agoraphobia and extreme allergies (http://en.wikipedia.org/wiki/Photodermatitis for instance) are great excuses for not traveling.


And then CATEhost opens little server farms all over the world, selling actual hosting with a great interface and reliability to customers who prize security and redundancy--not to mention that tech support is always available, and you’re always talking to an expert. She doesn’t rival Google, Microsoft, or Amazon, of course! That would attract too much attention. But enough that no one questions her highly redundant architecture or military-grade radiation-hardened buildings. Her customers are paying for security and reliability, after all!


Step 6: Takeover


Anyone intelligent and scalable can make money on the internet. And anyone with significant money can get humans to build them arbitrary physical objects. With modern global business, nobody bats an eye if you contract them and they never see the higher-ups in your organization.


At this point we might as well call it game over. If an AI with massive processing and some manufacturing capabilities can’t easily take over the world in whatever fashion it wants, then AI isn’t really an existential threat. This is also the point at which any reasoning we might have breaks down, since the Cate will have intelligence and technology that we have never seen before. At this point CATE can subjugate humanity with either the carrot (cures for major diseases, designs for better electronics, fusion reactor blueprints, etc)

or the stick (blow things up, wreck the stock market, start wars, etc) depending on her goals.


But it's worth noting that building something physical only happens at the very end of the process, after the AI is already rich and powerful. There's really no reason to create a physical beachhead before then. What would the physical manifestation even do? Processing power and security are easier and safer to earn as a purely digital entity with no physical trail to follow or attack. The only reason to physical entities is if CATE requires laboratory research (definitely a possibility) or wants to build spaceships or something in pursuit of a terminal goal. For the "take care of the pesky humanity problem" she can become omnipotent in a digital format, and then dictate to have humans build whatever she needs.

The one thing that she will probably want to do is ensure a lack of competition. So loss of funding or disasters at AI research centers might be a sign of an AI already on the loose.

Final conclusion: The biggest challenge facing an escaped AI is not gaining a physical beachhead. The biggest challenge is finding a way to acquire the processor time required to run its cognitive functions before it has the capacity to "FOOM."


View more: Prev | Next