All of Dagon's Comments + Replies

Dagon
10

Interesting, but I worry that the word "Karma" as a label for a legibly-usable resource token makes it VERY different from common karma systems on social websites, and that the bid/distribute system is even further from common usage.

For the system described, "karma" is a very misleading label.  Why not just use "dollars" or "resource tokens"?

Dagon
20

The rabbit hole can go deep, and probably isn't worth getting too fancy for single-digit hosts.  Fleets of thousands of spot instances benefit from the effort.  Like everything, dev-time vs runtime-complexity vs cost-efficiency is a tough balance.

When I was doing this often, I had different modes for "dev mode, which includes human-timeframe messing about" and "prod mode", which was only for monitored workloads.   In both cases, automating the "provision, spin up, and initial setup", as well as the "auto-shutdown if not measurably used for N minutes (60 was my default)" with a one-command script made my life much easier.

Dagon
40

I've seen scripts (though I don't have links handy) that do this based on no active logins and no CPU load for X minutes as well.  On the other tack, I've seen a lot of one-off processes that trigger a shutdown when they complete (and write their output/logs to S3 or somewhere durable).  Often a Lambda is used for the control plane - it responds to signals and runs outside the actual host.

2jefftk
I like this idea a lot, but I'm nervous about setting the right CPU threshold. Too low and it never shuts off, too high and it shuts down in the middle of something when waiting for a slow download. But possibly if I looked at load logs I'd see it's so clearly either ~zero or >>zero that it's not fussy?
Dagon
30

There's a big presumption there.  If he was a p-zombie to start with, he still has non-experience after the training.  We still have no experience-o-meter, or even a unit of measure that would apply.

For children without major brain abnormalities or injuries, who CAN talk about it, it's a pretty good assumption that they have experiences.  As you get more distant from your own structure, your assumptions about qualia should get more tentative.

Dagon
20

Do you think that as each psychological continuations plays out, they'll remain identical to one another?

They'll differ from one another, and differ from their past singleton self.  Much like future-you differs from present-you.  Which one to privilege for what purposes, though, is completely arbitrary and not based on anything.  

Which psychological stream one-at-the-moment-of-brain-scan ends up in is a matter of chance.

I think this is a crux.  It's not a matter of chance, it's all of them.  They all have qualia.  They all hav... (read more)

Dagon
30

Reminder to all: thought experiments are limited in what you can learn.  Situations which are significantly out-of-domain for our evolved and trained experiences simply cannot be analyzed by our intuitions.  You can sometimes test a model to see if it remains useful in novel/fictional situations, but you really can't trust the results.

For real decisions and behaviors, details matter.  And thought experiments CANNOT provide the details, or they'd be just situations, not hypotheticals.

Dagon
20

Once we identify an optimal SOA

This is quite difficult, even without switching costs or fear of change.   Definition of optimal is elusive, and most SOA have so many measurable and unmeasurable, correlated and uncorrelated factors to them that comparison is not directly possible.

Add to this the common moral beliefs (incorrect IMO, but still very common) of "inaction is less blameworthy than wrong action, and only slightly blameworthy compared to correct action", and there needs to be a pretty significant expected gain from switching in order to undert... (read more)

Dagon
2-2

Wow, a lot of assumptions without much justification

Let's assume computationalism and the feasibility of brain scanning and mind upload. And let's suppose one is a person with a large compute budget. 

Already well into fiction.  

But one is not both. This means that when one is creating a copy one can treat it as a gamble: there's a 50% chance they find themselves in each of the continuations. 

There's a 100% chance that each of the continuations will find themselves to be ... themselves.  Do you have a mechanism to designate one as the "t... (read more)

1james oofou
Do you think that as each psychological continuations plays out, they'll remain identical to one another? Surely not. They will diverge. So although each is itself, each is a psychological stream distinct from the other, originating at the point of brain scanning. Which psychological stream one-at-the-moment-of-brain-scan ends up in is a matter of chance. As you say, they are all equally "true" copies, yet they are separate. So, which stream one ends up in is a matter of chance or, as I said in the original post, a gamble.  Think of it like this: if one had one continuation in which one lived a perfect life, one would be guaranteed to live that perfect life. But if one had 10 copies in which one lived a perfect life, one does benefit at all. It's the average that matters. But one is deciding how to use one's compute at time t (before any copies are made). Ones at time t is under no obligation to spend one's compute on someone almost entirely unrelated to one just because that person is perhaps still technically oneself. The "once they diverge" statement is beside the point - the decision is made prior to the divergence.  I go into more detail in a post on my Substack (although it's perhaps a lot less readable, and I still work from similar assumptions, and one would be best to read the first post in the series first). 
Dagon
20

This is a topic where macro and micro have a pretty big gap.  

If you're asking about measured large-group unemployment, you probably don't get very good causality from any given change, and there's no useful, simple model of the motivations and frictions of potential-employeers and potential-employees.  It's a very complicated matching market.

If you're asking about some specific reasons that an individual may be out of work or become out of work, you'll get a lot better result and some concrete reasons.   But everyone you talk to will say "t... (read more)

Dagon
20

I don't understand the question.  What intuition for not smoking are you talking about?  CDT prefers smoking.   Are you asking why EDT abstains from smoking?  I'm not the best defender, as I don't really think EDT is workable, but as I understand it EDT updates it's world state based on actions, meaning that it prefers the world where you don't have the lesion and don't WANT to smoke.

Dagon
20

The first one is only a metaphor - it's not possible now, and we don't know if it ever will be (because we don't know how to scan a being in enough detail to recreate it well enough).    

The second one is WAY TOO limited.  If you put a radio anywhere near your head, or really any other-controlled media, you can be programmed.  By trivial extension, you have been programmed.  Get used to it.

Dagon
30

Economists and other social theorists often take the concept of utility for granted.

Armchair economists and EAs even more so.  Take for granted, and fail to document WHICH version of the utility concept they're using.   

For me, utility is a convenient placeholder for the underlying model that our ordinal preferences expressed through action (I did X, meaning I prefer the expected sum of value of outcomes likely from X).  Utility is the "value" that is preferred.  Note that it's kind of a circular defining - it's the thing that driv... (read more)

Dagon
20

I think it's a different level of abstraction.  Decision theory works just fine if you separate the action of predicting a future action from the action itself.  Whether your prior-prediction influences your action when the time comes will vary by decision theory.

I think, for most problems we use to compare decision theories, it doesn't matter much whether considering, planning, preparing, replanning, and acting are correlated time-separated decisions or whether it all collapses into a sum of "how to act at point-in-time".  I haven't seen much detailed exploration of decision theory X embedded agents or capacity/memory-limited ongoing decisions, but it would be interesting and important, I think.

Dagon
20

Decision theory is fine, as long as we don't think it applies to most things we colloquially call "decisions".   In terms of instantaneous discrete choose-an-action-and-complete-it-before-the-next-processing-cycle, it's quite a reasonable topic of study.

2cubefox
A more ambitious task would be to come up with a model that is more sophisticated than decision theory, one which tries to formalize your previous comment about intent and prediction/belief.
Dagon
20

But if you only have a belief that you will do something in the future, you still have to decide, when the time comes, whether to carry out the action or not. So your previous belief doesn't seem to be an actual decision, but rather just a belief about a future decision -- about which action you will pick in the future

Correct.  There are different levels of abstraction of predictions and intent, and observation/memory of past actions which all get labeled "decision".    I decide to attend a play in London next month.  This is an intent ... (read more)

2cubefox
That's an interesting perspective. Only it doesn't seem fit into the simplified but neat picture of decision theory. There everything is sharply divided between being either a statement we can make true at will (an action we can currently decide to perform) and to which we therefore do not need to assign any probability (have a belief about it happening), or an outcome, which we can't make true directly, that is at most a consequence of our action. We can assign probabilities to outcomes, conditional on our available actions, and a value, which lets us compute the "expected" value of each action currently available to us. A decision is then simply picking the currently available action with the highest computed value. Though as you say, such a discretization for the sake of mathematical modelling does fit poorly with the continuity of time.
Dagon
40

When the decision is made, consideration ends. The action must be wholehearted in spite of uncertainty.

This seems like hyperbolic exhortation rather than simple description.  This is not how many decisions feel to me - many decisions are exactly a belief (complete with bayesean uncertainty).  A belief in future action, to be sure, but it's distinct in time from the action itself.  

I do agree with this as advice, in fact - many decisions one faces should be treated as a commitment rather than an ongoing reconsideration.  It's not actuall... (read more)

2cubefox
But if you only have a belief that you will do something in the future, you still have to decide, when the time comes, whether to carry out the action or not. So your previous belief doesn't seem to be an actual decision, but rather just a belief about a future decision -- about which action you will pick in the future. See Spohn's example about believing ("deciding") you won't wear shorts next winter:
2Richard_Kennaway
It is exhortation, certainly. It does not seem hyperbolic to me. It is making the same point that is illustrated by the multi-armed bandit problem: once you have determined which lever gives the maximum expected payout, the optimum strategy is to always pull that lever, and not to pull levers in proportion to how much they pay. Dithering never helps. Yes. But only as such changes come to be. Certainly not immediately on making the decision. "Commitment" is not quite the concept I'm getting at here. It's just that if I decided yesterday to do something today, then if nothing has changed I do that thing today. I don't redo the calculation, because I already know how it came out.
Answer by Dagon
40

I only see one downvoted post, and a bunch of comments and a few posts with very low voting at all.  That seems pretty normal to me, and the advice of "lurk for quite a bit, and comment occasionally" is usually good for any new users on any site.

8lsusr
While this is true, I applaud Oxidize for learning the fast way. Most users of this site do only the "lurk for quite a bit", and never attempt to write great top-level posts. Ultimately, there is no harm done by crashing and burning a few times—as long as you're nice about it (which Oxidize has been).
Dagon
20

A lot depends on what you mean by "required", and what specific classes or functions you're talking about.  The core skill of committing a position to writing and supporting it with logic is never going away.   It will shift from "do this with minimal spelling and grammar assistance" to "ensure that the prompt-review-revise loop generates output you can stand behind".   

This is already happening in many businesses and practical (grant-writing) aspects of academia.  It'll take a while for undergrad and MS programs to admit that their academic theories of what they're teaching needs revision.

1Anders Lindström
Oh, I mean "required" as in to get a degree in a certain subject you need to write a thesis as your rite of passage.  Yes, you are right. Adept or die. AI can be a wonderful tool for learning but as it is used right now, where everyone have to say that they don´t use it, it beyond silly. I guess there will be some kind of reckoning soon.
Dagon
50

This seems generally applicable.  Any significant money transaction includes expectations, both legible and il-, which some participants will classify as bullshit.  Those holding the expectations may believe it to be legitimately useful, or semi-legitimately necessary due to lack of perfect alignment.

If you want to specify a bit, we can probably guess at why it's being required.

2Kabir Kumar
What I liked about applying for VC funding was the specific questions.  "How is this going to make money?" "What proof do you have this is going to make money" and it being clear the bullshit that they wanted was numbers, testimonials from paying customers, unambiguous ways the product was actually better, etc. And then standard bs about progress, security, avoiding weird wibbly wobbly talk, 'woke', 'safety', etc.  With Alignment funders, they really obviously have language they're looking for as well, or language that makes them more and less willing to put more effort into understanding the proposal. Actually, they have it more than the VCs. But they act as if they don't. 
Dagon
*30

[Note: I apologize for being somewhat combative - I tend to focus on the interesting parts, which is those parts which don't add up in my mind.  I thank you for exploring interesting ideas, and I have enjoyed the discussion! ]

I was only saying that I don't see anything proving it won't work

Sure, proving a negative is always difficult.  
 

 I agree that this missile problem shouldn't happen in the first place. But it did happen in the past

Can you provide details on which incident you're talking about, and why the money-bond is the problem ... (read more)

1Knight Lee
Thank you for saying that. I thought about it a bit more, and while I still think it's possible in theory, I agree it's not that necessary.[1] When a country shoots down a civilian airliner, it's usually after they repeatedly sent it a warning, but the pilots never heard it. It's more practical to fix this problem rather than have the money system. Maybe a better solution would be a type of emergency warning signal that all airplanes can hear, even if they accidentally turned their radio off. There may be a backup receiver or two which is illegal to turn off, and only listens to such warnings. That would make it almost impossible for the pilots to ignore the warnings. 1. ^ I still think they money system might be useful for preventing automated militaries from waging accidental war, but that's another story.
Dagon
30

I've been in networking long enough to know that "can be less than", "often faster", and "can run" are all verbal ways of saying "I haven't thought about reliability or measured the behavior of any real systems beyond whole percentiles."

But really, I'm having trouble understanding why a civilian plane is flying in a war zone, and why current IFF systems can't handle the identification problem of a permitted entry.

 

1Knight Lee
Thank you. I admit you have more expertise in networking than me. It is indeed just a new idea I thought of today, not something I've studied the details of. I have nothing proving it will work, I was only saying that I don't see anything proving it won't work. Do you agree with this position? Maybe there will be technical issues preventing this system from moving information as fast as a video call, but maybe it can be fixed, right? I agree that this missile problem shouldn't happen in the first place. But it did happen in the past, so the idea might help. It's not the same thing as current IFF. From what I know, IFF can prove who's side you are on but not whether you are military or civilian. From an internet search, I read that Iran once disguised their military jets as civilian, which contributed to the disaster of Iran Air Flight 655. A civilian aircraft might be given permission in the form of a password, but there's nothing stopping a country from sharing that password with military jets. Also, if a civilian airliner is flying over international waters but gets too close to another country's ships, it might not have permission.
Dagon
30

Kind of unfortunate that a comms or systems latency destroys civilian airliners.  But nice to live in a world where all flyers have $10B per missile/aircraft pair lying around, and everyone trusts each other enough to hand it over (and hand it back later).  

1Knight Lee
Latency Think about video calls you had with someone on the other side of the world, you don't notice that much latency. The an internet signal can travel from the US to Australia and back again in less than 0.2 seconds, and is often faster than 80% the speed of light (fibre optic ping statistics). Computers are very fast: a lot of computer programs can run a million times a second (in sequence). $10 billion lying around There isn't $10 billion for each missile/aircraft pair, there is only one per alliance of countries, and it's only used when a missile asks a civilian airliner for a password. Maybe it's not cryptocurrency but another form of money, in which case it can be part of a foreign exchange reserve (rather than money set aside purely for this purpose). Trust Yes, there is no guarantee the country taking the money will hand it back. But if they are willing to "accidentally" launch a missile at a civilian airliner, and ransom it for $10 billion, and keep the money, they will be seen as a terrorist state. The world operates under the assumption that you can freely sail ships and fly airplanes, without worrying about another country threatening to blow them up and demanding ransom money. If you do not trust a country enough to pay their missile and save your civilian airliner, you should keep all your civilian aircraft and ships far far out of their reach. You should pull out any money you invested in them since they'll probably seize that too.
Dagon
-30

Sure.  There's lots of things that aren't yet possible to collect evidence about.  No given conception of God or afterlife options has been disproven.  However, there are lots of competing, incompatible theories, none of which have any evidence for or against.  Assigning any significant probability (more than a percent, say) to any of them is unjustified.  Even if you want to say 50/50 that some form of deism will be revealed after death, there are literally thousands of incompatible conceptions of how that works. And near-infinint... (read more)

Dagon
2-2

I didn't downvote this, because it seems good-faith and isn't harmful.  But I really dislike this "friendly" style of writing, and it doesn't fit well on lesswrong.  It's very hard to find things that are concrete enough to understand whether I disagree or not.  Rhetorical questions (especially that you don't answer) really detract from understanding your POV.  Some specifics:

But most of us patch together a little of this and a little of that and try to muddle through with a philosophy that’s something of a crazy quilt.

Citation needed. ... (read more)

2David Gross
I assume the audience here is a mix of sophisticated people who of course know all about the trolley problem, etc., and newbies who are attracted to rationalism or the LW ethos and are here to learn more about stuff. So I write in a mix of modes. I can't say I'm confident about how I navigate this... it's just kind of a gut feeling that there's room for multiple styles. As for your first point about "...crazy quilt," I expand on this later in the essay when I discuss how responses to the trolley problems show that commonly people sometimes lean on deontological reasoning, sometimes on consequentalist reasoning. For the second point, I think my "so from one perspective" caveat anticipates your objection. If you are first confronted with the lever-pulling scenario and think "well, this is just a matter of simple mathematics," the second scenario reminds you that there are other factors to consider. For the third point, congratulations on having an existentialist perspective on this matter, but I'm confident that this is far from universal.
Dagon
83

This would be a lot stronger if it acknowledged how few lies have the convenient fatal flaw of a chocolate allergy.  Many do, and it's a good overall process, but it's nowhere near as robust as implied.

Note that I disagree that it's not applicable when you don't already suspect deception - it's useful to look for details and inconsistency when dealing with any fallible source of information - doesn't matter whether it's an intentional lie, or a confused reporter, or an inapplicable model, truth the only thing that's consistent with itself and with observations.  

5ymeskhout
The example was intended to be unrealistically convenient, since the goal there was just an illustrative example. Had I used an actual lie narrative from one of my clients (for example) it would've gotten very convoluted and wordy, and more likely to confuse the reader. I acknowledge there are limitations when you're dealing with unknowable lies. Beyond that, it was really hard to figure out how rare "lies with convenient flaws" really are. I don't know what denominator I'd use (how many lies are in the universe? which ones count?) or how I'd calculate the numerator.
Dagon
104

This is a fundamental truth for all commodities and valuable things.  They're fungible, but not positionally identical, and not linearly aggregable.  This is why we prefer to talk about "utility" over "quantity" in game theory discussions.

Market cap is meaningful in some sense - the price in a liquid market isn't just randomly the last price used, it's the equilibrium price of a marginal share.  That's the price that current holders don't want to sell for less, and people with money don't want to buy or more.  That equilibrium is real i... (read more)

5purple fire
This is untrue. Order book data is available for the right price (and although it doesn't capture all capital flow, sophisticated algorithms can infer the full curves with high accuracy). For what it's worth, many finance firms view the market price as basically noise and use order book curves as the "real" price input. Most contracts that depend on equity value will calculate it by VWAP over the last 30 or 90 trading days, which is a much better estimate of true value.
Dagon
30

"something like that" isn't open enough.  "or something else entirely" seems more likely than "something like that".   Many more than 2 groups (family-sized coalitions) is an obvious possibility, but there are plenty of other strategies used by primitive malthusian societies - infanticide being a big one, and ritual killings being another.  According to Wikipedia, Jared Diamond suggests cannibalism for Rapa Nui.  

Looking at Wikipedia (which I should have done earlier), there's very little evidence for what specific things changed during the collapse.

 In any case, it's tenuous enough that one shouldn't take any lessons or update your models based on this.

Dagon
20

In the medium-term reduced-scarcity future, the answer is:  lock them into a VR/experience-machine pod. 

edit: sorry, misspoke.  In this future, humans are ALREADY mostly in these pods.  Criminals or individuals who can't behave in a shared virtual space simply get firewalled into their own sandbox by the AI.  Or those behaviors are shadowbanned - the perpetrator experiences them, the victim doesn't. 

Dagon
20

I nominate NYC, and I assert that LA is an inferior choice for this.  Source: John Carpenter/Kurt Russel movies.
 

Dagon
50

In a sufficiently wealthy society we would never kill anyone for their crimes.

In a sufficiently wealthy society, there're far fewer forgivable/tolerable crimes.  I'm opposed to the death penalty in current US situation, mostly for knowledge and incentive reasons (too easy to abuse, too hard to be sure).  All of the arguments shift in weight by a lot if the situation changes.  If the equilibrium shifts significantly so that there are fewer economic reasons for crimes, and fewer economic reasons not to investigate very deeply, and fewer economic reasons not to have good advice and oversight, there may well be a place for it.

Dagon
72

This was my thinking as well.  On further reflection, and based on OP's response, I realize there IS a balance that's unclear.  The list contains some false-positives.  This is very likely just by the nature of things - some are trolls, some are pure fantasy, some will have moved on, and only a very few are real threats.  

So the harm of making a public, anonymous, accusation and warning is definitely nonzero - it escalates tension for a situation that has passed.  The harm of failing to do so in the real cases is also nonzero, but ... (read more)

Dagon
4120

Can you explore a bit more about why you can't ethically dump it on the internet?  From my understanding, this is information you have not broken any laws to obtain, and have made no promises as to confidentiality.

If not true publication, what keeps you from sending it to prosecutors and police?  They may or may not act, but that's true no matter who you give it to (and true NOW of you).  

With regards to dumping the info on the internet, the files by definition contain extensive personal identifable information about people, names, addresses, photos, social media links often alongside allegations of their alleged crimes ranging such as infidelity, child abuse and financial fraud.

I can rarely substantiate these, and know for a fact based on the investigated cases that such allegations are often completely fabricated in order to frame the user's request for violence as more morally justified. I don't think it's fair to publish such informatio... (read more)

Dagon
30

People who have a lot of political power or own a lot of capital, are unlikely to be adversely affected if (say) 90% of human labor becomes obsolete and replaced by AI.

That's certainly the hope of the powerful.  It's unclear whether there is a tipping point where the 90% decide not to respect the on-paper ownership of capital.

so long as property rights are enforced, and humans retain a monopoly on decisionmaking/political power, such people are not-unlikely to benefit from the economic boost that such automation would bring.

Don't use passive voice for... (read more)

1rvnnt
Yes. Also unclear whether the 90% could coordinate to take any effective action, or whether any effective action would be available to them. (Might be hard to coordinate when AIs control/influence the information landscape; might be hard to rise up against e.g. robotic law enforcement or bioweapons.) Good point! I guess one way to frame that would be as And yeah, that seems very difficult to predict or reliably control. OTOH, if someone were to gain control of the AIs (possibly even copies of a single model?) that are running all the systems, that might make centralized control easier? </wild, probably-useless speculation>
Dagon
20

Specifically, "So, the islanders split into two groups and went to war." is fiction - there's no evidence, and it doesn't seem particularly likely.

1Zero Contradictions
Thanks for commenting. However, he also wrote in the same paragraph: He wrote "or something like that", so I think that allows some variation of two (main) groups fighting each other in a war. He gave his reasoning for why individuals would team into larger groups in the previous paragraph, but I will agree that it's mostly speculative how many warring groups there were. Regardless, I'm convinced that the island's environmental degradation and population collapse were both most likely caused by overpopulation.
Dagon
30

Well, there are possible outcomes that make resources per human literally infinite.  They're not great either, by my preferences.

In less extreme cases, a lot depends on your definition of "poverty", and the weight you put on relative poverty vs absolute poverty.  Already in most parts of the world the literal starvation rate is extremely low.  It can get lower, and probably will in a "useful AI" or "aligned AGI" world.  A lot of capabilities and technologies have already moved from "wealthy only" to "almost everyone, including technically impoverished people", and this can easily continue.  

 

Answer by Dagon
20
  1. There's a wide range of techniques and behaviors that can be called "hypnosis", and an even wider range of what can be called "a real thing, right?".  Things in the realm of hypnosis (meditation, guided-meditation, self-hypnosis, daily affirmations, etc. have plenty of anecdotal support from adherents, and not a lot of RCTs or formal proof of who it will work for and who it won't.
  2. There's a TON of self-help and descriptive writing on the topics of meditation and self-hypnosis.  For many people, daily affirmations seem to be somewhat effective in changing their attitude over time.  For many, a therapist or guide may be helpful in setting up and framing the hypnosis.  
Dagon
30

What does "unsafe" mean for this prediction/wager?  I don't expect the murder rate to go up very much, nor life expectancy to reverse it's upward trend.  "Erosion of rights" is pretty general and needs more specifics to have any idea what changes are relevant.

I think things will get a little tougher and less pleasant for some minorities, both cultural and skin-color.  There will be a return of some amount of discrimination and persecution.  Probably not as harsh as it was in the 70s-90s, certainly not as bad as earlier than that, but wo... (read more)

3artifex0
So, the current death rate for an American in their 30s is about 0.2%. That probably increases another 0.5% or so when you consider black swan events like nuclear war and bioterrorism. Let's call "unsafe" a ~3x increase in that expected death rate to 2%. An increase that large would take something a lot more dramatic than the kind of politics we're used to in the US, but while political changes that dramatic are rare historically, I think we're at a moment where the risk is elevated enough that we ought to think about the odds. I might, for example, give odds for a collapse of democracy in the US over the next couple of years at ~2-5%- if the US were to elect 20 presidents similar to the current one over a century, I'd expect better than even odds of one of them making themselves into a Putinesque dictator. A collapse like that would substantially increase the risk of war, I'd argue, including raising a real possibility of nuclear civil war. That might increase the expected death rate for young and middle-aged adults in that scenario by a point or two on its own. It might also introduce a small risk of extremely large atrocities against minorities or political opponents, which could increase the expected death rate by a few tenths of a percent. There's also a small risk of economic collapse. Something like a political takeover of the Fed combined with expensive, poorly considered populist policies might trigger hyperinflation of the dollar.  When that sort of thing happens overseas, you'll often see reduced health outcomes and breakdown in civil order increasing the death rate by up to a percent- and, of course, it would introduce new tail risks, increasing the expected death rate further. I should note that I don't think the odds of any of this are high enough to worry about my safety now- but needing to emigrate is much more likely outcome than actually being threatened, and that's a headache I am mildly worried about.
Dagon
31

This seems like a story that's unsupported by any evidence, and no better than fiction.
 

They could have fought over resources in a scramble of each against all, but anarchy isn't stable.

This seems most likely, and "stable" isn't a filter in this situation - 1/3 of the population will die, nothing is stable.   It wouldn't really be "each against all", but "small (usually family) coalitions against some of the other small-ish coalitions".   The optimal size of coalition will be dependend on a lot of factors, including ease of defection and strength of non-economic bonds between members.

1Zero Contradictions
Not at all. It's just a description of the island's population over time, followed by a logical conclusion of what most likely happened when the ecosystem becomes overpopulated. Without sufficient famine, disease, or predation to cull the population back below the carrying capacity, and without new crops, technologies, or resources to satisfy the population, the inevitable outcome is conflict over resources. Which sentences are "unsupported" in your opinion? The ecocide hypothesis is not a minority position either. There is criticism against it, but we also know that there's a strong and general humanist academic bias to oppose it in general. I think this is pedantic, but I understand what you meant. Parents would compete against other parents to feed their starving children, and siblings may compete against their siblings to some extent and others for care and resources. Coalitions could form to attack other coalitions, but the possibility of defection or betrayal effectively turns the competition for survival into each against all.
Dagon
*64
  1. If you could greatly help her at small cost, you should do so.

This needs to be quantified to determine whether or not I agree.  In most cases I imagine (and a few I've experienced), I would (and did) kill the animal to end it's suffering and to prevent harm to others if the animal might be subject to death throes or other violent reactions to their fear and pain.

In other cases I imagine, I'd walk away or drive on, without a second thought.  Neither the benefit nor the costs are simple, linear, measurable things.

  1. Her suffering is bad.

I don't have a... (read more)

4G Wood
Agree with Dagon here, when omnizoid say's "Its obvious that you should" they are calling on the rules of their own morality. Its similar with "Her suffering is bad", that's a direct moral judgment. Both statements fall apart when you consider that someone may have different moral rules than you. For example, in NZ we have an issue with deer destroying our native bush which in turn hurts our native birds. Deer are considered an invasive species and are actively eradicated. In the case when you are actively in the presence of a hurting deer empathy drives you to help, suffering is not pleasant to witness. However I suspect that many NZ's would condemn every deer in NZ to a painful death, as long as they didn't have to witness it, in order to save our trees and birdlife.
Dagon
20

One challenge I'd have for you / others who feel similar to you, is to try to get more concrete on measures like this, and then to show that they have been declining.

I've given some thought to this over the last few decades, and have yet to find ANY satisfying measures, let alone a good set.  I reject the trap of "if it's not objective and quantitative, it's not important" - that's one of the underlying attitudes causing the decline.

I definitely acknowledge that my memory of the last quarter of the previous century is fuzzy and selective, and beyond t... (read more)

Dagon
20

Do you think that the world is getting worse each year? 


Good clarification question!  My answer probably isn’t satisfying, though. “It’s complicated” (meaning: multidimensional and not ordinally comparable).

On a lot of metrics, it’s better by far, for most of the distribution.  On harder-to-operationally-define dimensions (sense of hope and agency for the 25th through 75th percentile of culturally normal people), it’s  quite a bit worse.

4ozziegooen
Thanks for the specificity!  > On harder-to-operationally-define dimensions (sense of hope and agency for the 25th through 75th percentile of culturally normal people), it’s  quite a bit worse. I think it's likely that many people are panicking and losing hope each year. There's a lot of grim media around. I'm far less sold that something like "civilizational agency" is declining. From what I can tell, companies have gotten dramatically better at achieving their intended ends in the last 30 years, and most governments have generally been improving in effectiveness.  One challenge I'd have for you / others who feel similar to you, is to try to get more concrete on measures like this, and then to show that they have been declining. My personal guess is that a bunch of people are incredibly anxious over the state of the world, largely for reasons of media attention, and then this spills over into them assuming major global ramifications without many concrete details or empirical forecasts.   
Dagon
40

would consider the end of any story a loss. 

Unfortunately, now you have to solve the fractal-story problem.  Is the universe one story, or does each galaxy have it's own?  Each planet?  Continent?  Human?  Subpersonal individual goals/plotlines? Each cell?

1Aram Panasenco
I see where you're coming from, but I think any term in anything anyone writes about alignment can be picked apart ad infinitum. This can be useful to an extent, but beyond a certain point talking about meanings and definitions becomes implementation-specific. Alignment is an engineering problem first and a philosophical problem second. For example, if RLHF is used to achieve alignment, the meaning of "story" will get solidified through thousands of examples and interactions. The AI will get reinforced to not care about cells or individuals, care about ecosystems and civilizations, and not care as much about the story-of-the-universe-as-a-whole. If a different alignment method is used, the meaning of "story" will be conveyed differently. If the overall idea is good and doesn't have any obvious failure modes other than simple definitions (e.g. "story" seems to be orders of magnitude simpler to define than "human happiness" or "free will"), I'd consider that a huge success and a candidate for the community to focus real alignment implementation efforts on.
Dagon
20

I feel like you're talking in highly absolutist terms here.

You're correct, and I apologize for that.  There are plenty of potential good outcomes where individual autonomy reverses the trend of the last ~70 years.  Or where the systemic takeover plateaus at the current level, and the main change is more wealth and options for individuals.  Or where AI does in fact enable many/most individual humans to make meaningful decisions and contributions where they don't today.

I mostly want to point out that many disempowerment/dystopia failure scenarios don't require a step-change from AI, just an acceleration of current trends. 

2ozziegooen
Do you think that the world is getting worse each year?  My rough take is that humans, especially rich humans, are generally more and more successful.  I'm sure there are ways for current trends to lead to catastrophe - line some trends dramatically increasing and others decreasing, but that seems like it would require a lengthy and precise argument. 
Dagon
20

Presumably, if The Observer has a truly wide/long view, then destruction of the Solar System, or certainly loss of all CHON-based lifeforms on earth, wouldn't be a problem - there have got to be many other macroscopic lifeforms out there, even if The Great Filter turns out to be "nothing survives the Information Age, so nobody ever detects another lifeform".  

Also, you're describing an Actor, not just an Observer.  If has the ability to intervene, even if it rarely chooses to do so, that's it's salient feature.

1Aram Panasenco
The Observer gets invested in the macro-stories of the evolution/civilization it observes and would consider the end of any story a loss. Just like you would get annoyed if a show you're watching on Netflix gets cancelled after one season and it's not consolation that there are a bunch of other shows on Netflix that also got cancelled after one season. The Observer wants to see all stories unfold fully, it's not going to let squiggle maximizers cancel them. And regarding the naming, yeah I just couldn't come up with anything better. Watcher? I'm open to suggestions lol.
Dagon
42

This seems like it would require either very dumb humans, or a straightforward alignment mistake risk failure, to mess up. 

I think "very dumb humans" is what we have to work with.  Remember, it only requires a small number of imperfectly aligned humans to ignore the warnings (or, indeed, to welcome the world the warnings describe).

4ozziegooen
In many worlds, if we have a bunch of decently smart humans around, they would know what specific situations "very dumb humans" would mess up, and take the corresponding preventative measures. A world where many small pockets of "highly dumb humans" could cause an existential catastrophe is one that's very clearly incredibly fragile and dangerous, enough so that I assume reasonable actors would freak out until it stops being so fragile and dangerous. I think we see this in other areas - like cyber attacks, where reasonable people prevent small clusters of actors from causing catastrophic damage.  It's possible that the offense/defense balance would dramatically favor tiny groups of dumb actors, and I assume that this is what you and others expect, but I don't see it yet. 
Dagon
20

a lot of people have strong low-level assumptions here that a world with lots of strong AIs must go haywire.

For myself, it seems clear that the world has ALREADY gone haywire.  Individual humans have lost control of most of our lives - we interact with policies, faceless (or friendly but volition-free) workers following procedure, automated systems, etc.   These systems are human-implemented, but in most cases too complex to be called human-controlled.  Moloch won.

Big corporations are a form of inhuman intelligence, and their software and op... (read more)

2ozziegooen
I feel like you're talking in highly absolutist terms here. Global wealth is $454.4 trillion. We currently have ~8 Bil humans, with an average happiness of say 6/10. Global wealth and most other measures of civilization flourishing that I know of seem to be generally going up over time. I think that our world makes a lot of mistakes and fails a lot at coordination. It's very easy for me to imagine that we could increase global wealth by 3x if we do a decent job. So how bad are things now? Well, approximately, "We have the current world, at $454 Trillion, with 8 billion humans, etc". To me that's definitely something to work with. 
Dagon
61

In non-trivial settings, (some but not all) structural differences between programs lead to differences in input/output behaviour, even if there is a large domain for which they are behaviourally equivalent.

I think this is a crux (of why we're talking past each other; I don't actually know if we have a substantive disagreement).  The post was about detecting "smaller than a lookup table would support" implementations, which implied that the input/output functionally-identical-as-tested were actually tested in the broadest possible domain.  I full... (read more)

1Alfred Harwood
Cool, that all sounds fair to me. I don't think we have any substantive disagreements.
Dagon
20

might be true if you just care about input and output behaviour

Yes, that is the assumption for "some computable function" or "black box which takes in strings and spits out other strings."  

I'm not sure your example (of an AI with a much wider range of possible input/output pairs than the lookup table) fits this underlying distinction.  If the input/output sets are truly identical (or even identical for all tests you can think of), then we're back to the "why do we care" question.

3Alfred Harwood
hmm, we seem to be talking past each other a bit. I think my main point in response is something like this: In non-trivial settings, (some but not all) structural differences between programs lead to differences in input/output behaviour, even if there is a large domain for which they are behaviourally equivalent. But that sentence lacks a lot of nuance! I'll try to break it down a bit more to find if/where we disagree (so apologies if a lot of this is re-hashing). * I agree that if two programs produce the same input output behaviour for literally every conceivable input then there is not much of an interesting difference between them and you can just view one as a compression of the other. * As I said in the post, I consider a program to be 'actually calculating the function', if it is a finite program which will return f(x) for every possible input x. * If we have a finite length lookup table, it can only output f(x) for a finite number of inputs. * If that finite number of inputs is less than the total number of possible inputs, this means that there is at least one input (call it x_0) for which a lookup table will not output f(x_0). * I've left unspecified what it will do if you query the lookup table with input x_0. Maybe it doesn't return anything, maybe it outputs an error message, maybe it blows up. The point is that whatever it does, by definition it doesn't return f(x_0). * Maybe the number of possible inputs to f is finite and the lookup table is large enough to accommodate them all. In this case, the lookup table would be a 'finite program which will return f(x) for every possible input x' so I would be happy to say that there's not much of a distinction between the lookup table and a different method of computing the function. (Of course, there is a trivial difference in the sense that they are different algorithms, but its not the one we are concerned with here). * However, this requires that the size of the program is on the order of (or la
Dagon
62

i don't exactly disagree with the methodology, but I don't find the "why do we care" very compelling.  For most practical purposes, "calculating a function" is only and exactly a very good compression algorithm for the lookup table.  

Unless we care about side-effects like heat dissipation or imputed qualia, but those seem like you need to distinguish among different algorithms more than just "lookup table or no".

2Alfred Harwood
I think I disagree. Something like this might be true if you just care about input and output behaviour (it becomes true by definition if you consider that any functions with the same input/output behaviour are just different compressions of each other). But it seems to me that how outputs are generated is an important distinction to make.  I think the difference goes beyond 'heat dissipation or imputed qualia'. As a crude example, imagine that, for some environment f(x) is an 'optimal strategy' (in some sense) for inputs x. Suppose we train AIs in this environment and AI A learns to compute f(x) directly whereas AI B learns to implement a lookup table. Based on performance alone, both A and B are equal, since they have equivalent behaviour. But it seems to me that there is a meaningful distinction between the two. These differences are important since you could use them to make predictions about how the AIs might behave in different environments or when exposed to different inputs. I agree that there are more degrees of distinction between algorithms than just "lookup table or no". These are interesting, just not covered in the post!
Dagon
52

(I’m using time-sensitive words, even though we are stepping out of the spacetime of our universe for parts of this discussion.)

Maybe use different words, so as not to imply that there is a temporal, causal, or spacial relation.

Many people realize that, conceptually “below” or “before” any “base universe,” there is

I don't realize or accept that.  Anything that would be in those categories are inaccessible to our universe, and not knowable or reachable from within.  They are literally imaginary.

1LVSN
What are the other describable or possible-though-indescribable hypotheses? If it's intuitive that there are no other hypotheses to start from — if the explanations have been reduced to some small number of all imaginable possibilities — that's a non-nothing sort of evidence which ought to be contended with at the very least, rather than scoffed at with 'you didn't see an epistemic polylemma therefor there's no evidence that there was one'.
Load More