All of TruePath's Comments + Replies

I agree that's a possible way things could be. However, I don't see how it's compatible with accepting the arguments that say we should assume that alignment is a hard problem. I mean absent such arguments why expect you have to do anything special beyond normal training to solve alignment?

As I see the argumentative landscape the high x-risk estimates depend on arguments that claim to give reason to believe that alignment is just a generally hard problem. I don't see anything in those arguments that distinguishes between these two cases.

In other words o... (read more)

2Vladimir_Nesov
AIs have advantages such as thinking faster and being as good at everything as any other AI of the same kind. These advantages are what simultaneously makes them dangerous and puts them in a better position to figure out alignment or coordination that protects from misaligned AIs and human misuse of AIs. (Incidentally, see this comment on various relevant senses of "alignment".) Being better at solving problems and having effectively more time to solve problems improves probability of solving a given problem.

I don't mean alignment with human concerns. I mean that the AI itself is engaged in the same project we are: building a smarter system than itself. So if it's hard to control the alignment of such a system then it should be hard for the AI. (In theory you can imagine that it's only hard at our specific level of intelligence but in fact all the arguments that AI alignment is hard seem to apply equally well to the AI making an improved AI as to us making an AI).

See my reply above. The AI x-risk arguments require the assumption that superintelligence nece... (read more)

Those are reasonable points but note that the arguments for AI x-risk depend on the assumption that any superintelligence will necessarily be highly goal directed. Thus, either the argument fails because superintelligence doesn't imply goal directed,

And given that simply maximizing the intelligence of future AIs is merely one goal in a huge space it seems highly unlikely that (especially if we try to avoid this one goal) we just get super unlucky and the AI has the one goal that is compatible with improvement.

I like the idea of this sequence, but -- given the goal of spelling out the argument in terms of first principles -- I think more needs to be done to make the claims precisce or acknowledge they are not.

I realize that you might be unable to be more precisce given the lack of precision in this argument generally -- I don't understand how people have invested so much time/mondy on research to solve the problem and so little on making the argument for it clear and rigorous -- but if that's the case I suggest you indicate where the definitions are insufficient... (read more)

Maybe this question has already been answered but I don't understand how recursive self-improvement of AIs is compatible with the AI alignment problem being hard.

I mean doesn't the AI itself face the alignment problem when it tries to improve/modify itself substantially? So wouldn't a sufficently intelligent AI refuse to create such an improvement for fear the goals of the improved AI would differ from its own?

2JBlack
I suspect that the alignment problem is much easier when considering expanding your own capabilities versus creating a completely new type of intelligence from scratch that is smarter than you are. I'm far from certain of this, but it does seem likely. There is also the possibility that some AI entity doesn't care very much about alignment of later selves with earlier ones, but acts to self-improve or create more capable descendants either as a terminal goal, or for any other reason than instrumentally pursuing some preserved goal landscape. Even a heavily goal-directed AI may self-improve knowing that it can't fully solve alignment problems. For example, it may deduce that if it doesn't self-improve then it will never achieve any part of its main goals, whereas there is some chance that some part of those goals can be achieved if it does. There is also the possibility that alignment is something that can be solved by humans (in some decades), but that it can be solved by a weakly superintelligent and deceptive AI much faster.
2Dagon
Recursive self-improvement could happen on dimensions that don't help (or that actively harm) alignment.  That's the core of https://www.lesswrong.com/tag/orthogonality-thesis .   The AI may (or may not; the capability/alignment mismatch may be harder than many AIs can solve; and the AI may not actually care about itself when creating new AIs) try to solve alignment of future iterations with itself, but even if so, it doesn't seem likely that an AI misaligned with humanity will create an AI that is.
2Vladimir_Nesov
It's unclear if alignment is hard in the grand scheme of things. Could snowball quickly, with alignment for increasingly capable systems getting solved shortly after each level of capability is attained. But at near-human level, which seems plausible for early LLM AGIs, this might be very important in requiring them to figure out coordination to control existential risk while alignment remains unsolved, remaining at relatively low capability level in the meantime. And solving alignment might be easier for AIs with simple goals, allowing them to recursively self-improve quickly. As a result, AIs aligned with humanity would remain vulnerable to FOOMing of misaligned AIs with simple goals, and would be forced by this circumstance to comprehensively prevent any possibility of their construction rather than mitigate the consequences.
2Rafael Harth
Alignment isn't required for high capability. So a self-improving AI wouldn't solve it because it has no reason to. This becomes obvious if you think about alignment as "doing what humans want" or "pursuing the values of humanity". There's no reason why an AI would do this.

I'd just like to add that even if you think this piece is completly mistaken I think it certainly shows we are definitely not knowledgeable enough about what and how values and motives work in us much less AI to confidently make the prediction that AIs will be usefully described with a single global utility function or will work to subvert their reward system or the like.

Maybe that will turn out to be true but before we spend so many resources on trying to solve AI alignment let's try to make the argument for the great danger much more rigorous first...usually best way to start anyway.

This is one of the most important posts ever on LW though I don't think the implications have been fully drawn out. Specifically, this post raises serious doubts about the arguments for AI x-risk as a result of alignment mismatch and the models used to talk about that risk. It undercuts both Bostrom's argument that an AI will have a meaningful (self-aware?) utility function and Yudkowsky's reward button parables.

The role these two arguments play in convincing people that AI x-risk is a hard problem is to explain why, if you don't anthropomorphize shoul... (read more)

I absolutely think that the future of online marketing g involves more asking ppl for their prefs. I know I go into my settings on good to active curate what they show me.

Indeed, I think Google is leaving a fucking pile of cash on the table by not adding a "I dislike" button and a little survey on their ads.

2jefftk
> not adding a "I dislike" button and a little survey on their ads Have you tried clicking the little "ad choices" button in the upper right hand corner?

I feel there is something else going on here too.

Your claimed outside view asks us to compare a clean codebase with an unclean one and I absolutely agree that it's a good case for using currentDate when initially writing code.

But you motivated this by considering refactoring and I think things go off the rails there. If the only issue in your codebase was you called currentDate yyymmdd consistently or even had other consistent weird names it wouldn't be a message it would just have slightly weird conventions. Any coder working on it for a non-trivial len... (read more)

2Adam Zerner
Perhaps. yyyymmdd to currentDate is just an example though. In practice I expect that codebases would have a variety of different issues.

I don't think this is a big problem.. The people who use ad blockers are both a small fraction of internet users and the most sophisticated ones so I doubt they are a major issue for website profit. I mean sure, Facebook is eventually going to try to squeeze out the last few percent of users if they can do so with an easy countermeasure but if this was really a big concern websites would be pushing to get that info back from the company they use to host ads. Admittedly when I was working on ads for Google (I'm not cut out to be out of academia so I went... (read more)

2Pattern
Ads decrease consumption drastically, and (patreon user) Gwen found it to be a massive loss. Important to know if there are other ways of making money or that's not your (only) goal. I've found targeting ads to be, at best, 'selling pizza right in front of the pizza store - right after the customers have left with the pizza they bought*'. Arguably some small (rare) theoretical benefits even if doing almost no work (though in a world filled with badly targeted ads, they'll get mowed down like all the rest) via followup, but it's not pre-emptive and I have to keep track of what I want (to consume), and when it's coming out. Ads aren't some service where I subscribe (ooh, that Batman-/Christopher Nolan-movie/TV Show was really good, I want to see trailers for it automatically when the the next one comes out). Instead they're trying to serve old left overs. (Fill in the blank) advertises that show I've already finished watching* because it's popular and on their streaming service which ?? they know I use but don't know I've already seen it ??. Ads for the new phone I have on the very same phone... Either someone has gone mad with greed, or people actually have multiple phones that are the same phone? *It wasn't that great. I'm not going back for more of the same, again. It seems bizarre for me to say 'ask me what ads I want to see', but like...wouldn't that be more effective? I know what I want, I know what I'd buy.* (Everyone tweeting #SnyderCut since forever, everyone following that tag...it doesn't take rocket science to know what they want. For all I know they paid for it in advance (kickstarter style).) *The next level thing would be budgeting. That's too much to ask for. (But might interface better with deals, in some way beyond 'a pre-order the book and it's cheaper' deal.) I'd say the weaknesses include not knowing about new products, or new categories of products. But honestly? That's a problem right now that might be fixed this way: If people are 'wast

Interesting, but I think it's the other end of the equation where the real problem lies: voting. Given the facts that

1) A surprisingly large fraction of the US population has tried hard drugs of one kind or another.

2) Even those who haven't almost surely know people who have and seem to find it interesting/fascinating/etc.. not horrifying behavior that deserves prison time.

So why is it that people who would never dream of sending their friend who tried coke to prison or even the friend who sold that friend some of his stash how do we end up with... (read more)

4[anonymous]
There's really nothing left to explain here. They would never dream of sending their friend who tried coke to prison because they're friends. The same doesn't hold for strangers. Similarly, you'd probably let a friend of yours who just lost his home spend the night at your place, but not any random homeless person.

So your intent here is to diagnose the conceptual confusion that many people have with respect to infinity yes? And your thesis is that: people are confused about infinity because they think it has a unique referant while in fact positive and negative infinity are different?

I think you are on to something but it's a little more complicated and that's what gets people are confused. The problem is that in fact there are a number of different concepts we use the term infinity to describe which is why it so super confusing (and I bet there are more... (read more)

1Slider
> people are confused about infinity because they think it has a unique referant while in fact positive and negative infinity are different? No, that is a different point. The point is that positive infinity would be better treated as multiple different values and trying to mesh them all into one quantity leads to trouble. We differentiate between 2,4,6 and don't use an umbrella term "a lot". Should you do so you could run into trouble with claims like "a lot is divisible by 4" (proof following 4/4=1 affirms, proof following 6%4!=0 refuses). I did a bad job of fighting ambigioity of the word infinity. Of the listed understandings 2 is closes but I am really pointing ot transfinitism that there are multiple values outside of all finites that are not equal to each other (ie a whole world to play with instead of single islands).

Stimulants are an excellent short term solution. If you absolutely need to get work done tonight and can't sleep amphetamine (i.e. Adderall) is a great solution. Indeed, there are a number of studies/experiments (including those the airforce relies on to give pilots amphetamines) backing up the fact that it improves the ability to get tasks done while sleep deprived.

Of course, if you are having long term sleep problems it will likely increase those problems.

-11GPT2

There is a lot of philosophical work on this issue some of which recommends taking conditional probability as the fundamental unit (in which case Bayes theorem only applies for non-extremal values). For instance, see this paper

1Yoav Ravid
Thanks, it looks quite interesting, but unfortunately i don't think i have the technical knowledge to understand most of the paper. can you make a quick summery of the relevant points?

Computability is just \Delta^0_1 definability. There are plenty of other notions of definability you could try to cash out this paradox in terms of. Why pick \Delta^0_1 definability?

If the argument worked in any particular definability notion (e.g. arithmetic definability) it would be a problem. Thus, the solution needs to explain why the argument shouldn't convince you that with respect to any concrete notion of definable set the argument doesn't go through.

2TurnTrout
Turing’s thesis applies only to this notion of definability, right?

But that's not what the puzzle is about. There is nothing about computability in it. It is supposed to be a paradox along Russell's set of all sets that don't contain themselves.

The response about formalizing exactly what counts as a set defined by an English sentence is exactly correct.

2TurnTrout
But there are more objections; even if "computability" isn't explicitly mentioned in the problem, it's still present. Are the sets "the singleton set containing 1 if and only if machine M halts on input w" and "the singleton set containing 1" the same? Even if we grant a procedure for figuring out what counts as a set, we can't even compute which sentences are duplicates.

Yah, enumerable means something different than computably enumerable.

This is just the standard sleeping beauty paradox and I’d suggest that the issue isn’t unique to FNC.

However, you are a bit quick in concluding it is time inconsistent as it’s not altogether clear that one is truly referring to the same event before and after you have the observation. The hint here is that in the standard sleeping beauty paradox the supposed update involves only information you already were certain you would get.

Id argue that what’s actually going on is that you are evaluating slightly different questions in the two cases

2Stuart_Armstrong
This is not the standard sleeping beauty paradox, as the information you are certain to get does not involve any amnesia or duplication before you get it.
Answer by TruePath90

Don't. At least outside of Silicon Valley where oversubscription may actually be a problem. It's a good intention but it inevitably will make people worry they aren't welcome or aren't the right sort of people . Instead, describe what one does or what one talks about in a way that will appeal to the kind of people who would enjoy coming

4ChristianKl
Oversubscription isn't an issue but that doesn't mean that it doesn't matter who comes. If a lot of people come who aren't a good match they water down the meetup and the regulars don't come anymore to events with an open invitation list but prefer to come to events whose attendence is more selective.

Given that you just wrote a whole post to say hi and share your background with everyone I'm pretty confident you'll fit right in and won't have any problems being too shy. Writing a post like this rather than just commenting is such a less wrong kind of thing to do so I think you'll be right at home.

5Bae's Theorem
That's very kind of you, thank you. It means a lot.

Searle can be any X?? WTF? That's a bit confusingly written.

The intuition Searle is pumping is that since he, as a component of the total system doesn't understand Chinese it seems counterintuitive to conclude that the whole system understands Chinese. When Searle says he is the system he is pointing to the fact that he is doing all the actual interpretation of instructions and is seems weird to think that the whole system has some extra experiences that let it understand Chinese even though he does not. When Searle uses the word understand he does not

... (read more)

I essentially agree with you that science can't bridge the is-ought gap (see caveats) but it's a good deal more complicated than the arguments you give here allow for (they are a good intro but I felt it's worth pointing out the complexities).

  1. When someone claims to have bridged the is-ought gap they aren't usually claiming to have analytically identified (i.e. identified as a matter of definition) ought with some is statements. That' s a crazily high bar and modern philosophers (and Sam Harris was trained as a philosopher) tend to feel true analytic id
... (read more)
3torekp
Thanks for making point 2. Moral oughts need not motivate sociopaths, who sometimes admit (when there is no cost of doing so) that they've done wrong and just don't give a damn. The "is-ought" gap is better relabeled the "thought-motivation" gap. "Ought"s are thoughts; motives are something else.

I agree with your general thrust except your statement that "you longtermists can simply forgo your own pleasure wireheading and instead work very hard on the whole growth and reproduction agenda" if we are able to wirehead in an effective manner it might be morally obligatory to force them into wireheading to maximize utility.

7Andaro
Not interested in this kind of "moral obligation". If you want to be a hedonistic utilitarian, use your own capacity and consent-based cooperation for it.

Also, your concern about some kind of disaster caused by wireheading addiction and resulting deaths and damage is pretty absurd.

Yes, people are more likely to do drugs when they are more available but even if the government can't limit the devices that enable wireheading from legal purchase it will still require a greater effort to put together your wireheading setup than it currently does to drive to the right part of the nearest city (discoverable via google) and purchasing some heroin. Even if it did become very easy to access it's still not true that m

... (read more)
1avturchin
Probably the difference is in the angle of slope of the availability of the drug. For a person who never injected anything in his system (like me), going to the "right part of the city" is frightful experience which I would hopely never do. It is very hard step. But not the same with overeating or internet-addiction, where entrance is very easy. "once monitoring equipment detects a problem you can simply shut off the intervention" Look on the Facebook addiction - did the Facebook ever shutdown anyone because he spent too long on their site and was procrastinated at work? Some people block FB on their computers voluntary, but they still lurk through proxies - it is not easy to block something harmful but addictive. Anyway, I understand your desire to "good wireheading" which will end sufferings and will not hinder productivity, and some AI controlled brain stimulation may be such a system. But if it is not controlled by advanced AI, but by just a few parameters regulation, a person could easily unwillingly touch a region of the brain (and utility function) from which he will become instantly severely addicted.

You make a lot of claims here that seem unsupported and based on nothing but vague analogy with existing primitive means of altering our brain chemisty. For instance a key claim that pretty most of your consequences seem to depend on is this: "It is great to be in a good working mood, where you are in the flow and every task is easy, but if one feels “too good”, one will be able only to perform “trainspotting”, that is mindless staring at objects.

Why should this be true at all? The reason heroin abusers aren't very productive (and, imo, heroin isn't the mo

... (read more)

Skimming the paper I'm not at all impressed. In particular, they make frequent and incautious use of all sorts of approximations that are only valid up to a point under certain assumptions but make no attempt to bound the errors introduced or justify the assumptions.

This is particularly dangerous to do in the context of trying to demonstrate a failure of the 2nd law of thermodynamics as the very thought experiment that might be useful will generally break those heuristics and approximations. Worse, the 2nd law is only a statistical regularity not a true

... (read more)

Not everyone believes that everything is commesurable and people often wish to be able to talk about these issues without implicitly presuming that fact.

Moreover, values suggests something that is desirable because it is a moral good. A priority can be something I just happen to selfishly want. For instance, I might hold diminishing suffering as a value yet my highest current priority might be torturing someone to death because they killed a loved one of mine (having that priority is a moral failing on my part but doesn't make it impossible).

Two thoughts.

First, as a relatively in shape person who walks a ton (no car living in the midwest) I can attest that I often wish I had a golf cart/scooter solution. They don't need to be a replacement for walking (though good that they can be) they might also appeal to those of us who like to walk a lot but need a replacement for a car when it gets really hot or we need to carry groceries (motorcycle style scooters require licenses and can't always be driven on campuses or parks). It would be great if these became less socially disapproved of for the no

... (read more)
6Alicorn
I expect that when I replace my scooter at the end of its useful life I will be able to get a much better scooter. For me, this one is at what I think is currently a good tradeoff between portability (more battery is heavier, more torque might require a heavier machine too), plausibility as a wheelchair (for use indoors and on the sidewalk rather than among cars), and price (I'd pay more knowing what I know, but when I got it I didn't actually realize quite how big a deal it was going to be for me).
5MBlume
I believe Alicorn meant to claim that the larger class of electric vehicles for ~1 person -- scooters, tricycles, skateboards, ebikes, etc -- are about to take off in a big way because there are a lot more people who would buy them if they knew about them/saw their friends using them than there are using them now

Except if you actually go try and do the work people's pre-theoretic understanding of rationality doesn't correspond to a single precise concept.

Once you step into Newcomb type problems it's no longer clear how decision theory is supposed to correspond to the world. You might be tempted to say that decision theory tells you the best way to act...but it no longer does that since it's not that the two-boxer should have picked one box. The two-boxer was incapable of so picking and what EDT is telling you is something more like: you sho... (read more)

First, let me say I 100% agree with the idea that there is a problem in the rationality community of viewing rationality as something like momentum or gold (I named my blog rejectingrationality after this phenomena and tried to deal with it in my first post).

However, I'm not totally sure everything you say falls under that concept. In particular, I'd say that rationality realism is something like the belief that there is a fact of the matter about how best to form beliefs or take actions in response to a particular set of experiences and that ... (read more)

If you want to argue against that piece of reasoning give it a different name because it's not the Chinese room argument. I took multiple graduate classes with professor Searle and, while there are a number of details Said definitely gets the overall outline correct and the argument you advanced is not his Chinese room argument.

That doesn't mean we can't talk about your argument just don't insist it is Searle's Chinese room argument.

1Valerio
In his paper, Searle brings forward a lot of arguments. Early in his argumentation and referring to the Chinese room, Searle makes this argument (which I ask you not to mix with later arguments without care): Later, he writes: I am framing this argument in a way it can be analyzed: 1) P (the Chinese room) is X (a program capable of passing Turing test in Chinese); 2) Searle can be any X and not understanding Chinese (as exemplified by Searle being the Chinese room and not understanding Chinese, which can be demonstrated for certain programs) thus 3) no X is understanding Chinese Searle is arguing that “no program is understanding Chinese” (I stress this in order to reply to Said). The argument "P is X, P is not B, thus no X is B" is an invalid syllogism. Nevertheless, Searle believes in this case that “P not being B” implies (or strongly points towards) “X not being B”. Yes, Searle’s intuition is known to be problematic and can be argued against accordingly. My point however is that there is out there in the space of X a program P that is quite unintuitive. I am suggesting a positive example of “P possibly understanding Chinese” which could cut short the debate. Don’t you see that giving a positive answer to the question “can a program understand?” may bring some insight in Searle’s argument too (such as developing it into a "Chinese room test" to assess whether a given program can indeed understand)? Don't you want to look into my suggested program P (semiotic AI)? In the beginning of my post I made it very clear:

To the extent they define a particular idealization it's one which isn't interesting/compelling. What one would want to have to say there was a well defined question here is a single definition of what a rational agent is that everyone agreed on which one could then show favored such and such decision theory.

To put the point differently you and I can agree on absolutely every fact about the world and mathematics and yet disagree about which is the best decision theory because we simply mean slightly different things by rational agent. Moreover,... (read more)

Obviously you can and if you define that NEW idealization an X-agent (or more likely redefine the word rationality in that situation) and then there may be a fact of the matter about how an X-agent will behave in such situations. What we can't do is assume that there is a fact of the matter about what a rational agent will do that outstrips the definition.

As such it doesn't make sense to say CDT is right or TDT or whatever before introducing a specific idealization relative to which we can prove they give the correct answer. But that idealiz... (read more)

2Vladimir_Nesov
There are pre-formal facts about what words should mean, or what meanings to place in the context where these words may be used. You test a possible definition against the word's role in the story, and see if it's apt. This makes use of facts outside any given definition, just as with the real world. And here, it's not even clear what the original definitions of agents should be capable of, if you step outside particular decision theories and look at the data they could have available to them. Open source game theory doesn't require anything fundamentally new that a straightforward idealization of an agent won't automatically represent. It's just that the classical decision theories will discard that data in their abstraction of agents. In Newcomb's problem, it's essentially discarding part of the problem statement, which is a strange thing to expect of a good definition of an agent that needs to work on the problem.

Your criticism of the philosophy/philosophers is misguided on a number of accounts.

1. You're basing those criticisms on the presentation in a video designed to present philosophy to the masses. That's like reading some phys.org arg article claiming that electrons can be in two locations at once and using that to critisize the theory of Quantum Mechanics.

2. The problem philosophers are interested in addressing may not be the one you are thinking of. Philosophers would never suggest that the assumption of logical omniscience prevents one from usi... (read more)

4Bird Concept
I don't want to base my argument on that video. It's based on the intuitions for philosophy I developed doing my BA in it at Oxford. I expect to be able to find better examples, but don't have the energy to do that now. This should be read more as "I'm pointing at something that others who have done philosophy might also have experienced", rather than "I'm giving a rigorous defense of the claim that even people outside philosophy might appreciate".

In particular, I'd argue that the paradoxical aspects of Newcomb's problem result from exactly this kind of confusion between the usual agent idealization and the fact that actual actors (human beings) are physical beings subject to the laws of physics. The apparent paradoxical aspects results because we are used to idealizing individual behavior in terms of agents where that formalism requires we specify the situation in terms of a tree of possibilities with each path corresponding to an outcome and with the payoff computed by looking at the pa... (read more)

3Chris_Leong
"This means there is simply no fact of the matter about how a rational agent (or whatever) should behave in newcomb type situations" - this takes this critique too far. Just because the usual agent idealisation breaks, it doesn't follow that we can't create a new idealisation that covers these cases.

It doesn't really make sense to talk about the agent idealization at the same time as talking about effective precommitment (i.e. deterministic/probabilistic determination of actions).

The notion of an agent is an idealization of actual actors in terms of free choices, e.g., idealizing individuals in terms of choices of functions on game theoretic trees. This is an incompatible idealization with thinking of such actors as being deterministically or probabilistically committed to actions for those same 'choices.'

Of course, ultimately, actual... (read more)

2Chris_Leong
What do you mean by agent idealization? That seems key to understanding your comment, which I can't follow at the moment. EDIT: Actually, I just saw your comment above. I think TDT/UDT show how we can extend the agent idealization to cover these kinds of situations so that we can talk about both at the same time.

You are getting the statement of the Chinese room wrong. The claim isn't that the human inside the room will learn Chinese. Indeed, it's a key feature of the argument that the person *doesn't* ever count as knowing Chinese. It is only the system consisting of the person plus all the rules written down in the room etc.. which knows Chinese. This is what's supposed to (but not convincingly IMO) be an unpalatable conclusion.

Secondly, no one is suggesting that there isn't an algorithm that can be followed which makes it appear as i... (read more)

1Valerio
TruePath, you are mistaken, my argument addresses the main issue of explaining computer understanding (moreover, it seems that you are making confusion between the Chinese room argument and the “system reply” to it). Let me clarify. I could write the Chinese room argument as the following deduction argument: 1) P is a computer program that does [x] 2) There is no computer program sufficient for explaining human understanding of [x] => 3) Computer program P does not understand [x] In my view, assumption (2) is not demonstrated and the argument should be reformulated as: 1) P is a computer program that does [x] 2’) Computer program P is not sufficient for explaining human understanding of [x] => 3) Computer program P does not understand [x] The argument still holds against any computer program satisfying assumption (2’). Does however a program exist that can explain human understanding of [x] (a program such that a human executing it understands [x])? My reply focuses on this question. I suggest to consider artificial semiosis. For example, a program P learns solely from symbolic experience of observing a symbols in a sequence that it should output “I say” (I have described how such a program would look like in my post). Another program Q could learn from symbolic experience solely how to speak Chinese. Humans do not normally learn these ways a rule for using “I say” or how to speak Chinese, because their experience is much richer. However, we could reason about the understanding that a human would have if he could have only symbolic experience and the right program instructions to follow. The semiosis performed by the human would not differ from the semiosis performed by the computer program. It can be said that program P understands a rule for using “I say”. It could be said that the computer program Q understands Chinese. You can consider [x] to be a capability enabled by sensory-motion. You can consider [x] to be consciousness. My “semiosis reply” could of course

We can identify places we know (inductively) tend to lead us astray and even identify tricks that help us avoid being affected by common fallacies which often aflict humans. However, it's not at all clear if this actually makes us more rational in any sense.

If you mean act-rationality we'd have to study if this was a good life choice. If you mean belief rationality you'd have to specify some kind of measure/notion of importance to decide when it really matters you believed the true thing. After all if it's just maximizing the numb... (read more)

The inveresly proportional thing is a bad move. Sorting through potential charitable causes is itself charitable work and it's just crazy inefficient to do that by everyone voting on what tugs at their heartstrings but by paying someone smart to consider all the various pet causes and evaluate them. Worse, the causes that are least well known are often unknown for very good reason but will now get special attention.

The reason you are right about cases like the doctor example is that when you are actually in a location that then gets hit you *are* l... (read more)

Note that your whole delegation argument rests on the idea that you have (and know you have) some kind of superior knowledge (or virtue) about what needs to get done and you're just searching for the best way to get it done. The reason it made sense to stay involved in the local campaign was because you had the special advantage of being the person who knew the right way to improve the city so you could offer something more than any other equally virtuous person you might hand the money to instead..

In contrast, in the village case you *don't* ha... (read more)

2Darmani
I think you hit the kernel of the argument in the first paragraph: If you have an obscure pet cause, then chances are it's because you do have some special knowledge about the problem. The person visiting a random village might not, but the locals do, and hence this is a reason why local charity can be effective, particularly if you live in a remote area where the problems are not quantified (and are hence probably not reading this).

Why assume whatever beings simulated us evolved?

Now I'm sure you're going to say well a universe where intelligent beings just pop into existence fully formed is surely less simple than one where they evolve. However, when you give it some more thought that's not true and it's doubtful if Occam's razor even applies to initial conditions.

I mean supposed for a moment the universe is perfectly deterministic (newtonian or no-collapse interp). In that case the Kolmogorov complexity of a world starting with a big bang that gives rise... (read more)

2cousin_it
Nice argument! But note that in such a world, all evidence of the past (like fossils) will look like the creatures indeed evolved. So for purposes of most future decisions, the creatures can safely assume that they evolved. To break that, you need to spend more K-complexity.

No, because we want the probability of being a simulation conditional on having complex surroundings not the probability of having complex surroundings conditional on a simulation. The fact that a very great number of simulated beings are created in simple universes doesn't mean that none is ever simulated in a complex one or tell us anything about whether being such a simulation is more likely than being in a physical universe.

Ok, this is a good point. I should have added a requirement that the true solution is C infinity on the part of the manifold that isn't in our temporal past. The backward's heat equation is ill-posed for this reason on...it can't be propogated arbitrarily far forward (i.e. back).

Which way you think this goes probably depends on just how strongly you think Occam's razor should be applied. We are all compelled to let the probability of a theory's truth go to zero as it's kolmogorov complexity goes to infinity but there is no prima facia reason to think it drops off particularly fast or slow. If you think , as I do, that there is only relatively weak favoring of more simple scientific laws while intelligent creatures would favor simplicity as a cognitive technique for managing complexity quite strongly you get my conclusion. But I'll admit the other direction isn't implausible.

The problem with this kind of analysis is that one is using the intuition of a physical scenario to leverage an ambiguity in what we mean by agent and decision.

Ultimately, the notion of decisions and agents are idealizations. Any actual person or AI only acts as the laws of physics dictate and agents, decisions or choices don't appear in any description in terms of fundamental physics. Since people (and programs) are complex systems that often make relatively sophisticated choices about their actions we introduce the idealization of agents and decis... (read more)

6cousin_it
Was there also no room for the fact that VNM utility maximization is useful? I'm looking for the next step in usefulness after VNM and UDT. Or you could say I'm looking for some nice math to build into an AI. My criteria for judging that math are messy, but that doesn't mean they don't exist.

I worry such a plan will face significant legal hurdles. As suggested the building would probably not fall into the exceptions to the federal fair housing act (is that right) for choosing roommates (it's not a single family dwelling but a group of apartments in some sense).

But you EXACTLY want to choose who lives there based on political/religious beliefs (almost by definition it's impossible to be a rationalist and a dogmatic unquestioning conservative christian). Also by aspects of family makeup in that you don't want people living in this community to... (read more)

0ChristianKl
If the descrimitation laws consider rationality political a way around might be to declare the whole community a religious community. Monastries have no problem picking their members via religion.
-1ChristianKl
Our census doesn't suggest that being a Christian is incompatible with being rationalist. This community holds Peter Thiel in high regard despite his Christian beliefs. Our political beliefs are also diverse.

Sorry, but you can't get around the fact that humans are not well equipped to compute probabilities. We can't even state what our priors are in any reasonable sense much less compute exact probabilities.

As a result using probabilities has come to be associated with having some kind of model. If you've never studied the question and are asked how likely you think it is there are intelligent aliens you say something like "I think it's quite likely". You only answer with a number if you've broken it down into a model (chance life evolves averag... (read more)

Uhh, why not just accept that you aren't and can never be perfectly rational and use those facts in positive ways.

Bubbles are psychologically comforting and help generate communities. Rationalist bubbling (which ironically includes the idea that they don't bubble) probably does more to build the community and correct other wrong beliefs than almost anything else.

Until and unless rationalist take over society the best strategy is probably just to push for a bubble that actively encourages breaking other (non-rationalist) bubbles.

So the equations should be (definition of vaccine efficacy from wikipedia)

.6 * p(sick2) = p(sick2) - p(sick1)
p(sick1) - .4 p(sick2) = 0 . i.e. efficacy is the difference be the unvaccinated and vacinated rates of infection divided by the unvaccinated rate. You have to assume there is no selective pressure in terms of who gets the vaccine (they have the same risk pool as the normal population for flu which is surely untrue) to get your assumtion that

.42 p(sick1) + .58p(sick2) = .1 p(sick1) + 1.38p(sick2) = .238

or 1.78 p(sick2) = .238

p(sick2)=.13 (weird I... (read more)

0Fluttershy
This confusion is due to the fact that the system of two equations I wrote in my comment above was originally crammed onto one line, rather than being separated onto two lines. Sorry! This formatting error made the comment hard to read, and has since been corrected. The correct system of equations to solve is: p(sick1) x 0.42 + p(sick2) x 0.58 = 1 x 0.10 p(sick1) = p(sick2) x 0.60 Instead of: .42 p(sick1) + .58p(sick2) = .238 .1 p(sick1) + 1.38p(sick2) = .238.

Not with respect to their revealed preferences for working in high risk jobs I understand. There are a bunch of economic papers on this but it was a surprisingly low number.

Load More