All of Arepo's Comments + Replies

Arepo20

Did the ringing go away over time, or was it permanent?

Arepo20

What do you mean the leadership is shared? That seems much less true now Effective Ventures have started spinning off their orgs. It seems like the funding is still largely shared, but that's a different claim.

Arepo2-1

I would also be interested to see this. Also, could you clarify:

I have definitely taken actions within the bounds of what seems reasonable that have aimed at getting the EA community to shut down or disappear (and will probably continue to do so).

Are you talking here about 'the extended EA-Alignment ecosystem', or do you mean you've aimed at getting the global poverty/animal welfare/other non-AI-related EA community to shut down or disappear?

habryka117

The leadership of these is mostly shared. There are many good parts of EA, and reform would be better than shutting down, but reform seems unlikely at this point.

My world model mostly predicts effects on technological development and the long term future dominate, so in as much as the non-AI related parts of EA are good or bad, I think what matters is their effect on that. Mostly the effect seems small, and quibbling over the sign doesn't super seem worth it.

I do think there is often an annoying motte and bailey going on where people try to critique EA for... (read more)

Arepo20

Another effect I'm very concerned about is the unseen effect on the funding landscape. For all EVF organisations are said to be financially independent, none of them seem to have had any issue getting funding, primarily from Open Phil (generally offering market rate or better salaries and in some cases getting millions of dollars on marketing alone), while many other EA orgs - and, contra the OP, there many more* outside the EVF/RP net than within - have struggled to get enough money to pay a couple of staff a living wage.

* That list excludes regional EA s... (read more)

Arepo150

It doesn't matter whether you'd have been hypothetically willing to do something for them. As I said on the Facebook thread, you did not consult with them. You merely informed them they were in a game, which, given the social criticism Chris has received, had real world consequences if they misplayed. In other words, you put them in harm's way without their consent. That is not a good way to build trust.

Arepo*140

The downvotes on this comment seem ridiculous to me. If I email 270 people to tell them I've carefully selected them for some process, I cannot seriously presume they will give up >0 of their time to take part in it. 

Any such sacrifice they make is a bonus, so if they do give up >0 time, it's absurd to ask that they give up even more time to research the issue.

Any negative consequences are on the person who set up the game. Adding the justification that 'I trust you' does not suddenly make the recipient more obligated to the spammer.

habryka*110

It's not like we asked 270 random people. We asked 270 people, each one of which had already invested many hundreds of hours into participating on LessWrong, many of which I knew personally and considered close friends. Like, I agree, if you message 270 random people you don't get to expect anything from them, but the whole point of networks of trust is that you get to expect things from each other and ask things from each other.

If any of the people in that list of 270 people had asked me to spend a few minutes doing something that was important to them, I would have gladly obliged.

Arepo110

My impression is that many similar projects are share houses or other flat hierarchies. IMO a big advantage of the model here is a top-down approach, where the trustees/manager view it as a major part of our job to limit and mitigate interpersonal conflicts, zero sum status games etc.

Arepo00

Whatever you call it, they've got to identify some alternative, even if only tacitly by following some approximation of it in their daily life.

Arepo00

I would like to write an essay about that eventually, but I figured persuading PUs of the merits of HU was lower hanging fruit.

For what it's worth, I have a lot of sympathy with your scepticism - I would rather (and believe it possible to) build a system resembling ethics up without reference to normativity, 'oughts', or any of their associated baggage. I think the trick will be to properly understand the overlap of ethics and epistemology, both of which are subject to similar questions (how do we non question-beggingly 'ground' 'factual' questions?), but ... (read more)

0[anonymous]
Very true! And this is precisely why I'm outright suspicious of non-naturalistic theoretical ethics and it's magical "oughts". In my case, in fact, I'm especially suspicious of Peter Singer and his simplistic form of hedonic utilitarianism, because it seems to me to rely overmuch on intuition pumps and too little on naturalized descriptions of how actual agents judge value.
0torekp
Good thing Bayesians don't need to identify the null hypothesis. Upvoted for mentioning that ethics and epistemology are subject to similar questions. That's a huge insight, familiar in academic philosophy, but AFAICT rare among self-identified rationalists and little discussed on lesswrong.
Arepo50

How do we know EY isn't doing the same?

0[anonymous]
Indeed. Given a lack of transcripts being released, I give a reasonable amount of probability that there is a trick of some sort involved (there have been some proposals of what that might be, e.g. "this will get AI research to get more donations"), although I don't think that would necessarily defeat the purpose of the trick: after all, the AI got out of the box either way!
Arepo00

‘A charity that very efficiently promoted beauty and justice’ would still be a utilitarian charity (where the form of util defined utility as beauty and justice), so if that’s not EA, then EA does not = utilitarianism, QED.

Also, as Ben Todd and others have frequently pointed out, many non-utilitarian ethics subsume the value of happiness. A deontologist might want more happiness and less suffering, but feel that he also had a personal injunction against violating certain moral rules. So long as he didn’t violate those codes, he might well want to maximise efficient use of welfare.

Arepo20

I'd guess these effects are largely not causation, but correlation caused by conscientiousness/ambition causing both double majors and higher earnings.

Unless you're certain of this or have some reason to suspect a factor pulling in the other direction, this still seems to suggest higher expectation from doing a double major.

0D_Malik
I definitely agree, especially considering that double majors aren't even that hard if you plan ahead. (For instance, you can major in math and management science while doing less units than the minimum allowed.) I've edited the original post with a table on good major/minor combinations.
Arepo50

Written a full response to your comments on Felicifia (I'm not going to discuss this in three different venues), but...

your opponent's true rejection seems to be "cryonics does not work"

This sort of groundless speculation about my beliefs (and its subsequent upvoting success), a) in a thread where I’ve said nothing about them, b) where I’ve made no arguments to whose soundness the eventual success/failure of cryo would be at all relevant, and c) where the speculator has made remarks that demonstrate he hasn’t even read the arguments he’s dism... (read more)

2Viliam_Bur
I apologize for misinterpreting your position. I wrote what at the moment seemed to me as the most likely explanation.
2lavalamp
My guess is that people are mostly upvoting for the tribal identity warning in the first paragraph.
0somervta
To be fair, I upvoted the parent for a lot more than that bit.
Arepo60

Assuming you accept the reasoning, 90% seems quite generous to me. What percentage of complex computer programmes when run for the first time exhibit behaviour the programmers hadn't anticipated? I don't have much of an idea, but my guess would be close to 100. If so, the question is how likely unexpected behaviour is to be fatal. For any programme that will eventually gain access to the world at large and quickly become AI++, that seems (again, no data to back this up - just an intuitive guess) pretty likely, perhaps almost certain.

For any parameter of ... (read more)

7falenas108
That's for normal programs, where errors don't matter. If you look at ones where people carefully look over the code because lives are at stake (like NASA rockets), then you'll have a better estimate. Probably still not accurate, because much more is at stake for AI than just a few lives, but it will be closer.
7TheOtherDave
You know, the idea that SI might at any moment devote itself to suppressing AI research is one that pops up from time to time, the logic pretty much being what you suggest here, and until this moment I have always treated it as a kind of tongue-in-cheek dig at SI. I have only just now come to realize that the number of people (who are not themselves affiliated with SI) who really do seem to consider suppressing AI research to be a reasonable course of action given the ideas discussed on this forum has a much broader implication in terms of the social consequences of these ideas. That is, I've only just now come to realize that what the community of readers does is just as important, if not more so, than what SI does. I am now becoming genuinely concerned that, by participating in a forum that encourages people to take seriously ideas that might lead them to actively suppress AI research, I might be doing more harm than good. I'll have to think about that a bit more. Arepo, this is not particularly directed at you; you just happen to be the data point that caused this realization to cross an activation threshold.
Arepo10

Seems like a decent reply overall, but I found the fourth point very unconvincing. Holden has said 'what he knows know' - to wit that whereas the world's best experts would normally test a complicated programme by running it, isolating out what (inevitably) went wrong by examining the results it produced, rewriting it, then doing it again.

Almost no programmes are glitch free, so this is at best an optimization process and one which - as Holden pointed out - you can't do with this type of AI. If (/when) it goes wrong the first time, you don't get a second chance. Eliezer's reply doesn't seem to address this stark difference between what experts have been achieving and what SIAI is asking them to achieve.

0bogdanb
I agree with the glitch problems. But (1) programmers and techniques are improving; (2) people are more careful when aware of danger; (3) if it’s hard but inevitable, giving up doesn’t sound like a winning strategy. I mean, if people make mistakes at some important task, how isn’t it a good idea to get lots of smart mathematicians to think hard about how to avoid mistakes? Note that all doctors, biologists, nuclear physicists and rocket scientists are also not glitch free, but those that work with dangerous stuff do tend to err less often. But they have to be aware of the dangers (or at least anticipate their existence). A doctor might try a different pill if the first one doesn’t seem to work well against the sniffles, but will be much less inclined to experiments when they know the problem is a potential pandemic. (By the way, it is probably possible that the first possible AGI is buggy, and a killer, and will foom in a few seconds (or before anyone can react, anyway); it might even be likely. But it’s still possible we’ll get several chances. My point is not that we don’t have to worry about anything, but that even if the chances might be low it still makes sense to try harder. And, hey, AFAIK the automatic trains in Paris work much better than the human-driven ones. It’s not quite a fair comparison in any direction, but there is evidence that we can make stuff work pretty well at least for a while.) ETA: You know, now that I think about it, it seems plausible that programmer errors would lean towards the AGI not working (e.g. you divide by zero; core dump; the program stops), while a mathematician’s error would lean towards the AGI working but doing something catastrophic (e.g. your encryption program has exactly zero bugs, it works exactly as designed, but ROT13 has been proven cryptographically unsound after you used it to send that important secret). So maybe it’s a good idea if the math guys start thinking hard long in advance?
Arepo00

Hm. Interesting piece. I'm partially sold, but not on this: 'Further, I see little difference between how a Muslim "chooses" to get upset at disrespect to Mohammed, and how a Westerner might "choose" to get upset if you called eir mother a whore.'

I'm pretty content to call that a sort of choice, especially if you make it a fair comparison, ie a general remark not victimising one person that all mothers are whores. After all, there’s still a pretty big difference between that (or even the rather more inflammatory ‘all Western mothers are... (read more)

Arepo91

I don’t know how relevant improv is to Less Wrongers, but I find it helpful for everyday social interactions, so:

Primary recommendation: Salinsky & Frances-White’s The Improv Handbook.

Reason It’s one of the only improv books which actually suggests physical strategies for you to try out that might improve your ability rather than referring to concepts that the author has a pet phrase for that they use as a substitute for explaining what it means. Not all of the suggetions worked for me, and they’re based on primarily on anecdotal evidence (plus the s... (read more)

1Karmakaiser
This will be a study project to me after the semester so thanks for the recommendations.
Arepo00

How does one create an open thread? The only options I had available were this and comments. Is it something you need minimum karma for?

1Oscar_Cunningham
This is the most recent open thread. It's a post in "Main" to which people can comment if they have something to say which isn't even worth a "Discussion" level post. A new one is created twice a month.
Arepo-10

If I was reliable enough to be sure, then I'd probably know where the comment in question was :P But I am fairly confident. They weren't professionally made videos (nor was the website professional-looking), just a set by this one guy explaining one bit of maths at a time.

-1Anubhav
Sounds to me like you're confabulating.
Arepo20

We've changed it to a more standard font now. How's that?

0Larks
Much better!
Arepo10

We've been experimenting a bit with the font. A bit of Googling failed to yield any scientific data on what font types people prefer, though there's a fair bit on font size. Sounds like a new test is in order, though... (incidentally, if you or anyone else reading this happen to know of any decent studies on fonts for reading online, I'd be keen to see it).

0kilobug
Hum, no, I did see some studies about print fonts, but I don't remember any about screen fonts (the issues being slightly differents). But at least for me, regularity is important in a font, if we take http://80000hours.org/blog/11-health-vs-education for example (which I found the content to be interesting), and the word 'donations' in the second line, it seems to me that the letters "d", "o" and "s" are bigger than the other letters, which make it feels a bit like DOnatiOnS and is painful/distracting to read.
0Larks
I agree with Nesov; the font is fine, but size and inter-line spacing mean I can only see ~130 words on my netbook's screen at a time.
1Vladimir_Nesov
Since the topic came up: The font seems fine to me, but the distance between lines in a paragraph seems too big, harder to parse the text into "chunks".
Arepo80

Hitting on desperate boys(/girls) is an unusual strategy by definition...

0[anonymous]
5wedrifid
(Nitpick: This is not technically by definition.)
Arepo00

Apologies for being a bit OT, but new account so I can't post this question directly (if that's even the proper approach) - and they're semi-relevant to this question.

I'm trying to remember the location of some maths teaching videos, or enough about them to find them again. They were made by a member of the LW community, though hosted on his own page, and mentioned in a post about self-improvement or similar. I think I mentioned them on here before - they seemed like great videos individually, though sadly lacking signposts from one to the next as a collection. Anyone have any idea what I'm talking about?

[This comment is no longer endorsed by its author]Reply