People will be like ‘we have these correlational studies so you should change your entire diet to things your body doesn’t tell you are good and that bring you zero joy.’
I mean, seriously, f*** that s***. No.
I do buy that people have various specific nutritional requirements, and that not eating vegetables and fruits means you risk having deficits in various places. The same is true of basically any exclusionary diet chosen for whatever reason, and especially true for e.g. vegans.
In practice, the only thing that seems to be an actual issue is fiber.
"I don'...
“Meaningless” is vaguely defined here. You defined free will at the beginning, so it must have some meaning in that sense.
It seems like “meaningless” is actually a placeholder for “doesn’t really exist”.
Which would make the trilemma boil down to:
And your basis for rejecting point 1 is that “truth wouldn’t matter, anything would be justified, therefore it’s false”.
I don’t think this follows.
Ultimately, what you’re pointing out is an issue of distinguishing between a non-free op...
Then of what use is the test? Of what use is this concept?
You seem to be saying “the true Turing test is whether the AI kills us after we give it the chance, because this distinguishes it from a human”.
Which essentially means you’re saying “aligned AI = aligned AI”
It’s true any job can find unqualified applicants. What I’m saying is that this in particular relies on an untenably small niche of feasible candidates that will take an enormous amount of time to find/filter through on average.
Sure, you might get lucky immediately, but without a reliable way to find the “independently wealthy guy who’s an intellectual and is sufficiently curious about you specifically that he wants to sit silently and watch you for 8 hours a day for a nominal fee”, your recruitment time will, on average, be very long, especially in compar...
Do you genuinely think that you can find such people “reliably”?
Unless you’re paying gratuitously, the only people who would reliably be interested in doing this would be underqualified randoms. Expect all benefit to be counteracted by the time it takes to get into a productive rhythm, at which point they’ll likely churn in a matter of weeks anyway.
I cannot imagine losing this game as the gatekeeper either, honestly.
Does anyone want to play against me? I’ll bet you $50 USD.
I agree and I am putting my money where my mouth is.
I will play this game under the rules linked in the OP with me as the gatekeeper and anyone as the AI. I will bet at odds 10:1 (favorable to you) that I will not let the AI out. The minimum bet is my 10 k USD against your 1 k USD and the maximum bet my 100 k USD against your 10 k USD. We will use a mutually accepted LW community member as referee.
If you believe you have at least a 10% chance of making me let the AI out, you should take this bet. I predict no one will take me up on it.
I speculate that the ...
I play on lichess often. I can tell you that a lichess rating of 2900 absolutely corresponds to grandmaster level strength. It is rare for FMs or IMs to exceed a 2800 blitz rating. Most grandmasters hover around 2600-2800 blitz.
The discussion on attack surfaces is very useful, intuitive and accessible. If a better standalone resource doesn’t already exist, such a (perhaps expanded) list/discussion would be a useful intro for people unfamiliar with specific risks.
This was excruciatingly frustrating to read, well done.
This is well-reasoned, but I have difficulty understanding why this kind of takeover would be necessary from the perspective of a powerful, rational agent. Assuming AGI is indeed worth its name, it seems the period of time needed for it to "play nice" would be very brief.
AGI would be expected to be totally unconcerned with being "clean" in a takeover attempt. There would be no need to leave no witnesses, nor avoid rousing opposition. Once you have access to sufficient compute, and enough control over physical resources, why wait 10 years for humanity to be...
My first thought as well. IGF-1 exists for a reason. Growth is universally necessary for development, repair and function.
shift the
Minor edit - should be "shift in the"
It's encouraging to see more emphasis recently on the political and public-facing aspects of alignment. We are living in a far-better-than-worst-case world where people, including powerful ones, are open to being convinced. They just need to be taught - to have it explained to them intuitively.
It seems cached beliefs produced by works like you get about five words have led to a passive, unspoken attitude among many informed people that attempting to explain anything complicated is futile. It isn't futile. It's just difficult.
In another of your most downvoted posts, you say
I kind of expect this post to be wildly unpopular
I think you may be onto something here.
You can fail to get rid of balls. All of your energy and effort can go into not allowing something to crash or fall, averting each disaster shortly before it would be too late. Speaking for ten minutes with each of fifty of sources every day can be a good way to keep any of them from being completely neglected, but it’s a terrible way to actually finish any of those projects. The terminal stage of this is a system so tied up in maintaining itself and stopping from falling behind that it has no slack to clear tasks or to improve its speed.
This is the ...
It seems intuitively bad:
All good practices. Although, isn't this just "more metal", rather than "less ore"? I imagine one would want to maximize both the inputs and outputs, even if opportunities for increasing inputs are exhausted more quickly.
How is such a failure of imagination possible?
It's odd to claim that, contingent upon AGI being significantly smarter than us, and wanting to kill us, that there is no realistic pathway for us to be physically harmed.
Claims of this sort by intelligent, competent people likely reveal that they are passively objecting to the contingencies rather than disputing whether these contingencies would lead to the conclusion.
The quotes you're responding to here superficially imply "if smart + malicious AI, it can't kill us", but it seems much more likely this is a warped translation of either "AI can't be smart", or "AI can't be malicious".
I would happily play the role of B.
I do not have an established FIDE rating, but my strength is approximately 1850 FIDE currently (based on playing against FIDE rated players OTB quite often, as well as maintaining 2100-2200 blitz ratings on Lichess & Chess.com, and 2200-2300 bullet). I'd be available after 6:30 pm (UTC+10) until ~12:00 pm (UTC+10). Alternatively, weekends are very flexible. I could do a few hours per week.
I agree short/long time controls are a relevant, because speed is a skill that is almost entirely independent of conceptual knowledge and is mostly a function of baseline playing ability.
Edit: Would also be fine with C
Strongly agree. To my utter bewilderment, Eliezer appears to be exacerbating this vulnerability by making no efforts whatsoever to appear credible to the casual person.
In nearly all of his public showings in the last 2 years, he has:
As a result, to the layperson, he comes off as an egotistical, pessimistic nerd with fringe views - a perfect clown from which to retreat to a "mid...
A high VO2 Max is a superior metric. It is harder to achieve, and more predictive of health outcomes.
It’s possible and even somewhat common to have a low resting heart rate with a below average VO2 max, but it’s virtually impossible to have a high VO2 max and a high resting heart rate.