All of Hide's Comments + Replies

Hide30

A high VO2 Max is a superior metric. It is harder to achieve, and more predictive of health outcomes.


It’s possible and even somewhat common to have a low resting heart rate with a below average VO2 max, but it’s virtually impossible to have a high VO2 max and a high resting heart rate.

Hide78

People will be like ‘we have these correlational studies so you should change your entire diet to things your body doesn’t tell you are good and that bring you zero joy.’

I mean, seriously, f*** that s***. No.

I do buy that people have various specific nutritional requirements, and that not eating vegetables and fruits means you risk having deficits in various places. The same is true of basically any exclusionary diet chosen for whatever reason, and especially true for e.g. vegans.

In practice, the only thing that seems to be an actual issue is fiber.


"I don'... (read more)

Hide20

“Meaningless” is vaguely defined here. You defined free will at the beginning, so it must have some meaning in that sense.

It seems like “meaningless” is actually a placeholder for “doesn’t really exist”.

Which would make the trilemma boil down to:

  1. Free will doesn’t exist
  2. It exists and I have it
  3. It exists and I don’t have it

And your basis for rejecting point 1 is that “truth wouldn’t matter, anything would be justified, therefore it’s false”.

I don’t think this follows.

Ultimately, what you’re pointing out is an issue of distinguishing between a non-free op... (read more)

Hide20

Then of what use is the test? Of what use is this concept?

You seem to be saying “the true Turing test is whether the AI kills us after we give it the chance, because this distinguishes it from a human”.

Which essentially means you’re saying “aligned AI = aligned AI”

2Roko
no, because a human might also kill you when you give them the chance. To pass the strong-form Turing Test it would have to make the same decision (probabilistically: have the same probability of doing it) It is useful because we know what kind of outcomes happen when we put millions of humans together via human history, so "whether an AI will emulate human behavior under all circumstances" is useful.
Hide20

It’s true any job can find unqualified applicants. What I’m saying is that this in particular relies on an untenably small niche of feasible candidates that will take an enormous amount of time to find/filter through on average.

Sure, you might get lucky immediately, but without a reliable way to find the “independently wealthy guy who’s an intellectual and is sufficiently curious about you specifically that he wants to sit silently and watch you for 8 hours a day for a nominal fee”, your recruitment time will, on average, be very long, especially in compar... (read more)

Hide10

Do you genuinely think that you can find such people “reliably”?

2nim
Oops! I only realized in your reply that you're considering "reliability" the load-bearing element. Yes, the hiring pipeline will look like a background noise of consistent interest from the unqualified, and sporadic hits from excellent candidates. You're approaching it from the perspective that the background noise of incompetents is the more important part, whereas I think that the availability of an adequate candidate eventually is the important part. I think this because basically anywhere that hires can reliably find unqualified applicants. For a role where people stay in the job for 6 months, for instance, you only need to find a suitable replacement once every 6 months... so "reliably" being able to find an excellent candidate every day seems simply irrelevant.
Hide0-7

Unless you’re paying gratuitously, the only people who would reliably be interested in doing this would be underqualified randoms. Expect all benefit to be counteracted by the time it takes to get into a productive rhythm, at which point they’ll likely churn in a matter of weeks anyway.

7nim
Counterexample: financially self-sufficient individual who is curious about the work that the thinker is doing, and wants to learn more of how it is done.
Hide238

I cannot imagine losing this game as the gatekeeper either, honestly.

Does anyone want to play against me? I’ll bet you $50 USD.

1Morpheus
I'd also bet $50 as a gatekeeper. I won this game as a gatekeeper before and now need someone to put my ego in place. I'd prefer to play against someone who won as the AI before. This post prompted me to wonder to which degree there might be publication bias going on in that people don't report when they "predictably" won as the gatekeeper (as I did).
3interstice
I also don't think I would lose as the gatekeeper(against a human), and would be willing to make a similar bet if anyone's interested.
Bojadła4228

I agree and I am putting my money where my mouth is.

I will play this game under the rules linked in the OP with me as the gatekeeper and anyone as the AI. I will bet at odds 10:1 (favorable to you) that I will not let the AI out. The minimum bet is my 10 k USD against your 1 k USD and the maximum bet my 100 k USD against your 10 k USD. We will use a mutually accepted LW community member as referee.

If you believe you have at least a 10% chance of making me let the AI out, you should take this bet. I predict no one will take me up on it.

I speculate that the ... (read more)

0the gears to ascension
I can think of lots of ways to manipulate someone but I can't think of any that would win this game with the constraint that I have to sound like GPT4
Hide00

I play on lichess often. I can tell you that a lichess rating of 2900 absolutely corresponds to grandmaster level strength. It is rare for FMs or IMs to exceed a 2800 blitz rating. Most grandmasters hover around 2600-2800 blitz.

Hide40

The discussion on attack surfaces is very useful, intuitive and accessible. If a better standalone resource doesn’t already exist, such a (perhaps expanded) list/discussion would be a useful intro for people unfamiliar with specific risks.

Hide10

This was excruciatingly frustrating to read, well done.

Hide50

This is well-reasoned, but I have difficulty understanding why this kind of takeover would be necessary from the perspective of a powerful, rational agent. Assuming AGI is indeed worth its name, it seems the period of time needed for it to "play nice" would be very brief.

AGI would be expected to be totally unconcerned with being "clean" in a takeover attempt. There would be no need to leave no witnesses, nor avoid rousing opposition. Once you have access to sufficient compute, and enough control over physical resources, why wait 10 years for humanity to be... (read more)

9Daniel Kokotajlo
It bothers with the charade until it no longer needs to. It's unclear how long that'll take.
Hide20

My first thought as well. IGF-1 exists for a reason. Growth is universally necessary for development, repair and function.

Hide10

shift the

 

Minor edit -  should be "shift in the"

Hide20

It's encouraging to see more emphasis recently on the political and public-facing aspects of alignment. We are living in a far-better-than-worst-case world where people, including powerful ones, are open to being convinced. They just need to be taught - to have it explained to them intuitively.

It seems cached beliefs produced by works like you get about five words have led to a passive, unspoken attitude among many informed people that attempting to explain anything complicated is futile. It isn't futile. It's just difficult. 

5Thane Ruthenis
I think "you get about five words" is mostly right... It's just that it's "five words per message", not "five words on the issue, total". You have to explain in short bursts, but you can keep building on your previous explanations.
Answer by Hide117

In another of your most downvoted posts, you say 

I kind of expect this post to be wildly unpopular

I think you may be onto something here.

Hide42

You can fail to get rid of balls. All of your energy and effort can go into not allowing something to crash or fall, averting each disaster shortly before it would be too late. Speaking for ten minutes with each of fifty of sources every day can be a good way to keep any of them from being completely neglected, but it’s a terrible way to actually finish any of those projects. The terminal stage of this is a system so tied up in maintaining itself and stopping from falling behind that it has no slack to clear tasks or to improve its speed.

 

This is the ... (read more)

Hide30

It seems intuitively bad:

  • Capabilities and accelerationist-focused researchers have gone from diluted and restrained to concentrated and encouraged
  • Microsoft now has unbounded control, rather than a 49% stake
  • Microsoft cannot be expected to have any meaningful focus on alignment/safety
  • They are not starting from scratch: a huge chunk of their most capable staff and leadership will be involved
  • The "superalignment" project will be at best dramatically slowed, and possibly abandoned if OpenAI implodes
  • Other major labs smell blood in the water, possibly exacerbating race dynamics, not to mention a superficial increase (by 1) in the number of serious players. 
Hide10

All good practices. Although, isn't this just "more metal", rather than "less ore"? I imagine one would want to maximize both the inputs and outputs, even if opportunities for increasing inputs are exhausted more quickly.

1Logan Kieller
This is a reasonable note and I do agree with you that the ideas presented in this essay only capture the increased "glean" of metal (the desirable output of happiness, insights, stories).  Less ore for me was both a creative liberty I took with the metaphor, but also represented a shift in what levers should be focused on. I would posit that most people now seek more metal with more ore. "I will be happier if I make more money" So, less ore is relative to this, not relative to the stable level of inputs a person would be currently at.  Otherwise, I would argue that not only are "opportunities for increasing inputs are exhausted more quickly", but some would run counterproductive to the fundamental goals of whatever optimization problem. It's hard to determine what will meaningfully contribute to vague and complex ideas like happiness, insightfulness, story-worthiness, etc. Would definitely be interested in expanding on these ideas in a further essay. 
3AnthonyC
In the case of the hedonic treadmill, possibly more inputs could have no effect at all? I was wondering this too.
Hide*-3-3

How is such a failure of imagination possible?

It's odd to claim that, contingent upon AGI being significantly smarter than us, and wanting to kill us, that there is no realistic pathway for us to be physically harmed. 

Claims of this sort by intelligent, competent people likely reveal that they are passively objecting to the contingencies rather than disputing whether these contingencies would lead to the conclusion.

The quotes you're responding to here superficially imply "if smart + malicious AI, it can't kill us", but it seems much more likely this is a warped translation of either "AI can't be smart", or "AI can't be malicious".

3Viliam
I imagine there could also be some unexplored assumption, such as "but we are many, and the AI is one" (a strong intuition that many always defeat one, which was true for our ancestors), and they don't realize that "one" superhuman AI could still do thousand things in parallel and build backups and extra robotic brains.
Hide0-2

I would happily play the role of B.

I do not have an established FIDE rating, but my strength is approximately 1850 FIDE currently (based on playing against FIDE rated players OTB quite often, as well as maintaining 2100-2200 blitz ratings on Lichess & Chess.com, and 2200-2300 bullet). I'd be available after 6:30 pm (UTC+10) until ~12:00 pm (UTC+10). Alternatively, weekends are very flexible. I could do a few hours per week. 

I agree short/long time controls are a relevant, because speed is a skill that is almost entirely independent of conceptual knowledge and is mostly a function of baseline playing ability. 

 

Edit: Would also be fine with C

Hide120

Strongly agree. To my utter bewilderment, Eliezer appears to be exacerbating this vulnerability by making no efforts whatsoever to appear credible to the casual person. 

In nearly all of his public showings in the last 2 years, he has:

  • Rocked up in a trilby
  • Failed to adequately introduce himself
  • Spoken in condescending, aloof and cryptic tones; and
  • Failed to articulate the central concerns in an intuitive manner

As a result, to the layperson, he comes off as an egotistical, pessimistic nerd with fringe views - a perfect clown from which to retreat to a "mid... (read more)