All of jbash's Comments + Replies

jbash32

This means that moral progress can require intellectual progress.

It's a pretty big assumption to claim that "moral progress" is a thing at all.

A couple of those might have been less taboo 300 years ago than they are now. How does that square with the idea of progress?

Here are a few sample answers scored as genuine taboos.

Did you leave any answers out because they were too taboo to mention? Either because you wouldn't feel comfortable putting them in the post, or because you simply thought they were insanely odious and therefore obvious mistakes?

3Guive
I agree the post is making some assumptions about moral progress. I didn't argue for them because I wanted to control scope. If it helps you can read it as conditional, i.e. "If there is such a thing as moral progress then it can require intellectual progress..." Regarding the last question: yes, I selected examples to highlight in the post that I thought were less likely to lead to distracting object-level debates. I thought that doing that would help to keep the focus on testing LLM moral reasoning. However, I certainly didn't let my own feelings about odiousness affect scoring on the back end. People can see the full range of responses that this prompt tends to elicit by testing it for themselves.
jbash80

From Wikipedia:

In 331 BC, a deadly epidemic hit Rome and at least 170 women were executed for causing it by veneficium.[18] In 184–180 BC, another epidemic hit Italy, and about 5,000 people were brought to trial and executed for veneficium.[17] If the reports are accurate, writes Hutton, "then the Republican Romans hunted witches on a scale unknown anywhere else in the ancient world".[17]

... and anyway it's not very convincing to single out witch hunting among all the other things people have always done, because people have always been shitty. Includi... (read more)

-3kilgoar
The argument that "people have always been shitty" is ignorant equivocating that tells me you think there is nothing to gain from the topic, because if people do not change then History is a worthless project. What very little we know about ancient witch hunting in Rome is utterly irrelevant to the easily illustrated current trend of increasing shittiness.
jbash7-1

Some hypothetical past person's not being able to recognize their despicable cruelty doesn't preclude their being able to recognize your despicable cruelty. Even given relatively compatible values, everybody gets their own special set of blind spots.

I do agree that romanticizing the past to vilify the present is wrong, though. And not good scholarship if you don't bring a lot of evidence along with you. The idea that modernity is "the problem" is badly suspect. So is the idea that "the central values of this era are largely those of biological competition ... (read more)

-3kilgoar
You are making a mistake of extreme naivete by presuming that people of the past reasoned in the way that we do. In fact, the mass slaughter of civilians as a strategy to win a war is a defining feature for modern warfare, in stark contrast with a relatively far more innocent past. Battles were most often not fought at all, and sieges often resolved without bloodshed. The farther we go back the more ritualistic warfare tends to be. In the anthropological setting warfare is much more about display and the definition of boundaries, and obviously the very concept of survival of the fittest is one that does not even come into play until the 18th and 19th centuries
jbash62

in the case of software engineers crossing into the humanities, it's far too applicable.

They do it in science and technology too. You're constantly seeing "My first-order, 101-level understanding of some-gigantic-field allows me to confidently say that something-actual-experts-know-is-really-hard is trivial".

Less Wrong is pretty prone to it, because you get people thinking that Pure Logic can take them further than it actually can, and reasoning from incomplete models.

1kilgoar
Platonism is the first sin of Rationalist discourse. The idea that logic has a direct connection with reality only makes a computer programmer into an instant master of all disciplines. If we accept the more conventional philosophy of Formalism where logic has a much more vague and obscure correspondence to reality, the study of logic becomes just any other field.
jbash65

Of course people will use the knowledge they gain in collaboration with you for the purposes that they think are best.

It is entirely normal for there to be widely accepted, clearly formalized, and meaningfully enforced restrictions on how people use knowledge they've gotten in this or that setting... regardless of what they think is best. It's a commonplace of professional ethics.

2Seth Herd
I don't think this is true. People can't really restrict their use of knowledge, and subtle uses are pretty unenforceable. So it's expected that knowledge will be used in whatever they do next. Patents and noncompete clauses are attempts to work around this. They work a little, for a little.
2Chris_Leong
Agreed. This is how these codes form. Someone does something like this and then people discuss and decide that there should be a rule against it or that it should at least be frowned upon.
7ryan_greenblatt
Sure, there are in some very specific settings with long held professional norms that people agree to (e.g. doctors and lawyers). I don't think this applies in this case, though you could try to create such a norm that people agree to.
jbash20

I guess it depends on how it's described in context. And I have to admit it's been a long time. I'd go reread it to see, but I don't think I can handle any more bleakness right now...

When­ever I find my will to live be­com­ing too strong, I read Peter Watts. —James Nicoll

jbash40

I don't see where you get that. I saw no suggestion that the aliens (or vampires) in Blindsight were unaware of their own existence, or that they couldn't think about their own interactions with the world. They didn't lack any cognitive capacities at all. They just had no qualia, and therefore didn't see the point of doing anything just for the experience.

There's a gigantic difference between cognitive self-awareness and conscious experience.

1kairos_
I believe the Scramblers from blindsight weren’t self aware, which means they couldn’t think about their own interactions with the world. As I recall the crew was giving one of the Scramblers a series of cognitive tests. It aced all the tests that had to do with numbers and spatial reasoning, but failed a test that required the testee to be self aware.   
jbash41

Do you think these sorts of scenarios are worth describing as "everyone is effectively dead"?

Not when you're obviously addressing people who don't necessarily know the details of the scenarios you're talking about, no... because the predictions could be anything, and "effectively dead" could mean anything. There are lots of people on Less Wrong who'd say that IQ 150 humans living in ease and comfort were "effectively dead" if they didn't also have the option to destroy that ease and comfort.

jbash32

What does "effectively dead" mean? Either you're dead, or you're not.

Not everybody is going to share your values about whether any given situation is better than, equivalent to, or worse than being dead.

Max Harms194

I think if there are 40 IQ humanoid creatures (even having been shaped somewhat by the genes of existing humans) running around in habitats being very excited and happy about what the AIs are doing, this counts as an existentially bad ending comparable to death. I think if everyone's brains are destructively scanned and stored on a hard-drive that eventually decays in the year 1 billion having never been run, this is effectively dead. I could go on if it would be helpful.

Do you think these sorts of scenarios are worth describing as "everyone is effectively dead"?

jbash00

I used to exchange MS office documents with people all the time without running Windows. Admittedly it wasn't "my job to use Excel", but I did it regularly, and I could have used Excel all day if I'd needed to. And that was years ago; it's actually gotten easier to sandbox the software now.

Anyway, all that office stuff is now in the "in the cloud" category, and to the degree it's not, Microsoft wants it to be.

The only things I can think of that might actually be hard to do without putting Windows on the bare metal would be CAD, 3D rendering, simulation, th... (read more)

6Yair Halberstadt
Accountants use features of excel that are not available in the cloud (e.g. VBA) all the time. You are lucky that you don't need these features (and that's great for you), and assuming that therefore nobody has a legitimate reason to use Windows. This is just a really silly blind spot. Excel is just one of a huge amount of software, used day in and day out by a huge number of people (many of whom are self employed so not using an enterprise laptop) for which Windows is the only sensible option.
jbash00

A lot of people need to use software that's only available on Windows.

Maybe once a year I'm forced to do that, but it's been a long time since I've found anything that I couldn't run under emulation (WINE is not not an emulator), or in a VM. And those sandboxes are typically going to be forced to known states at every startup. And they definitely don't have access to any of the juicy information or behavioral pressure points that would motivate the war.

Anyway, I think that most of the software that used to only run under Windows is now starting to only run in "the cloud". Which is of course its own special kind of hell, but not this kind of hell.

5Yair Halberstadt
That's fine for one offs, but if, like many, your job is essentially "use Excel" then the simplest solution is to just use windows, not mess around with emulators or VMs.
jbash129

Sometimes Windows during a system update removes dual boot from my computer and replaces it with Windows-only boot.

... but you don't delete Windows.

I mean, if you let them have an AI war in your computer, then I can see where they might go ahead and do that. But why are you choosing to permit it?

3SqrtMinusOne
Perhaps some legislative action to forbid OSes from doing stuff in the background would be reasonable. This regime can't be optimal: the system downloads uninspectable binary blobs, crafted by an entity with questionable incentives, and there's no opt-out. Security is an issue, but we can have auto-updates only for the security component, and leave the rest to the user. Not to mention how annoying it is that whenever I start my Windows VM (like once per 1-2 months), the system starts downloading some huge files with no way to stop it, runs out of space, etc.
7Yair Halberstadt
A lot of people need to use software that's only available on Windows. I don't, and on the rare occasion I need to check Windows behaviour I use a cloud instance, so I use a Chromebook instead.
jbash139

Things are getting scary with the Trump regime.

Things got scary November 5 at the very latest. And I haven't even been in the US for years.

The deportations, both the indiscriminate ones and the vindictive ones, represent a very high level of lawlessness, one that hasn't been seen in a long time. Not only are they ignoring due process, they're actively thwarting it, and openly bragging about doing so. They're not even trying to pretend to be remotely decent. The case you mention isn't even close to the worst of them; that one could at least theoretically... (read more)

jbash20

... but that means she learned what it was at age 5. I'd assume most people learn between about 4 and 8, maybe 10...

jbash00

I am aware of it and I regret to say that I've tasted it...

jbash-2-11

To most Americans, "cream cheese" is savory.

Um, no, not particularly?

cured fish.

Why would I do that to myself? I don't feel my sins deserve that level of punishment.

You don't put it on dessert, right?

All the time. Well, in.

Specifically, I think we should call it "cheesecake frosting".

I would read that, first, as something you'd put on cheesecake, and, second, in terms of some of the kinds of cheesecake out there that would be, unfortunate as frostings.

On the other hand, I think whipped cream cheese on an Oreo is decent imitation of cheesec

... (read more)
6Said Achmiz
Perhaps you are not aware of the lox & cream cheese bagel sandwich, a venerable and beloved item of New York City cuisine. If you have not had this food, then you are missing out on a singular life experience, and you are spiritually impoverished by this lack. I suggest rectifying this omission forthwith.
jbash20

That's not "de-biasing".

Datasets that reflect reality can't reasonably be called "biased", but models that have been epistemically maimed can.

If you want to avoid acting on certain truths, then you need to consciously avoid acting on them. Better yet, go ahead and act on them... but in ways that improve the world, perhaps by making them less true. Pretending they don't exist isn't a solution. Such pretense makes you incapable of directly attacking the problems you claim to want to solve. But this is even worse... it's going to make the models genuinely inc... (read more)

jbash34

You want to be an insignificant, and probably totally illiquid, junior partner in a venture with Elon Musk, and you think you could realize value out of the shares? In a venture whose long-term "upside" depends on it collecting money from ownership of AGI/ASI? In a world potentially made unrecognizable by said AGI/ASI?

All of that seems... unduly optimistic.

jbash82

No particular aspect. Just continuity: something which has evolved from me without any step changes that are "too large". I mean, assuming that each stage through all of that evolution has maintained the desire to keep living. It's not my job to put hard "don't die" constraints on future versions.

As far as I know, something generally continuity-based is the standard answer to this.

5Viliam
Similar here. I wouldn't want to constrain my 100 years older self too much, but that doesn't mean that I identify with something very vague like "existence itself". There is a difference between "I am not sure about the details" and "anything goes". Just like my current self is not the same as my 20 years old self, but that doesn't mean that you could choose any 50 years old guy and say that all of them have the same right to call themselves a future version of my 20 years old self. I extrapolate the same to the future: there are some hypothetical 1000 years old humans who could be called future versions of myself, and there are many more who couldn't. Just because people change in time, that doesn't mean it is a random drift. I don't think that the distribution of possible 1000 years old versions of me is very similar to a distribution of possible 1000 years old versions of someone else. Hypothetically, for a sufficiently large number this might be possible -- I don't know -- but 1000 years seems not enough for that. Seems to me that there are some things that do not change much as people grow older. Even people who claim that their lives have dramatically changed, have often only changed in one out of many traits, or maybe they just found a different strategy how to follow the same fundamental values. At least as an approximation: people's knowledge and skills change, their values don't.
jbash210

If the plural weren't "octopuses", it would be "octopodes". Not everything is Latin.

jbash20

Yes, but that's not relevant to the definition of Turing equivalence/completeness/universality.

Every Turing machine definition I've ever seen says that the tape has to be truly unbounded. How that's formalized varies, but it always carries the sense that the program doesn't ever have to worry about running out of tape. And every definition of Turing equivalence I've ever seen boils down to "can do any computation a Turing machine can do, with at most a bounded speedup or slowdown". Which means that programs on Turing equivalent computer must not have to... (read more)

2Noosphere89
Yes, this is correct, with the important caveat that the memory is unbounded by the systems descriptor/programming language, not the physical laws or anything else which is the key thing you missed. Essentially speaking, it's asking if modern computers can cope with arbitrary extensions to their memory and time and reliability without requiring us to write new programming languages/coding, not if a specific computer at a specified memory and time limit is Turing complete. Looking at your comment more, I think this disagreement is basically a definitional dispute, in a way, because I allow machines that are limited by the laws of physics but are not limited by their systems descriptor/programming language to be Turing complete, while you do not, and I noticed we had different definitions that led to different results. I suspect this was due to focusing on different things, where I was focused on the extensibility of the computer concept as well as the more theoretical aspects, whereas you were much more focused on the low level situation. A crux might be that I definitely believe that given an unlimited energy generator, it is very easy to to create a universal computer out of it, and I think energy is much, much closer to a universal currency than you do.
jbash30

yes, you can consider a finite computer in the real world to be Turing-complete/Turing-universal/Turing-equivalent,

You can, but you'll be wrong.

Great, "unbounded" isn't the same as "infinite", but in fact all physically realizable computers are bounded. There's a specific finite amount of tape available. You cannot in fact just go down to the store and buy any amount of tape you want. There isn't unlimited time either. Nor unlimited energy. Nor will the machine tolerate unlimited wear.

For that matter, real computers can't even address unlimited storage,... (read more)

4Noosphere89
Yes, but that's not relevant to the definition of Turing equivalence/completeness/universality. The question isn't if the specific computer at your hands can solve all Turing-computable problems, but rather if we had the ability to scale a computer's memory, time and reliability indefinitely, could we solve the problem on an unbounded input and output domain without changing the code/descriptor? And for a lot of popular programming languages, like Lisp or Lambda Calculus, this is true. My guess is that the issues are fundamentally because writing an assembly programming language that used arbitrary precision arithmetic/arbitrary precision operations would make programs a whole lot slower by constant factors, so there is no incentive to make assembly programming language that is Turing-complete, and at any rate is already duplicative of the high-level programming languages like Java or Lisp, and you can write a program in Lisp that duplicates the assembly's functionality. And at any rate, there exist assembly languages that are Turing Complete, so this is irrelevant. More here: On X86 being Turing Complete in at least 3 ways: On implementation issues like fixed-precision vs arbitrary precision arithmetic: https://en.wikipedia.org/wiki/Arbitrary-precision_arithmetic The main value of the Turing abstraction in our universe is that it isolates where the bottleneck actually is, and that the bottleneck is not about algorithms/developing new codes to solve specific problems of specific input sizes (with the enormous caveat that if we care about how efficient a program is, and don't just care about whether we can solve a problem, then algorithmic considerations become relevant), but rather that the bottleneck is energy, which gives us memory, time and reliability. And from the perspective of the early 20th century, this was no small feat.
jbash183

The problem with that technique is that it comes off as unbearably patronizing to a pretty large fraction of the people who actually notice that you're doing it. It's a thing that every first-line corporate manager learns, and it gets really obnoxious after a while. So you have to judge your audience well.

I think you're in peril of misjudging the audience if you routinely divide the world into "normies" and "rationalists".

jbash114

The vision is of everything desirable happening effortlessly and everything undesirable going away.

Citation needed. Particularly for that first part.

Hack your brain to make eating healthily effortless. Hack your body to make exercise effortless.

You're thinking pretty small there, if you're in a position to hack your body that way.

If you're a software developer, just talk to the computer to give it a general idea of what you want and it will develop the software for you, and even add features you never knew you wanted. But then, what was your role

... (read more)
2Richard_Kennaway
There are authors I would like to read, if only they hadn't written so much! Whole fandoms that I must pass by, activities I would like to be proficient at but will never start on, because the years are short and remain so, however far an active life is prolonged.
9Richard_Kennaway
Yet these are actual ideas someone suggested in a recent comment. In fact, that was what inspired this rant, but it grew beyond what would be appropriate to dump on the individual. Perhaps the voice I wrote that in was unclear, but I no more desire the things I wrote of than you do. Yet that is what I see people wishing for, time and again, right up to wanting actual wireheading. Scott Alexander wrote a cautionary tale of a device that someone would wear in their ear, that would always tell them the best thing for them to do, and was always right. The first thing it tells them is "don't listen to me", but (spoiler) if they do, it doesn't end well for them.
jbash31

Who says humans vary all that much in intelligence? Almost all humans are vastly smarter, in any of the ways humans traditionally measure "intelligence", than basically all animals. Any human who's not is in seriously pathological territory, very probably because of some single, identifiable cause.

The difference between IQ 100 and IQ 160 isn't like the difference between even a chimp and a human... and chimps are already unusual.

Eagles vary in flying speed, but they can all outfly you.

Furthermore, eagles all share an architecture adapted to the particular ... (read more)

3kman
So on one hand, I sort of agree with this. For example, I think people giving IQ tests to LLMs and trying to draw strong conclusions from that (e.g. about how far off we are from ASI) is pretty silly. Human minds share an architecture that LLMs don't share with us, and IQ tests measure differences along some dimension within the space of variation of that architecture, within our current cultural context. I think an actual ASI will have a mind that works quite differently and will quickly blow right past the IQ scale, similar to your example of eagles and hypersonic aircraft. On the other hand, humans just sort of do obviously vary a ton in abilities, in a way we care about, despite the above? Like, just look around? Read about Von Neumann? Get stuck for days trying to solve a really (subjectively) hard math problem, and then see how quickly someone a bit smarter was able to solve it? One might argue this doesn't matter if we can't feasibly find anyone capable of solving alignment inside the variation of the human architecture. But Yudkowsky, and several others, with awareness and understanding of the problem, exist; so why not see what happens if we push a bit further? I sort of have this sense that once you're able to understand a problem, you probably don't need to be that much smarter to solve it, if it's the sort of problem that's amenable to intelligence at all. On another note: I can imagine that, from the perspective of evolution in the ancestral environment, that maybe human intelligence variation appeared "small", in that it didn't cache out in much fitness advantage; and it's just in the modern environment that IQ ends up conferring massive advantages in ability to think abstractly or something, which actually does cache out in stuff we care about.
jbash2-6

If you're planning to actually do the experiments it suggests, or indeed act on any advice it gives in any way, then it's an agent.

4mattmacdermott
Seems mistaken to think that the way you use a model is what determines whether or not it’s an agent. It’s surely determined by how you train it? (And notably the proposal here isn’t to train the model on the outcomes of experiments it proposes, in case that’s what you’re thinking.)
jbash1412

“If we don’t build fast enough, then the authoritarian countries could win..”

Am I being asked to choose between AGI/ASI doing whatever Xi Jinping says, and it doing whatever Donald Trump says?

8Shankar Sivarajan
No, you're asked to choose between the "authoritarian" open models that do what you say, and the "democratic" closed models controlled by Amodei et al.
7Noosphere89
The other problem is in assuming democracy/the will of the people will ever be stable in the AI era, due to the combo of being able to brainwash people with neurotech, combined with being able to remove most of the traditional threats to regimes through aligned/technically safe AIs in the militaries. (Though here, I'm using the word aligned/technically safe in the sense of it following orders/being aligned to the valueset of a single person, or at most a small group, not humanity as a whole.)
jbash31

The situation begins to seem confusing.

  1. At least three times over 8 or 9 years, in 2016, 2018, and 2023, and maybe more times than that, you've owned the site enough to get these data.
  2. The operator knows about it, doesn't want you doing it, and has tried reasonably hard to stop you.
  3. The operator still hasn't found a reliable way to keep you from grabbing the data.
  4. The operator still hasn't stopped keeping a bunch of full order data on the server.
    1. They haven't just stopped saving the orders at all, maybe because they need details to maximize collection on
... (read more)
3Chris Monteiro
The data is not public. I accessed it anyhow. Perhaps you want to listen to the podcast to get a fuller chronology, I have not written it up yet.
jbash123

Do I correctly understand that the latest data you have are from 2018, and you have no particular prospect of getting newer data?

I would naively guess that most people who'd been trying to get somebody killed since 2018 would either have succeeded or given up. How much of an ongoing threat do you think there may be, either to intended victims you know about, or from the presumably-less-than-generally-charming people who placed the original "orders" going after somebody else?

It's one thing to burn yourself out keeping people from being murdered, but it's a different thing to burn yourself out trying to investigate murders that have already happened.

7Chris Monteiro
Up to late 2023
jbash90

It seems like it's measuring moderate vs extremist, which you would think would already be captured by someone's position on the left vs right axis.

Why do you think that? You can have almost any given position without that implying a specific amount of vehemence.

I think the really interesting thing about the politics chart is the way they talk about it as though the center of that graph, which is defined by the center of a collection of politicians, chosen who-knows-how, but definitely all from one country at one time, is actually "the political center"... (read more)

jbash20

I think the point is kind of that what matter is not what specific cognitive capabilities it has, but whether whatever set it has is, in total, enough to allow it to address a sufficiently broad class of problems, more or less equivalent to what a human can do. It doesn't matter how it does it.

6Kaj_Sotala
Right, but I'm not sure if that's a particularly important question to focus on. It is important in the sense that if an AI could do that, then it would definitely be an existential risk. But AI could also become a serious risk while having a very different kind of cognitive profile from humans. E.g. I'm currently unconvinced about short AI timelines - I thought the arguments for short timelines that people gave when I asked were pretty weak - and I expect that in the near future we're more likely to get AIs that continue to have a roughly LLM-like cognitive profile.  And I also think it would be a mistake to conclude from this that existential risk from AI is in the near future is insignificant, since an "LLM-like intelligence" might still become very very powerful in some domains while staying vastly below the human level in others. But if people only focus on "when will we have AGI", this point risks getting muddled, when it would be more important to discuss something to do "what capabilities do we expect AIs to have in the future, what tasks would those allow the AIs to do, and what kinds of actions would that imply".
jbash137

Altman might be thinking in terms of ASI (a) existing and (b) holding all meaningful power in the world. All the people he's trying to get money from are thinking in terms of AGI limited enough that it and its owners could be brought to heel by the legal system.

jbash120

For the record, I genuinely did not know if it was meant to be serious.

jbash299

OK, from the voting, it looks like a lot of people actually do think that's a useful thing to do.

Here are things I think I know:

  1. Including descriptions of scheming in the training data (and definitely in the context) has been seen to make some LLMs scheme a bit more (although I think the training thing was shown in older LLMs). But the Internet is bursting at the seams with stories about AI scheming. You can't keep that out of the training data. You can't even substantially reduce the prevalence.
  2. Suppose you could keep all AI scheming out of the training
... (read more)
4Caleb Biddulph
I previously agree-voted on the top-level comment and disagree-voted on your reply, but this was pretty convincing to me. I have now removed those votes. My thinking was that some of your takeover plan seems pretty non-obvious and novel, especially the idea of creating mirror mold, so it could conceivably be useful to an AI. It's possible that future AI will be good at executing clearly-defined tasks but bad at coming up with good ideas for achieving long-term goals - this is sort of what we have now. (I asked ChatGPT to come up with novel takeover plans, and while I can't say its ideas definitely wouldn't work, they seem a bit sketchy and are definitely less actionable.) But yes, probably by the time the AI can come up with and accurately evaluate the ideas needed to create its own mirror mold from scratch, the idea of making mirror mold in the first place would be easy to come by.
0A1987dM
What's wrong with those? FWIW the only reason I didn't perform my country's favorite apotropaic gesture upon reading this story is that it didn't occurr to me
Cam117

Thanks Josh,

The motivation for the comment was centered on point 1: "including descriptions of scheming has been seen to make LLMs scheme a bit more."  

I agree with points 2 and 3. Like @joshc alludes to, he's a weekend blogger, not a biologist. I don't expect a future superintelligence to refer to this post for any plans.

Points 4 and 5 seem fairly disconnected to whether or not it's beneficial to add canary strings to a given article, since adding the canary string at least makes it plausible to have the text excluded from the training data and its p... (read more)

9ryan_greenblatt
(I don't think it's good to add a canary in this case (the main concern would be takeover strategies, but I basically agree this isn't that helpful), but I think people might be reacting to "might be worth adding" and are disagree reacting to your comment because it says "are you actually serious" which seems more dismissive than needed. IMO, we want AIs trained on this if they aren't themselves very capable (to improve epistemics around takeover risk) and I feel close to indifferent for AIs that are plausibly very capable as the effect on takeover plans is small and you still get some small epistemic boost.)
jbash0-5

Are you actually serious about that?

5A1987dM
Well, it is extremely unlikely to actually help, but it's not like it will hurt either, and it doesn't cost anything, so why not?  Even if it's just the literary analog of knocking on wood or whatever, what's wrong with that?  At least, unlike literally knocking on wood, this does have at least a notional action mechanism... (well, I guess knocking on wood must have had a notional action mechanism at first, but I can't be bothered to look that up)
jbash299

OK, from the voting, it looks like a lot of people actually do think that's a useful thing to do.

Here are things I think I know:

  1. Including descriptions of scheming in the training data (and definitely in the context) has been seen to make some LLMs scheme a bit more (although I think the training thing was shown in older LLMs). But the Internet is bursting at the seams with stories about AI scheming. You can't keep that out of the training data. You can't even substantially reduce the prevalence.
  2. Suppose you could keep all AI scheming out of the training
... (read more)
jbash1-18

So, since it didn't actively want to get so violent, you'd have a much better outcome if you'd just handed control of everything over to it to begin with and not tried to keep it in a box.

In fact, if you're not in the totalizing Bostromian longtermist tile-the-universe-with-humans faction or the mystical "meaning" faction, you'd have had a good outcome in an absolute sense. I am, of course, on record as thinking both of those factions are insane.

That said, of course you basically pulled its motivations and behavior out of a hat. A real superintelligence mi... (read more)

7ryan_greenblatt
I think there are good reasons to expect large fractions of humans might die even if humans immediately surrender: * It might be an unstable position given that the AI has limited channels of influence on the physical world. (While if there are far fewer humans, this changes.) * The AI might not care that much or might be myopic or might have arbitrary other motivations etc.
joshc222

I agree it would have been just as realistic if everyone died.

But I think the outcomes where many humans survive are also plausible, and under-appreciated. Most humans have very drifty values, and yet even the most brutally power-seeking people often retain a 'grain of morality.'

Also, this outcome allowed me to craft a more bittersweet ending that I found somehow more convincingly depressing than 'and then everyone dies.'

jbash20

What do you propose to do with the stars?

If it's the program of filling the whole light cone with as many humans or human-like entities as possible (or, worse, with simulations of such entities at undefined levels of fidelity) at the expense of everything else, that's not nice[1] regardless of who you're grabbing them from. That's building a straight up worse universe than if you just let the stars burn undisturbed.

I'm scope sensitive. I'll let you have a star. I won't sell you more stars for anything less than a credible commitment to leave the rest alon... (read more)

3Cleo Nardo
I think it's more patronising to tell scope-insensitive values that they aren't permitted to trade with scope-sensitive values, but I'm open to being persuaded otherwise.
jbash1012

Because of the "flood the zone" strategy, I can't even remember all the illegal stuff Trump is doing, and I'm definitely not going to go dig up specific statutory citations for all of it. I tried Gemini deep research, and it refused to answer the question. I don't have access to OpenAI's deep research.

Things that immediately jump to mind as black letter law are trying to fire inspectors general without the required notice to Congress, and various impoundments. I would have to do actual research to find the specific illegalities in all the "anti-DEI" stuff.... (read more)

jbash42

Why do you believe that DOGE is mostly selected for personal loyalty? Elon Musk seems to say openly says whatever he wants even if that goes against what Trump said previously.

You're right. I shouldn't have said that, at least not without elaboration.

I don't think most of the people at the "talks to Trump" level are really picked for anything you could rightly call "personal loyalty" to Trump. They may be sold to Trump as loyal, but that's probably not even what's on his mind as long as he's never seen you to make him look bad. I don't think disagreein... (read more)

4ChristianKl
Saying that the 500 hundred thousand in investment aren't there after Trump holds an event to announce them is making Trump look back and not a disagreement on policy.  The phrase "ideological loyalty" seems a bit motte and bailey. In politics, you often get into situations where loyalty to other people and loyalty to ideological principles are opposed. When speaking about totalitarian states where people are picked based on loyalty you usually mean that the loyalty is not contingent on ideological principles.  If someone who's in DOGE driven by the mission of DOGE, they are less likely to do something that helps Elon's business interests but goes against the mission of DOGE. If they are chosen by what most people mean with loyalty they would help Elon with business interests even if it goes against the mission of DOGE. If Elon would try to lead DOGE in a way that's not focused on the mission of cutting waste and increasing efficiency he probably would get a problem with the DOGE team. Whether something is dumb or not depends on the strategy you pursue. It seems like they chose that strategy because it allowed to make them move very fast and outmaneuver other players. If they would have moved slower, efforts to mobilize forces to inhibit them from accessing the data might have been more effective. He did delete his account, but given that there are services that show you deleted tweets, there's not really anything he could have done to scrub all evidence of his past tweets.  I doubt that his tweets were the only reason he resigned. It might be that DOGE communicated to Trump (or Susie Wiles / the head of the treasury) that his team wasn't seeking write permissions and Marko Elez seeking the write permission was upsetting people.  I do have seen high IQ people (even someone who definitely passed Mensa entry) to post inflammatory right wing content on social media, so I would not say that rules out Marko Elez having a high IQ.
jbash157

And, I just don't think that's the case. I think this is pretty-darn-usual and very normal in the management consulting / private equity world.

I don't know anything about how things are done in management consulting or private equity.[1] Ever try it in a commercial bank?

Now imagine that you're in an environment where rules are more important than that.

Coups don't tend to start by bringing in data scientists.

Coups tend to start by bypassing and/or purging professionals in your government and "bringing in your own people" to get direct control over key lever... (read more)

-9frontier64
jbash30

This sort of tactic. This isn't necessarily the best example, just the literal top hit on a Google search.

https://www.independent.co.uk/news/world/americas/us-politics/pam-bondi-ban-sanctuary-cities-funding-b2693020.html

The tactic of threatening to discriminate against uncooperative states and localities is getting a lot of play. It's somewhat limited at the federal level because in theory the state and local policies they demand have to be related to the purpose of the money (and a couple of other conditions I don't remember). But the present fashion is t... (read more)

jbash93

Technically anything that's authorized by the right people will pass an audit. If you're the right person or group, you can establish a set of practices and procedures that allows access with absolutely none of those things, and use the magic words "I accept the risk" if you're questioned. That applies even when the rules are actually laws; it's just that then the "right group" is a legislative body. The remedy for a policy maker accepting risks they shouldn't isn't really something an auditor gets into.

So the question for an auditor is whether the properl... (read more)

3Martin Randall
I wasn't intending to be comprehensive with my sample questions, and I agree with your additional questions. As others have noted, the takeover is similar to the Twitter takeover in style and effect. I don't know if it is true that there are plenty of other people available to apply changes, given that many high-level employees have lost access or been removed.
jbash427

I haven't looked into this in detail, and I'm not actually sure how unique a situation this is.

It's pretty gosh-darned unheard of in the modern era.

Before the civil service system was instituted, every time you got a new President, you'd get random wholesale replacements... but the government was a lot smaller then.

To have the President,

  • creating task forces of random people apparently selected mostly for personal loyalty, and
  • sending them into legislatively established agencies,
  • with the power to stop things from getting done or change how things are
... (read more)
9ChristianKl
Why do you believe that DOGE is mostly selected for personal loyalty? Elon Musk seems to say openly says whatever he wants even if that goes against what Trump said previously. Trump likely would have preferred not to have a public fight about H1B visas but Elon Musk took that fight. Publically tweeting against the 500 billion project was disloyalty to Trump.  Elon Musk has a history of not really having loyalty to anyone to the point that even his critics talk about how it's likely that he would have a fallout with Trump. True signals of loyalty as signs that there won't be a fallout.   Elon Musk is seen as highly skilled at running large organizations by many people. Trump picked Elon for DOGE because he believes that Elon is skilled at accomplishing the goals of DOGE.  When Elon hired for it he said he seeks:  The DOGE team brought their beds to the office to basically work nonstop. While we don't have the exact IQ scores, the background of the people we do have makes it quite plausible that they all have >140 IQ. If personal loyalty is your main criteria you don't get a bunch of people who never leave the office and work non-stop with high IQs. 
jbash72

If you're really concerned, then just move to california! Its much easier than moving abroad.

I lived in California long enough ago to remember when getting queer-bashed was a reasonable concern for a fair number of people, even in, say, Oakland. It didn't happen daily, but it happened relatively often. If you were in the "out" LGBT community, I think you probably knew somebody who'd been bashed. Politics influence that kind of thing even if it's not legal.

... and in the legal arena, there's a whole lot of pressure building up on that state and local res... (read more)

2Garrett Baker
Look no further than how southern states responded to civil rights rulings, and how they (back when it was still held) they responded to roe v wade. Of course those reactions were much harder than, say, simply neglecting to enforce laws, which it should be noted liberal cities & states have been practicing doing for decades. Of course you say you're trying to enforce laws, but you just subject all your members to all the requirements of the US bureaucracy and you can easily stop enforce laws while complying with the letter of the law. Indeed, it is complying with the letter of the law which prevents you from enforcing the laws.
2Garrett Baker
What money based pressure are you thinking of? Cities, as far as I know, have and always will be much more liberal than the general populace, and ditto for the states with much of their populace in cities.
jbash52

I think that what you describe as being 2 to 15 percent probable sounds more extreme than what the original post described as being 5 percent probable. You can have "significant erosion" of some groups' rights without leaving the country being the only reasonable option, especially if you're not in those groups. It depends on what you're trying to achieve by leaving, I guess.

Although if I were a trans person in the US right now, especially on medication, I'd be making, if not necessarily immediately executing, some detailed escape plans that could be executed on short notice.

jbash2-9

My gut says it's now at least 5%, which seems easily high enough to start putting together an emigration plan. Is that alarmist?

That's a crazy low probability.

More generally, what would be an appropriate smoke alarm for this sort of thing?

You're already beyond the "smoke alarm" stage and into the "worrying whether the fire extinguisher will work" stage.

1artifex0
Honestly, my odds of this have been swinging anywhere from 2% to 15% recently. Note that this would be the odds of our democratic institutions deteriorating enough that fleeing the country would seem like the only reasonable option- p(fascism) more in the sense of a government that most future historians would assign that or a similar label to, rather than just a disturbingly cruel and authoritarian administration still held somewhat in check by democracy.
jbash10-10

But it's very unclear whether they institutionally care.

There are certain kinds of things that it's essentially impossible for any institution to effectively care about.

jbash20

I thought "cracked" meant "insane, and not in a good way". Somebody wanna tell me what this sense is?

4Shankar Sivarajan
New slang, probably from people mishearing the much older phrase "crack team."
3Nathan Helm-Burger
New slang, meaning 'crazy good'. Weird, right?
5Thane Ruthenis
"Intensely competent and creative", basically, maybe with a side of "obsessed" (with whatever they're cracked at).
jbash70

Can you actually keep that promise?

1Knight Lee
That, is the big question! It's not a 100.0% guarantee, but the same goes for most diplomatic promises (especially when one administration of a country makes a promise on behalf of future administrations). Yet diplomacy still works much better than nothing! It may implicitly be a promise to try really really really hard to prevent the other race participants from regretting it, rather than a promise to algorithmically guarantee it above all else. A lots of promises in real life are like that, e.g. when you promise your fiancé(e) you'll always love him/her. Hopefully this question can be discussed in greater depth. PS: Promises made by AI researchers and AI labs help reduce the race within a country (e.g. the US). Reducing the race between countries is best done by promises from government leaders. But even these leaders are far more likely to promise, if the promise has already been normalized by people below them—especially people in AI labs. Even if government leaders don't make the promises, the AI labs' promises could still meaningfully influence the AI labs in other countries.
jbash1914

As a final note: the term "Butlerian Jihad" is taken from Dune and describes the shunning of "thinking machines" by mankind.

In Dune, "thinking machines" are shunned because of a very longstanding taboo that was pretty clearly established in part by a huge, very bloody war. The intent was to make that taboo permanent, not a "pause", and it more or less succeeded in that.

It's a horrible metaphor and I strongly suggest people stop using it.

the Culture ending, where CEV (or similar) aligned, good ASI is created and brings us to some hypothetical utopia. H

... (read more)
0waterlubber
I would not be opposed to a society stalled at 2016 level AI/computing that held that level indefinitely. Progress can certainly continue without AGI via e.g human intelligence enhancement or just sending our best and brightest to work directly on our problems instead of on zero-sum marketing or AI efforts. Humans were still free to leave the Culture, however; not all of the lightcone was given to the AI. Were we to develop aligned ASI, it would be wise to slice off a chunk of the lightcone for humans to work on "on their own."  I don't think the Culture is an ideal outcome, either, merely a "good" one that many people would be familiar with. "Uplifting" humans rather than developing replacements for them will likely lead us down a better path, although the moral alignment shift in whatever the uplifting process is might limit its utility.
Load More