All of robo's Comments + Replies

robo40

Yes, it means figure out how the notation works.

robo10

That's a good Coasian point.  Talking out of my butt, but I think the airlines don't carry the risk.  The sale channel (airlines, Expedia, etc.) take commissions distributing an insurance product designed another company (Travel Insured International, Seven Corners) who handles product design compliance, with the actual claims being handled by another company and the insurance capital by yet another company (AIG, Berkshire Hathaway).

LLMs tell me the distributors get 30–50% commission, which tells you that it's not a very good product for consumers.

2Gurkenglas
I know less than you here, but last-minute flights are marked up because businesspeople sometimes need them and maybe TII/SC get a better price on those?
robo10

But fear of death does seem like a kind of value systematization

I don't think it's system 1 doing the systemization.  Evolution beat fear of death into us in lots of independent forms (fear of heights, snakes, thirst, suffocation, etc.), but for the same underlying reason.  Fear of death is not just an abstraction humans invented or acquired in childhood; is a "natural idea" pointed at by our brain's innate circuitry from many directions.  Utilitarianism doesn't come with that scaffolding.  We don't learn to systematize Euclidian and Minkowskian spaces the same way either.

robo90

Quick takes are presented inline, posts are not.  Perhaps posts could be presented as title + <80 (140?) character summary.

robo12

You may live in a place where arguments about the color of the sky are really arguments about tax policy.  I don't think I live there?  I'm reading your article saying "If Blue-Sky-ism is to stand a chance against the gravitational pull of Green-Sky-ism, it must offer more than talk of a redistributionist tax system" and thinking "...what on earth...?".  This might be perceptive cultural insight about somewhere, but I do not understand the context.  [This my guess as to why are you are being voted down]

robo184

You might be[1] overestimating the popularity of "they are playing god" in the same way you might overestimate the popularity of woke messaging.  Loud moralizers aren't normal people either.  Messages that appeal to them won't have the support you'd expect given their volume.

Compare, "It's going to take your job, personally".  Could happen, maybe soon, for technophile programmers!  Don't count them out yet.

  1. ^

    Not rhetorical -- I really don't know

robo10

Eliezer Yudkowsky wrote a story Kindness to Kin about aliens who love(?) their family members proportionally to the Hamilton's "I'd lay down my life for two brothers or eight cousins" rule.  It gives an idea to how alien it is.

Then again, Proto-Indo-European had detailed family words that correspond rather well to confidence of genetic kinship, so maybe it's a cultural thing.

robo7-2

Sure, I think that's a fair objection!  Maybe, for a business, it may be worth paying the marginal security costs of giving 20 new people admin accounts, but for the federal government that security cost is too high. Is that what people are objecting to? I'm reading comments like this:

Yeah, that's beyond unusual. It's not even slightly normal. And it is in fact very coup-like behavior if you look at coups in other countries.

And, I just don't think that's the case.  I think this is pretty-darn-usual and very normal in the management consulting / p... (read more)

jbash157

And, I just don't think that's the case. I think this is pretty-darn-usual and very normal in the management consulting / private equity world.

I don't know anything about how things are done in management consulting or private equity.[1] Ever try it in a commercial bank?

Now imagine that you're in an environment where rules are more important than that.

Coups don't tend to start by bringing in data scientists.

Coups tend to start by bypassing and/or purging professionals in your government and "bringing in your own people" to get direct control over key lever... (read more)

robo*244

Huh, I came at this with the background of doing data analysis in large organizations and had a very different take.

You're a data scientist.  You want to analyze what this huge organization (US government) is spending its money on in concrete terms.  That information is spread across 400 mutually incompatible ancient payment systems.  I'm not sure if you've viscerally felt the frustration of being blocked, spending all your time trying to get permission to read from 5 incompatible systems, let alone 400.  But it would take months or yea... (read more)

khafra*2818

As someone who has been allowed access into various private and government systems as a consultant, I think the near mode view for classified government systems is different for a reason. 


E.g., data is classified as Confidential when its release could cause damage to national security. It's Secret if it could cause serious damage to national security, and it's Top Secret if it could cause exceptionally grave damage to national security. 
People lose their jobs for accidentally putting a classified document onto the wrong system, even if it's still... (read more)

robo10

Checking my understanding: for the case of training a neural network, would S be the parameters of the model (along with perhaps buffers/state like moment estimates in Adam)?  And would the evolution of the state space be local in S space?  In other words, for neural network training, would S be a good choice for H?
In a recurrent neural networks doing in-context learning, would S be something like the residual stream at a particular token?

robo*10

I'll conjecture the following in a VERY SPECULATIVE, inflammatory, riff-on-vibes statements:

  • Gradient descent solves problem in the complexity class P[1].  It is P-Complete.
  • Learning theory (and complexity theory) have for decades been pushing two analogous bad narratives about the weakness of gradient descent (and P).
  • These narratives dominate because it is easy prove impossibility results like "Problem X can't be solved by gradient descent" (or "Problem Y is NP-Hard").  It's academically fecund -- it's a subject aspiring academics can write a lot
... (read more)
robo40

Thanks!  I'm not a GPU expert either.  The reason I want to spread the toll units inside GPU itself isn't to turn the GPU off -- it's to stop replay attacks.  If the toll thing is in a separate chip, then the toll unit must have some way to tell the GPU "GPU, you are cleared to run".  To hack the GPU, you just copy that "cleared to run" signal and send it to the GPU.  The same "cleared to run" signal must always make the GPU work, unless there is something inside the GPU to make sure won't accept the same "cleared to run" signal tw... (read more)

4Yonatan Cale
Oh yes the toll unit needs to be inside the GPU chip imo.   Alternatively the key could be in the central authority that is supposed to control the off switch. (same tech tho)   Nvidia (or whoever signs authorization for your GPU to run) won't sign it for you if you don't update the software (and send them a proof you did it using similar methods, I can elaborate).
robo20

I used to assume disabling a GPU in my physical possession would be impossible, but now I'm not so sure.  There might be ways to make bypassing GPU lockouts on the order of difficulty of manufacturing the GPU (requiring nanoscale silicon surgery). Here's an example scheme:

Nvidia changes their business models from selling GPUs to renting them.  The GPU is free, but to use your GPU you must buy Nvidia Dollars from Nvidia.  Your GPU will periodically call Nvidia headquarters and get an authorization code to do 10^15 more floating point operatio... (read more)

2Yonatan Cale
I love the direction you're going with this business idea (and with giving Nvidia a business incentive to make "authentication" that is actually hard to subvert)! I can imagine reasons they might not like this idea, but who knows. If I can easily suggest this to someone from Nvidia (instead of speculating myself), I'll try I'll respond to the technical part in a separate comment because I might want to link to it >>
robo30

Not a billion billion times.  You need ≈2^100 presses to get any signal, and ≈O(2^200) presses to figure out which way the signal goes.  2^200≈10^60.  Planck time's about 10^-45 seconds.  If you try to press the button more than 10^45 times per second the radiation the electrons in the button will emit will be so high frequency an small wavelength that it will collapse into a black hole.

robo10

Incentives for NIMBYism is an objection I've seldom seen stated.  "Of course I don't want to up-zone my neighborhood to allow more productive buildings -- that would triple my taxes!".

4cousin_it
People pretty much already say this about gentrification and rent. Like that joke about shooting in the air a few times every morning to keep your rent low. Maybe under LVT homeowners will end up doing the same.
robo70

You're being downvoted and nobody's telling you why :-(, so I thought I'd give some notes.

  • You're not talking to the right audience.  Few groups are more emotionally in favor of a glorious transhumanist future than people on LessWrong.  This is not technophobes who are afraid of change.  It's technophiles who have realized, in harsh constrat to the conclusion they emotionally want, that making a powerful AI would likely be bad for humanity.
  • Yes, it's important to overly anthropomorphize AIs, and you are doing that all over the place in your ar
... (read more)
Answer by robo72

Conjunction Fallacy.  Adding detail make ideas feel more realistic, and strictly less likely to be true.

Virtues for communication and thought can be diametrically opposed.

robo00

In a world where AI progress has wildly accelerated chip manufacture

 

This world?

1yams
Yes this world.
robo10

What distinction are you making between "visualising" and "seeing"?

Good question!  By "seeing" I meant having qualia, an apparent subjective experience.  By "visualizing" I meant...something like using the geometric intuitions you get by looking at stuff, but perhaps in a philosophical zombie sort of way?  You could use non-visual intuitions to count the vertices on a polyhedron, like algebraic intuitions or 3D tactile intuitions (and I bet blind mathematicians do).  I'm not using those.  I'm thinking about a wireframe image, drawn... (read more)

7Richard_Kennaway
I have qualia for imagined scenes. I'm not seeing them with my physical eyes, and they're not superimposed on the visual field that comes from my physical eyes. It's like they exist in a separate three-dimensional space that does not have any particular spatial relationship to the physical space around me.
robo*2-1

...I do not believe this test.  I'd be very good at counting vertices on a polyhedron through visualization and very bad at experiencing the sensation of seeing it.  I do "visualize" the polyhedra, but I don't "see" them.  (Frankly I suspect people who say they experience "seeing" images are just fooling themselves based on e.g. asking them to visualize a bicycle and having them draw it)

6Kaj_Sotala
I have a friend with eidetic imagination who says that for her, there is literally no difference between seeing something and imagining it. Sometimes she's worried about losing track of reality if she were to imagine too much.
6Matt Goldenberg
No, people really do see it, that whispiness can be crisp and clear I'm not the most visual person. But occasionally when I'm reading I'll start seeing the scene. I then get jolted out of it when I realize I don't know how I'm seeing the words as they've been replaced with the imagined visuals
7Richard_Kennaway
What distinction are you making between "visualising" and "seeing"? I've heard of that study about drawing bicycles. I can draw one just fine without having one before me. I have just done so, checked it, and every detail (that I included — this was just a two-minute sketch) was correct. Anyway, if people are as astonishingly bad at the task as the paper says, that just reflects on their memory, not the acuity of their mind's eye. I expect there are people who can draw a map of Europe with all the country borders, whereas I probably wouldn't even remember all of the countries.
6Eric Neyman
I think very few people have a very high-fidelity mind's eye. I think the reason that I can't draw a bicycle is that my mind's eye isn't powerful/detailed enough to be able to correctly picture a bicycle. But there's definitely a sense in which I can "picture" a bicycle, and the picture is engaging something sort of like my ability to see things, rather than just being an abstract representation of a bicycle. (But like, it's not quite literally a picture, in that I'm not, like, hallucinating a bicycle. Like it's not literally in my field of vision.)
robo10

Thanks for crossposting!  I've highly appreciated your contributions and am glad I'll continue to be able to see them.

robo*30

Quick summary of a reason why constituent parts like of super-organisms, like the ant of ant colonies, the cells of multicellular organisms, and endosymbiotic organelles within cells[1] are evolutionarily incentivized to work together as a unit:

Question: why do ants seem to care more about the colony than themselves?  Answer: reproduction in an ant colony is funneled through the queen.  If the worker ant wants to reproduce its genes, it can't do that by being selfish.  It has to help the queen reproduce.  Genes in ant workers have ... (read more)

3Yudhister Kumar
I haven't looked much at the extended phenotype literature, although that is changing as we speak. Thanks for pointing me in that direction! The thing I wanted to communicate was less "existing groups of things we call species are perfect examples of how super-organisms should work" and more "the definition of an ideal species captures something quite salient about what it means for a super-organism to be distinct from other super-organisms and its environment." In practice, yes, looking at structure does seem to be better.
1Purplehermann
Inside the super organism you are correct, but the genome is influenced by outside forces as whole over the ages - and any place where this breaks down for long enough you eventually get two species instead of one. Therefore outside groups can treat the species as a super organism in general, the individual members must be dealt with individually when there is previous loyalty to another member of the other species. For example, an Englishman and his dog vs an eskimo and his dog. The two humans may be against each other, the dogs may be against each other, but the opposite human/dog interactions would be standard if they weren't already attached to other in-species members.
robo*01

EDIT Completely rewritten to be hopefully less condescending.

There are lessons from group selection and the extended phenotype which vaguely reduce to "beware thinking about species as organisms".  It is not clear from this essay whether you've encountered those ideas.  It would be helpful for me reading this essay to know if you have.

3robo
Quick summary of a reason why constituent parts like of super-organisms, like the ant of ant colonies, the cells of multicellular organisms, and endosymbiotic organelles within cells[1] are evolutionarily incentivized to work together as a unit: Question: why do ants seem to care more about the colony than themselves?  Answer: reproduction in an ant colony is funneled through the queen.  If the worker ant wants to reproduce its genes, it can't do that by being selfish.  It has to help the queen reproduce.  Genes in ant workers have nothing to gain by making their ant more selfish and have much to gain by making their worker protect the queen. This is similar to why cells in your pancreas cooperate with cells in your ear.  Reproduction of genes in the body is funned through gametes.  Somatic evolution does pressure the cells in your pancreas to reproduce selfishly at the expense of cells in your ear (this is pancreatic cancer).  But that doesn't help the pancreas genes long term.  Pancreas-genes and the ear-genes are forced to cooperate with each other because they can only reproduce when bound together in a gamete. This sort of bounding together of genes making disperate things cooperate and act like a "super organism" is absent in members of a species.  My genes do not reproduce in concert with your genes.  If my genes figure out a way to reproduce at your expense, so much the better for them. 1. ^ Like mitochondria and chloroplasts, which were separate organisms but evolved to work so close with their hosts that they are now considered part of the same organism.
robo20

Hijacking this thread, has anybody worked through Ape in the coat's anthropic posts and understood / gotten stuff out of them?  It's something I might want to do sometime in my copious free time but haven't worked up to it yet.

2Ape in the coat
I'm very available to answer questions about my posts as soon as people actuall engage with the reasoning, so feel free to ask if you feel confused about anything.  If I am to highlight the core principle it would be: Thinking in terms of what happens in the probability experiment as a whole, to the best of your knowledge and from your perspective as a participant.  Suppose this experiment happened to you multiple times. If on iteration of the experiment something happens 2/3 of times then the probability of such event is 2/3. If something happens 100% of times then its probability is 1 and realizationof such event doesn't give you you any evidence.  All the rest is commentary. 
3Canaletto
I propose to sic o1 on them to distill it all into something readable/concise. (I tried to comprehend it and failed / got distracted). I think some people pointed out in comments that their model doesn't represent prob of "what day it is NOW" btw
robo21

Sorry, that was an off-the-cuff example I meant to help gesture towards the main idea.  I didn't mean to imply it's a working instance (it's not).  The idea I'm going for is:

  • I'm expecting future AIs to be less single LLMs (like Llama) and more loops and search and scaffolding (like o1)
  • Those AIs will be composed of individual pieces
  • Maybe we can try making the AI pieces mutually dependent in such a way that it's a pain to get the AI working at peak performance unless you include the safety pieces
robo2-3

This might be a reason to try to design AI's to fail-safe and break without controlling units.  E.g. before fine-tuning language models to be useful, fine-tune them to not generate useful content without approval tokens generated by a supervisory model.

2Seth Herd
I don't see how that would work technically. It seems like any small set of activating tokens would be stolen along with the weights, and I don't see how to train it for a large shifting set. I'm not saying this is impossible, just htat I'm not sure it is. Can you flesh this idea out any further?
robo*20

I suspect experiments with almost-genetically identical twin tests might advance our understanding about almost all genes except sex chromosomes.

Sex chromosomes are independent coin flips with huge effect sizes.  That's amazing!  Natural provided us with experiments everywhere!  Most alleles are confounded (e.g.. correlated with socioeconomic status for no causal reason) and have very small effect sizes.

Example: Imagine an allele which is common in east asians, uncommon in europeans, and makes people 1.1 mm taller.  Even though the alle... (read more)

robo*40

The player can force a strategy where they win 2/3 of the time (guess a door and never switch).  The player never needs to accept worse

The host can force a strategy where the player loses 1/3 of the time (never let the player switch).  The host never needs to accept worse.

Therefore, the equilibrium has 2/3 win for the player.  The player can block this number from going lower and the host can block this number from going higher.

robo30

I want to love this metaphor but don't get it at all.  Religious freedom isn't a narrow valley; it's an enormous Shelling hyperplane.  85% of people are religious, but no majority is Christian or Hindu or Kuvah'magh or Kraẞël or Ŧ̈ř̈ȧ̈ӎ͛ṽ̥ŧ̊ħ or Sisters of the Screaming Nightshroud of Ɀ̈ӊ͢Ṩ͎̈Ⱦ̸Ḥ̛͑..  These religions don't agree on many things, but they all pull for freedom of religion over the crazy *#%! the other religions want.

robo*50

Suppose there were some gears in physics we weren't smart enough to understand at all.  What would that look like to us?

It would look like phenomena that appears intrinsically random, wouldn't it?  Like imagine there were a simple rule about the spin of electrons that we just. don't. get.  Instead noticing the simple pattern ("Electrons are up if the number of Planck timesteps since the beginning of the universe is a multiple of 3"), we'd only be able to figure out statistical rules of thumb for our measurements ("we measure electrons as up ... (read more)

5Hastings
If you do set out on this quest, Bell's inequality and friends will at least put hard restrictions on where you could look for a rule underlying seemingly random wave function collapse. The more restricted your search, the sooner you'll find a needle!
robo*93

Humans are computationally bounded, Bayes is not.  In an ideal Bayesian perspective:

  • Your prior must include all possible theories a priori.  Before you opened your eyes as a baby, you put some probability of being in a universe with Quantum Field Theory with  gauge symmetry and updated from there.
  • Your update with unbounded computation.  There's not such thing as proofs, since all poofs are tautological.

Humans are computationally bounded and can't think this way.

(riffing)

"Ideas" find paradigms for modeling the univer... (read more)

1Abhimanyu Pallavi Sudhir
Yes, I also realized that "ideas" being a thing is due to bounded rationality -- specifically they are the outputs of AI search. "Proofs" are weirder though, and I haven't seen them distinguished very often. I wonder if this is a reasonable analogy to make: * Ideas : search * Answers : inference * Proofs: alignment
robo-40

Statements made to the media pass through an extremely lossy compression channel, then are coarse-grained, and then turned into speech acts.

That lossy channel has maybe one bit of capacity on the EA thing.  You can turn on a bit that says "your opinions about AI risk should cluster with your opinions about Effective Altruists", or not.  You don't get more nuance than that.[1]

If you have to choose between outputting the more informative speech act[2] and saying something literally true, it's more cooperative to get the output speech act corre... (read more)

robo12

If someone wants to distance themselves from a group, I don't think you should make a fuss about it.  Guilt by association is the rule in PR and that's terrible.  If someone doesn't want to be publicly coupled, don't couple them.

11a3orn
Whether someone is or was a part of a group is in general an actual fact about their history, not something they can just change through verbal disavowals. I don't think we have an obligation to ignore someone's historical association with a group in favor of parroting their current words. Like, suppose someone who is a nominee for the Supreme Court were to say "No, I totally was never a part of the Let's Ban Abortion Because It's Murder Group." But then you were to look at the history of this person and you found that they had done pro-bono legal work for the "Abortion Is Totally Murder" political action group; and they had founded an organization that turned out to be currently 90% funded by the "Catholics Against Murdering Babies"; and in fact had gone many times to "Let's Make Laws Be Influenced by the Catholic Church" summit; and he was a close personal friend to a bunch of archbishops and Catholic philosophers. In such a case, it's reasonable to be like "No, you're lying about what groups you were and are a part of." I think that you should be able to reasonably say this -- regardless of whether you think abortion is murder or not. The nominee is in fact lying; it is possible to lie about the group that you are a part of. Similarly -- well, the linked article from OP doesn't actually contain a disavowal from Dan Hendryks, afaict? This one contains the claim he was "never an EA adherent," which is closer to a disavowal. Whether or not this claim is true, it is the kind of claim that certainly admits truth. Or lies.
Answer by robo30

I think the classic answer to the "Ozma Problem" (how to communicate to far-away aliens what earthlings mean by right and left) is the Wu experiment.  Electromagnetism and the strong nuclear force aren't handed, but the weak nuclear force is handed.  Left-handed electrons participate in weak nuclear force interactions but right-handed electrons are invisible to weak interactions[1].

(amateur, others can correct me)

  1. ^

    Like electrons, right-handed neutrinos are also invisible to weak interactions.  Unlike electrons, neutrinos are also invisi

... (read more)
robo*30

Can you symmetrically put the atoms into that entangled state?  You both agree on the charge of electrons (you aren't antimatter annihilating), so you can get a pair of atoms into  |↑,↑⟩, but can you get the entangled pair to point in opposite directions along the plane of the mirror?

Edit Wait, I did that wrong, didn't I?  You don't make a spin up atom by putting it next to a particle accelerator sending electrons up.  You make a spin up atom by putting it next to electrons you accelerate in circles, moving the electrons in the direction your fingers point when a (real) right thumb is pointing up.  So one of you will make a spin-up atom and the other will make a spin-down atom.

2Charlie Steiner
Well, I was going by the original post where you're rotationally symmetric, not mirror-symmetric. But for mirror symmetry, pick the mirror plane to be the z direction, then you just make your spins either both +z or both -z.
robo54

No, that's a very different problem.  The matrix overlords are Laplace's demon, with god-like omniscience about the present and past.  The matrix overlords know the position and momentum of every molecule in my cup of tea.  They can look up the microstate of any time in the past, for free.

The future AI is not Laplace's demon.  The AI is informationally bounded.  It knows the temperature of my tea, but not the position and momentum of every molecule.  Any uncertainties it has about the state of my tea will increase exponentiall... (read more)

robo30

Oh, wait, is this "How does a simulation keep secrets from the (computationally bounded) matrix overlords?"

1lukemarks
This should be an equivalent problem, yes.
robo10

I don't think I understand your hypothetical.  Is your hypothetical about a future AI which has:

  • Very accurate measurements of the state of the universe in the future
  • A large amount of compute, but not exponentially large
  • Very good algorithms for retrodicting* the past

I think it's exponentially hard to retrodict the past.  It's hard in a similar way as encryption is hard.  If an AI isn't power enough to break encryption, it also isn't powerful enough to retrodict the past accurately enough to break secrets.

If you really want to keep something s... (read more)

3robo
Oh, wait, is this "How does a simulation keep secrets from the (computationally bounded) matrix overlords?"
1lukemarks
Yes, your description of my hypothetical is correct. I think it's plausible that approximating things that happened in the past is computationally easier than breaking some encryption, especially if the information about the past is valuable even if it's noisy. I strongly doubt my hypothetical will materialize, but I think it is an interesting problem regardless. My concern with approaches like the one you suggest is that they're restricted to small parts of the universe, so with enough data it might be possible to fill in the gaps.
robo51

I think a lot of people losing their jobs would probably do the trick, politics-wise.  For most people the crux is "will AIs will be more capable than humans", not "might AIs more capable than humans be dangerous".

1davekasten
You know, you're not the first person to make that argument to me recently.  I admit that I find it more persuasive than I used to. Put another way: "will AI take all the jobs" is another way of saying* "will I suddenly lose the ability to feed and protect those I love."  It's an apocalypse in microcosm, and it's one that doesn't require a lot of theory to grasp.   *Yes, yes, you could imagine universal basic income or whatever.  Do you think the average person is Actually Expecting to Get That ? 
robo20

I do this (with "sev") when counting to myself.  Nice to see the other people chose the same shelling point!

robo43

Yearly rent on the house is greater than yearly taxes on the house, right?  As you give the government shares of your house, tax shares will convert into rental shares and you will have to pay the government more and more.  A death spiral ensues and you lose the house.

"What if the government doesn't charge rent on its shares?"  Then everyone lets the government own 99% of their house to avoid the taxes.

A lot of the value of Georgism is incentivizing people who don't value a property to move out so people who do value the property can move in.

(off-the-cuff opinion)

1Benaya Koren
Right, missed that it makes your being there renting. How fast would the spiral be in your estimation? How fast would you lose the house in normal circumstances? By the way, I'm not sure that the tax is universally less than the rent. My understanding is that land speculation in Israel is mostly about expeting the prices to keep going up rather than the current rent.
robo30

I'm very not sure how to do this, but are there ways to collect some counteracting or unbiased samples about Sam Altman?  Or to do another one-sided vetting for other CEOs to see what the base rate of being able to dig up questionable things is?  Collecting evidence in that points in only one direction just sets off huge warning lights 🚨🚨🚨🚨 I can't quiet.

6gwern
Yes, it should. And that's why people are currently digging so hard in the other direction, as they begin to appreciate to what extent they have previously had evidence that only pointed in one direction and badly misinterpreted things like, say, Paul Graham's tweets or YC blog post edits or ex-OAer statements.
robo10

I think this is the sort of conversation we should be having!  [Side note: I think restricting compute is more effective than restricting research because you don't need 100% buy in.

  1. it's easier to prevent people from manufacturing semiconductors than to keep people from learning ideas that fit on a napkin
  2. It's easier to prevent scientists in Eaccistan from having GPUs than to prevent scientists in Eaccistan from thinking.

The analogy to nuclear weapons is, I think, a good one.  The science behind nuclear weapons is well known -- what keeps them fro... (read more)

5TsviBT
This is not true for AGI.
robo10

I think the weakness with KL divergence is that the potentially harmful model can do things the safe model would be exponentially unlikely to do.  Even if the safe model has a 1 in 1 trillion chance of stabbing me in the face, the KL penalty to stabbing me in the face is log(1 trillion) (and logs make even huge numbers small).

What about limiting the unknown model to chose one of the cumulative 98% most likely actions for the safe model to take?  If the safe model never has more than a 1% chance of taking an action that will kill you, then the unknown model won't be able to take an action that kills you.  This isn't terribly different from the Top-K sampling many language models use in practice.

robo12061

Our current big stupid: not preparing for 40% agreement

Epistemic status: lukewarm take from the gut (not brain) that feels rightish

The "Big Stupid" of the AI doomers 2013-2023 was AI nerds' solution to the problem "How do we stop people from building dangerous AIs?" was "research how to build AIs".  Methods normal people would consider to stop people from building dangerous AIs, like asking governments to make it illegal to build dangerous AIs, were considered gauche.  When the public turned out to be somewhat receptive to the idea of regulating ... (read more)

2Mitchell_Porter
There's a paper from ten years ago, "Testing Theories of American Politics: Elites, Interest Groups, and Average Citizens", which says that public opinion has very little effect on government, compared to the opinion of economic elites. That might be a start in figuring out what you can and can't do with that 40%.
2jbash
I think the "crux" is that, while policy is good to have, it's fundamentally a short-term delaying advantage. The stuff will get built eventually no matter what, and any delay you can create before it's built won't really be significant compared to the time after it's built. So if you have any belief that you might be able to improve the outcome when-not-if it's built, that kind of dominates.
8Ebenezer Dukakis
If LW takes this route, it should be cognizant of the usual challenges of getting involved in politics. I think there's a very good chance of evaporative cooling, where people trying to see AI clearly gradually leave, and are replaced by activists. The current reaction to OpenAI events is already seeming fairly tribal IMO.
Orpheus16*199

There are some conversations about policy & government response taking place. I think there are a few main reasons you don't see them on LessWrong:

  1. There really aren't that many serious conversations about AI policy, particularly in future worlds where there is greater concern and political will. Much of the AI governance community focuses on things that are within the current Overton Window.
  2. Some conversations take place among people who work for governments & aren't allowed to (or are discouraged from) sharing a lot of their thinking online.
  3. [Edited
... (read more)
Thomas Kwa4520

As recently as early 2023 Eliezer was very pessimistic about AI policy efforts amounting to anything, to the point that he thought anyone trying to do AI policy was hopelessly naive and should first try to ban biological gain-of-function research just to understand how hard policy is. Given how influential Eliezer is, he loses a lot of points here (and I guess Hendrycks wins?)

Then Eliezer updated and started e.g. giving podcast interviews. Policy orgs spun up and there are dozens of safety-concerned people working in AI policy. But this is not reflected in the LW frontpage. Is this inertia, or do we like thinking about computer science more than policy, or is it something else?

4TsviBT
Well I asked this https://www.lesswrong.com/posts/X9Z9vdG7kEFTBkA6h/what-could-a-policy-banning-agi-look-like but roughly no one was interested--I had to learn about "born secret" https://en.wikipedia.org/wiki/Born_secret from Eric Weinstein in a youtube video. FYI, while restricting compute manufacture is I would guess net helpful, it's far from a solution. People can make plenty of conceptual progress given current levels of compute https://www.lesswrong.com/posts/sTDfraZab47KiRMmT/views-on-when-agi-comes-and-on-strategy-to-reduce . It's not a way out, either. There are ways possibly-out. But approximately no one is interested in them.
gilch*184

the problem "How do we stop people from building dangerous AIs?" was "research how to build AIs".

Not quite. It was to research how to build friendly AIs. We haven't succeeded yet. What research progress we have made points to the problem being harder than initially thought, and capabilities turned out to be easier than most of us expected as well.

Methods normal people would consider to stop people from building dangerous AIs, like asking governments to make it illegal to build dangerous AIs, were considered gauche.

Considered by whom? Rationalists? T... (read more)

Haiku223

Strong agree and strong upvote.

There are some efforts in the governance space and in the space of public awareness, but there should and can be much, much more.

My read of these survey results is:
AI Alignment researchers are optimistic people by nature. Despite this, most of them don't think we're on track to solve alignment in time, and they are split on whether we will even make significant progress. Most of them also support pausing AI development to give alignment research time to catch up.

As for what to actually do about it: There are a lot of options... (read more)

6cousin_it
Yeah, that training took some time, but it worked. I can now write melodies and chords from imagination pretty easily. Have had this skill for awhile now. It's very useful, though of course not a golden ticket. My current challenge in music is just coming up with interesting stuff, I think this challenge isn't gonna run out anytime soon.
robo2-3

My comment here is not cosmically important and I may delete it if it derails the conversation.

There are times when I would really want a friend to tap me on the shoulder and say "hey, from the outside the way you talk about <X> seems way worse than normal.  Are you hungry/tired/too emotionally close?".  They may be wrong, but often they're right.
If you (general reader you) would deeply want someone to tap you on the shoulder, read on, otherwise this comment isn't for you.

If you burn at NYT/Cade Metz intolerable hostile garbage, are you hav... (read more)

FWIW, Cade Metz was reaching out to MIRI and some other folks in the x-risk space back in January 2020, and I went to read some of his articles and came to the conclusion in January that he's one of the least competent journalists -- like, most likely to misunderstand his beat and emit obvious howlers -- that I'd ever encountered. I told folks as much at the time, and advised against talking to him just on the basis that a lot of his journalism is comically bad and you'll risk looking foolish if you tap him.

This was six months before Metz caused SSC to shu... (read more)

8Alex_Altair
I really appreciate this comment! And yeah, that's why I said only "Note that...", and not something like "don't trust this guy". I think the content of the article is probably true, and maybe it's Metz who wrote it just because AI is his beat. But I do also hold tiny models that say "maybe he dislikes us" and also something about the "questionable understanding" etc that habryka mentions below. AFAICT I'm not internally seething or anything, I just have a yellow-flag attached to this name.

I think it is useful for someone to tap me on the shoulder and say "Hey, this information you are consuming, its from <this source that you don't entirely trust and have a complex causal model of>".

Enforcing social norms to prevent scapegoating also destroys information that is valuable for accurate credit assignment and causally modelling reality. I haven't yet found a third alternative, and until then, I'd recommend people both encourage and help people in their community to not scapegoat or lose their minds in 'tribal instincts' (as you put it), w... (read more)

robo91

I appreciate that you are not speaking loudly if you don't yet have anything loud to say.

robo330

Is that your family's net worth is $100 and you gave up $85?  Or your family's net worth is $15 and you gave up $85?

Either way, hats off!

The latter. Yeah idk whether the sacrifice was worth it but thanks for the support. Basically I wanted to retain my ability to criticize the company in the future. I'm not sure what I'd want to say yet though & I'm a bit scared of media attention. 

Reply171151
robo30

How close would this rank a program p with a universal Turing machine simulating p?  My sense is not very close because the "same" computation steps on each program don't align.

My "most naïve formula for logical correlation" would be something like put a probability distribution on binary string inputs, treat  and  as random variables , and compute their mutual information.

2niplav
The strongest logical correlation is -0.5, the lower the better. For p and sim(p), the logical correlation would be −ε, assuming that p and sim(p) have the same output. This is a pretty strong logical correlation. This is because equal output guarantees a logical correlation of at most 0, and one can then improve the logical correlation by also having similar traces. If the outputs have string distance 1, then the smallest logical correlation can be only 0.5.
Load More