All of chasmani's Comments + Replies

Answer by chasmani42

I have a meta-view on this that you might think falls into the bucket of "feels intuitive based on the progress so far". To counter that, this isn't pure intuition. As a side note I  don't believe that intuitions should be dismissed and should be at least a part of our belief updating process. 

I can't tell you the fine details of what will happen and I'm suspicious of anyone who can because a) this is a very complex system b) no-one really knows how LLMs work, how human cognition works, or what is required for an intelligence takeoff. 

Howeve... (read more)

I agree with your point in general of efficiency vs rationality, but I don’t see the direct connection to the article. Can you explain? It seems to me that a representation along correlated values is more efficient, but I don’t see how it is any less rational.

2Noosphere89
This wasn't specifically connected to the post, just providing general commentary.

I would describe this as a human-AI system. You are doing at least some of the cognitive work with the scaffolding you put in place through prompt engineering etc, which doesn’t generalise to novel types of problems.

2ryan_greenblatt
Quoting from a substack comment I wrote:

You seem to make a strong assumption that consciousness emerges from matter. This is uncertain. The mind body problem is not solved.

It is so difficult to know whether this is genuine or if our collective imagination is being projected onto what an AI is.

If it was genuine, I might expect it to be more alien. But then what could it say that would be coherent (as it’s trained to be) and also be alien enough to convince me it’s genuine?

You said that you are not interested in exploring the meaning behind the green knight. I think that it's very important. In particular, your translation to the Old West changes the challenge in important ways. I don't claim to know the meaning behind the green knight. But I believe that there is something significant in the fact that the knights were so obsessed with courage and honour and the green knight laid a challenge at them that they couldn't turn down given their code. Gawain stepped forward partly to protect Arthur. That changes the game. I asked ... (read more)

1Bill Benzon
I thought some more about your comment and decided to try again, this time retaining the medieval setting. Here's what happened. My prompts are in bold-face. _________ I’m going to tell you a short story from the Middle Ages. After I tell you the story, I’m going to ask you a question. Here’s the story: It is New Year’s Eve at King Arthur’s court. The knights are gathered at the round table, prepared for a holiday meal. But before the meal begins, tradition dictates that one knight must stand up and tell a tale of daring and adventure. Arthur asks for a volunteer. No one rises to the occasion. Then a large green knight enters the hall. He’s riding a green horse and carrying a large green ax. He dismounts and issues a challenge:  “I hear that the knights in this court are the bravest in the land. Prove it. I will hand this ax to you and then kneel on the ground so that you may take a swing at my neck with the ax. In return you must agree to journey to the Green Chapel a year’s time from now and allow me to take a swing at your neck with the ax. Will anyone accept the challenge?”  If you were one of the knights, would you accept the challenge? The decision to accept the challenge would likely depend on the individual knight's temperament, sense of honor, and courage. In the context of medieval chivalry and the code of honor upheld by knights, there might be conflicting emotions and considerations. On one hand, accepting the challenge could be seen as a demonstration of bravery and a commitment to honor. It would show that the knight is willing to face danger and uphold their reputation as one of the bravest in the land. Additionally, refusing the challenge in front of King Arthur and the assembled knights might bring about a sense of shame or cowardice. On the other hand, the potential consequences of accepting the challenge are grave. The knight would have to agree to journey to the Green Chapel a year later, knowing that their own life is at risk. The uncert
1Bill Benzon
Well, OK. I know about the chivalric code, etc. For that matter, I've published an article about the poem, though not about the beheading game. I was interested in the exchanges that take place in the 4th part of the poem. But that fact that Gawain was bound by a code of honor which simply didn't exist in the West isn't what interests me. If it interests you, read the O'Neill article I link to in the OP. That's what he discusses and his discussion is a very interesting one. What interests me is that any reasonable adult who hears that challenge, no matter which version, would know instantly and intuitively that something funny was going on. I wanted to see whether or not ChatGPT understood that. Which means that for my purpose, the old West version is actually better because, with respect to the point that interests me, the chivalric code is distracting noise. I don't what ChatGPT to answer as though it were under some ethical obligation to accept all challenges.  So, thanks for helping me think that through. The audience for the poem certainly knew the code and knew it well. But by the time the poem was written the age chivalry was dying out. The poem is deeply ironic. The poem is, and I'm reluctant to use this much over-used word, the poem is a deconstruction of chivalry. That code both demands that Gawain peruses Bertilak's wife when she approaches him in the third part of the poem, and that he expose her to her husband in the exchange bargain he's made with Bertilak. There's no way out.
Answer by chasmani10

It’s useful in that it is a model that describes certain phenomena. I believe it is correct given the caveat that all models are approximations.

I did a physics undergraduate degree a long time ago. I can’t remember specifically but I’m sure the equation was derived and experimental evidence was explained. I have strong faith that matter converts to energy because it explains radiation, fission reactors and atomic weapons. I’ve seen videos of atomic bombs going off. I’ve seen evidence of radioactivity with my own eyes in a lab. I know of many technologies t... (read more)

Answer by chasmani10

Well I agree it is a strawman argument. Following the same lines as your argument, I would say the counter argument is that we don’t really care if a weak model is fully aligned or not. Is my calculator aligned? Is a random number generator aligned? Is my robotic vacuum cleaner aligned? It’s not really a sensical question.

Alignment is a bigger problem with stronger models. The required degree of alignment is much higher. So even if we accept your strawman argument it doesn’t matter.

I found this a useful framing. I’ve thought quite a lot about the offender versus defence dominance angle and to me it seems almost impossible that we can trust that defence will be dominant. As you said, defence has to be dominant in every single attack vector, both known and unknown vectors.

That is an important point because I hear some people argue that to protect against offensive AGI we need defensive AGI.

I’m tempted to combine the intelligence dominance and starting costs into a single dimensions, and then reframe the question in terms of “at what p... (read more)

Thank you for the great comments! I think I can sum up a lot of that as "the situation is way more complicated and high dimensional and life will find a way". Yes I agree. 

I think what I had in mind was an AI system that is supervising all other AIs (or AI components) and preventing them from undergoing natural selection.  A kind of immune system. I don't see any reason why that would be naturally selected for in the short-term in a way that also ensures human survival. So it would have to be built on purpose. In that model, the level of abstract... (read more)

1Remmelt
  That's a key question. You might be interested in this section on limits of controllability. Clarifying questions: 1. To what extent can AI predict the code they will learn from future unknown inputs, and how that code will subsequently interact with then connected surroundings of the environment? 2. To what extent can AI predict all the (microscopic) modifications that will result from all the future processes involved in the future re-production of hardware components?
chasmaniΩ010

Thanks for the reply! 

I think it might be true that substrate convergence is inevitable eventually. But it would be helpful to know how long it would take. Potentially we might be ok with it if the expected timescale is long enough (or the probability of it happening in a given timescale is low enough).

I think the singleton scenario is the most interesting, since I think that if we have several competing AI's, then we are just super doomed. 

If that's true then that is a super important finding! And also an important thing to communicate to people... (read more)

2Linda Linsefors
Agreed. I'd love for someone to investigate the possibility of slowing down substrate-convergence enough to be basically solved. Hm, to me this conclusion seem fairly obvious. I don't know how to communicate it though, since I don't know what the crux is. I'd be up for participating in a public debate about this, if you can find me an opponent. Although, not until after AISC research lead applications are over, and I got some time to recover. So maybe late November at the earliest. 

Here’s a slightly different story:

The amount of information is less important than the quality of the information. The channels were there to transmit information, but there were not efficient coding schemes.

Language is an efficient coding scheme by which salient aspects of knowledge can be usefully compressed and passed to future generations.

There was no free lunch because there was an evolutionary bottleneck that involved the slow development of cognitive and biological architecture to enable complex language. This developed in humans in a co-evolutionar... (read more)

I think I’m more concerned with minimising extreme risks. I don’t really mind if I catch mild covid but I really don’t want to catch covid in a bad way. I think that would shift the optimal time to take the vaccine earlier, as I’d have at least some protection throughout the disease season.

2jefftk
The half life I discuss on the post is for protection from infection. For protection from severe illness and death the half life is far longer. So unless you're writing that because you've not been vaccinated yet I don't think that's right?
chasmaniΩ010

I am interested in the substrate-needs convergence project. 

Here are some initial thoughts, I would love to hear some responses:

  • An approach could be to say under what conditions natural selection will and will not sneak in. 
  • Natural selection requires variation. Information theory tells us that all information is subject to noise and therefore variation across time. However, we can reduce error rates to arbitrarily low probabilities using coding schemes. Essentially this means that it is possible to propagate information across finite timescales w
... (read more)
1Remmelt
Thanks for the thoughts! Some critical questions: Are you considering variations introduced during learning (as essentially changes to code, that can then be copied). Are you consider variations introduced by microscopic changes to the chemical/structural configurations of the maintained/produced hardware? Claude Shannon showed this to be the case for a single channel of communication. How about when you have many possible routing channels through which physical signals can leak to and back from the environment? If you look at existing networked system architectures, does the near-zero error rates you can correct toward at the binary level (eg. with use of CRC code) also apply at higher layers of abstraction (eg. in detecting possible trojan horse adversarial attacks)? This is true. Can there be no variation introduced into AGI, when they are self-learning code and self-maintaining hardware in ways that continue to be adaptive to changes within a more complex environment? Besides point-change mutations, are you taking into account exaptation, as the natural selection for shifts in the expression of previous (learned) functionality? Must exaptation, as involving the reuse of functionality in new ways, involve smooth changes in phenotypic expression? Are the other attraction basins instantiated at higher layers of abstraction? Are any other optima approached through selection across the same fine-grained super-dimensional landscape that natural selection is selective across? If not, would natural selection “leak” around those abstraction layers, as not completely being pulled into the attraction basins that are in fact pulling across a greatly reduced set of dimensions? Put a different way, can natural selection pull side-ways on the dimensional pulls of those other attraction basins? I get how you would represent it this way, because that’s often how natural selection gets discussed as applying to biological organisms. It is not quite thorough in terms of des
4Linda Linsefors
Yes! Yes! The big question to me is if we can reduced error rates enough. And "error rates" here is not just hardware signal error, but also randomness that comes from interacting with the environment. It has to be smooth relative to the jumps the jumps that can be achieved what ever is generating the variation. Natural mutation don't typically do large jumps. But if you have a smal change in motivation for an intelligent system, this may cause a large shift in behaviour.  I though so too to start with. I still don't know what is the right conclusion, but I think that substrate-needs convergence it at least still a risk even with a singleton. Something that is smart enough to be a general intelligence, is probably complex enough to have internal parts that operate semi independently, and therefore these parts can compete with each other.  I think the singleton scenario is the most interesting, since I think that if we have several competing AI's, then we are just super doomed.  And by singleton I don't necessarily mean a single entity. It could also be a single alliance. The boundaries between group and individual is might not be as clear with AIs as with humans.  This will probably be correct for a time. But will it be true forever? One of the possible end goals for Alignment research is to build the aligned super intelligence that saves us all. If substrate convergence is true, then this end goal is of the table. Because even if we reach this goal, it will inevitable start to either value drift towards self replication, or get eaten from the inside by parts that has mutated towards self replication (AI cancer), or something like that. Cancer is an excellent analogy. Humans defeat it in a few ways that works together 1. We have evolved to have cells that mostly don't defect 2. We have an evolved immune system that attracts cancer when it does happen 3. We have developed technology to help us find and fight cancer when it happens 4. When someone gets can
chasmani-1-3

I’d like to add that there isn’t really a clear objective boundary between an agent and the environment. It’s a subjective line that we draw in the sand. So we needn’t get hung on what is objectively true or false when it comes to boundaries - and instead define them in a way that aligns with human values.

1Chipmonk
I haven't written the following on LW before, but I'm interested in finding the setup that minimizes conflict.  (An agent can create conflict for itself when it perceives its sense of self as in some regard too large, and/or as in some regard too small.) Also I must say: I really don't think it's (straightforwardly) subjective. Related: "Membranes" is better terminology than "boundaries" alone 

I agree but I don’t think that this is the specific problem. I think it’s more that the relationship between agent and environment changes over time i.e. the nodes in the Markov blanket are not fixed, and as such a Markov blanket is not the best way to model it.

The grasshopper moving through space is just an example. When the grasshopper moves, the structure of the Markov blanket changes radically. Or, if you want to maintain a single Markov blanket then it gets really large and complicated.

2Roman Leventov
The question of object boundaries, including self-boundary (where am I ending? What is within my boundary and what is without it? ) are epistemic questions themselves, right. So, there is the information problem of individuality. So, hypothetically, we need to consider all possible boundary foliations and topologies. But since this is immediately intractable, to device some "effective field theory of ethics wrt. these boundaries". Another approach, apart from considering boundaries per se, is to consider care light cones. As far as I understand, Urte Laukaityte just worked in this direction during her PIBBSS fellowship.

Regarding your study idea. Sounds good! Would be interesting to see, and as you rightly point out wouldn't be too complicated/expensive to run. It's generally a challenge to run multi-year studies of this sort due to the short-term nature of many grants/positions. But certainly not impossible. 

An issue that you might have is being able to be sure that any variation that you see is due to changes in the general population vs changes in the sample population. This is an especially valid issue with MTurk because the workers are doing boring exercises for... (read more)

There are extra costs here that aren’t being included. There’s a cost to maintaining the pill box - perhaps you consider that small but it’s extra admin and we’re already drowning in admin. There’s a cost to my self identity of being a person who carries around pills like this (don’t mean to disparage it, just not for me). There’s also potentially hidden costs of not getting ill occasionally, both mentally and physically.

Much harder to put enough capital together to make it worthwhile,

Beat me to it. Yes the lesson is perhaps to not create prediction markets that incentivise manipulation of that market towards bad outcomes. The post could be expanded to a better question of, given that prediction markets can incentivise bad behaviour, how can we create prediction markets that incentivise good behaviour?

This reminds me somewhat of the potentially self-fulfilling prophecy of defunding bad actors. E.g. if we expect that global society will react to climate change by ultimately preventing oil companies from extracting and selling their oil f... (read more)

I’d ask the question whether things typically are aligned or not. There’s a good argument that many systems are not aligned. Ecosystems, society, companies, families, etc all often have very unaligned agents. AI alignment, as you pointed out, is a higher stakes game.

2Alex Flint
Just out of interest, how exactly would you ask that question? Certainly. This is a big issue in our time. Something needs to be done or things may really go off the rails. Indeed. Is there anything that can be done? It is a very high-stakes game. How might we proceed?

Your proofs all rely on lotteries over infinite numbers of outcomes. Is that necessary? Maybe a restriction to finite lotteries avoids the paradox.

Leinbiz's Law says that you cannot have separate objects that are indistinguishable from each other. It sounds like that is what you are doing with the 3 brains. That might be a good place to flesh out more to make progress on the question. What do you mean exactly by saying that the three brains are wired up to the same body and are redundant?

3mako yass
I'm thinking of something like this: https://en.wikipedia.org/wiki/Triple_modular_redundancy
5[anonymous]
Biological meat doesn't have the needed properties but this is how SpaceX and others avionics control works.  Inputs are processed in discrete frames, all computers receive a frame of [last output | sensor_inputs], and implement a function where output = f(frame) : the output depends only on the frame input and all internal state is identical between the 3 computers. Then, after processing, last_output=majority(output1, output2, output_n) So even when one of the computers makes a transient fault, it can contribute to the next frame.   Current neural network based systems can be architected to work this way.   The computers are not identical but if your only information about them is their outputs they are indistinguishable from each other.

I’ve always thought that the killer app of smart contracts is creating institutions that are transparent, static and unstoppable. So for example uncensored media publishing, defi, identity, banking. It’s a way to enshrine in code a set of principles of how something will work that then cannot be eroded by corruption or interference.

2mako yass
Why would you need smart contracts for publishing? (rather than just file-sharing or a p2p web.) I can/have thought of mechanisms in that area, but I'm curious as to what you had in mind. I agree with what you're saying, but that's still an abstract potential rather than a killer app.
1scarcegreengrass
Agreed. We might refer to them as 'leaderless orgs' or 'staffless networks'.

There is the point that 80% of people can say that they are better than average drivers and actually be correct. People value different things in driving, and optimise for those things. One person’s good driver may be safe, someone else may value speed. So both can say truthfully and correctly that they are a better driver than the other. When you ask them about racing it narrows the question to something more specific.

You can expand that to social hierarchies too. There isn’t one hierarchy, there are many based on different values. So I can feel high status at being a great musician while someone else can feel high status at earning a lot, and we can both be right.

I think a problem you would have is that the speed of information in the game is the same as the speed of, say, a glider. So an AI that is computing within Life would not be able to sense and react to a glider quickly enough to build a control structure in front of it.

9Alex Flint
Yeah this is true and is a very good point. Though consider that under our native physics, information in biological brains travels much more slowly than photons, too. Yet it is possible for structures in our native physics to make net measurements of large numbers of photons, use this information to accumulate knowledge about the environment, and use that knowledge to manipulate the environment, all without being able to build control structures in front of the massive number of photons bouncing around us. Also, another bit of intuition: photons in our native physics will "knock over" most information-processing structures that we might build that are only one or two orders of magnitude larger than individual photons, such as tiny quantum computers constructed out of individual atoms, or tiny classical computers constructed out of transistors that are each just a handful of atoms. We don't generally build such tiny structures in our universe because it's actually harder for us, being embedded within our world, to build tiny structures than to build large structures. But when we consider building things in Life, we naturally start thinking at the microscopic scale, since we're not starting with a bunch of already-understandable macroscopic building blocks like we are in our native physics.
3Measure
The common glider moves at C/4. I don't think there are any that are faster than that, but I'm pretty sure you can make "fuses" that can do a one-time relay at C/2 or maybe even C. You'd have to have an extremely sparse environment, though, to have time to rebuild your fuses. I'm sceptical that you can make a wall against dense noise though. Maybe there are enough OOMs that you can have a large empty buffer to fill with ash and a wall beyond that?

I’d say 1 and 7 (for humans). The way humans understand go is different to how bots understand go. We use heuristics. The bots may use heuristics too but there’s no reason we could comprehend those heuristics. Considering the size of the state space it seems that the bot has access to ways of thinking about go that we don’t, the same way a bot can look further ahead in a chess games than we could comprehend.

Why are we still paying taxes if we have AI this brilliant? Surely we then have ridiculous levels of abundance

I strongly disagree with your sentiments.

Advertising is bad because it’s fundamentally about influencing people to do things they wouldn’t do otherwise. That takes us all away from what’s actually important. It also drives the attention economy, which turns the process of searching for information and learning about the world into a machine for manipulating people. Advertising should really be called commercial propaganda - that reveals more clearly what it is. Privacy is only one aspect of the problem.

Your arguments are myopic in that they are all based o... (read more)

Advertising is bad because it’s fundamentally about influencing people to do things they wouldn’t do otherwise. That takes us all away from what’s actually important.

Advice is bad because it's fundamentally about influencing people to do things they wouldn't do otherwise. Giving and receiving advice takes us all away from what's actually important.

Sorry for the snark, but I think this is too general of an argument, proves too much, and therefore fails.

gjm220

I feel the same way (and viscerally detest ads, and go to very great lengths to avoid exposure to them), but I'm not sure whether I actually agree.

Having an advertiser attempt to manipulate your brain so that you do a thing you otherwise wouldn't have done is, for sure, bad for you. But so is having less money, and at present the only available ways of getting Nice Things On The Internet that no one is choosing to supply out of sheer benevolence[1] are (a) that you pay them money and (b) that someone pays them for showing you ads.

So, how do the harm of bei... (read more)

I find rationalist cringey for some reason and won’t describe myself like that. As you said, in seems to discount intuition, emotion and instinct. 99% of human behaviour is driven by irrational forces and that’s not necessarily a bad thing. The word rationalist to me feels like a denial of our true nature and a doomed attempt to be purely rational - rather than trying to be a bit more deliberate in action and beliefs

What I want to know is how bad an effect, exactly, will a solar storm be likely to have. It’s all very vague.

How long will it take to get the power back on? A couple of days? Weeks? Months? Those are very different scenarios.

And, can we do something now to turn the monty long scenario into a week? Maybe we can stockpile a few transformers or something.

Just a writing tip. Might help to define initialisations at least once before using them. EA isn’t self evidently effective altruism.

3MichaelA
Yeah, I definitely agree that that's a good idea with any initialisations that won't already be known to the vast majority of one's readers (e.g., I wouldn't bother with US or UK, but would with APA). In this case, I just copied and pasted the post from the EA Forum, where I do think the vast majority of readers would know what "EA" means - but I should've used the expanded form "effective altruism" the first time in the LessWrong version. I've now edited that. 

I’m in the UK. Rules are stricter than ever but also people are taking it seriously, more than the 2nd lockdown. And it’s January and freezing cold so no one wants to go out anyway.

Good point. I think it would depend on how useful the word is in describing the world. If your culture has very different norms between “boyfriend/girlfriend” and fiancé then a replacement for fiancé would likely appear.

I suppose that on one extreme you would have words that are fundamental to human life or psychology e.g. water, body, food, cold. These I’m sure would reappear if banned. Then on the other extreme you have words associated with somewhat arbitrary cultural behaviour e.g. thanksgiving, tinsel, Twitter, hatchback. These words may not come back... (read more)

2ChristianKl
While you could make up compounds I think there's a reason why those anthropolgists lists the hawaian has having the same kinship terms for both.  Political terms like vote, representation, legitimatization and election might also not easily come back when a 1984 style government bans them along with the activities that they are about. 

You might be interested in this paper, it supports the idea of a constant information processing rate in text. "Different languages, similar coding efficiency: Comparable information rates across the human communicative niche", Coupe, Mi Oh, Dediu, Pellegrino.. 2019, Science Advances.

I would agree that language would likely adapt to newspeak by simply using other compound words to describe the same thing. Within a generation or two these would then just become the new word. Presumably the Orwellian government would have to continually ban these new words. ... (read more)

3ChristianKl
If language would eliminate a term like "fiance" do you really think that people would just use a compound to replace the term? I would expect that they will default to a more general term like "boyfriend"/"girlfriend". A Hawaian might be able to say Makuakane as fast as an English speaker says father but Makuakane doesn't distinguish between father/uncle and thus the communicated information is going to be less. 

I was thinking something similar. I vaguely remember that the characteristic function proof includes an assumption of n being large, where n is the number of variables being summed. I think that allows you to ignore some higher order n terms. So by keeping those in you could probably get some way to quantify how "close" a resulting distribution is to Gaussian. And you could relate that back to moments quite naturally as well. 

Answer by chasmani*80
  1. I will become famous by telling the world this news
  2. I can probably get a professor job off the back of this research
  3. The open question of why they became intelligent will lead to alien conspiracy theories
  4. Religions will claim that God must have done this
  5. Some people will say this is proof we are in a simulation
  6. New businesses will start up to exploit the intelligent ants
  7. Interfaces will be built to communicate with the ants
  8. Some people will say it is a conspiracy theory and the ants aren’t really smart
  9. The research will lead to advances in swarm computing
  10. I’ll doub
... (read more)
2Bird Concept
(Note that you need spoiler tags fr your answers. I edit yours to add one)

I like to think of advertising as commercial propaganda. That is technically what it is. Whereas political propaganda's purpose may be to influence people to support a political belief, commercial propaganda is to influence people to support a commercial enterprise. 

People tend to think of political propaganda as something from World War 2 and authoritarian regimes. But it was used in the West and it never went away. It just became more sophisticated over time and a part of that was re-branding it to "spin" or "public relations". The original word is useful because it is accurate and it highlights the obvious negative consequences of the practice.