All of Lukas_Gloor's Comments + Replies

I thought about this and I'm not sure Musk's changes in "unhingedness" require more explanation than "power and fame have the potential to corrupt and distort your reasoning, making you more overconfident." The result looks a bit like hypomania, but I've seen this before with people who got fame and power injections. While Musk was already super accomplished (for justified reasons nonetheless) before taking over Twitter and jumping into politics, being the Twitter owner (so he can activate algorithmic godmode and get even more attention) probably boosted b... (read more)

It feels vaguely reasonable to me to have a belief as low as 15% on "Superalignment is Real Hard in a way that requires like a 10-30 year pause." And, at 15%, it still feels pretty crazy to be oriented around racing the way Anthropic is. 

Yeah, I think the only way I maybe find the belief combination "15% that alignment is Real Hard" and "racing makes sense at this moment" compelling is if someone thinks that pausing now would be too late and inefficient anyway. (Even then, it's worth considering the risks of "What if the US aided by AIs during takeoff... (read more)

2Noosphere89
My own take is I do endorse a version of the "pausing now is too late objection", more specifically I think that for most purposes, we should assume pauses are too late to be effective when thinking about technical alignment, and a big portion of the reason is that I don't think we will be able to convince many people that AI is powerful enough to need governance without them first hand seeing massive job losses, and at that point we are well past the point of no return for when we could control AI as a species. In particular, I think Eliezer is probably vindicated/made a correct prediction around how people would react to AI in there's no fire alarm for AGI (more accurately, the fire alarm will go off way too late to serve as a fire alarm.) More here: https://www.lesswrong.com/posts/BEtzRE2M5m9YEAQpX/there-s-no-fire-alarm-for-artificial-general-intelligence

The DSM-5 may draw a bright line between them (mainly for insurance reimbursement and treatment protocol purposes), but neurochemically, the transition is gradual.

That sounded mildly surprising to me (though in hindsight I'm not sure why it did) so I checked with Claude 3.7, and it said something similar in reply to me trying to ask a not-too-leading question. (Though it didn't talk about neurochemistry -- just that behaviorally the transition or distinction can often be gradual.) 

In my comments thus far, I've been almost exclusively focused on preventing severe abuse and too much isolation.

Something else I'm unsure about, but not necessarily a hill I want to die on given that government resources aren't unlimited, is the question of whether kids should have a right to "something at least similarly good as voluntary public school education." I'm not sure if this can be done cost-effectively, but if the state had a lot money that they're not otherwise using in better ways, then I think it would be pretty good to have standardized tes... (read more)

2habryka
This seems like it would punish variance a lot, and de-facto therefore be a huge tax on homeschooling. Some public schools are extremely bad, if no home schoolers are allowed to be as bad as the worst public schools, the costs of homeschooling increase a lot, constituting effectively a tax on homeschooling.  Maybe you mean "a right to an education at least as good as the worst public school education", but my guess is the worst public school education is so bad that these would already be covered by almost any reasonable approach to human rights (like, my guess is it already involves continuous ongoing threats of violence, being lied to, frequent physical violence, etc.).

More concretely, do you think parents should have to pass a criminal background check (assuming this is what you meant by "background check") in order to homeschool, even if they retain custody of their children otherwise?

I don't really understand why you're asking me about this more intrusive and less-obviously-cost-effective intervention, when one of the examples I spelled out above was a lower-effort, less intrusive, less controversial version of this sort of proposal.

I wrote above: 

Like, even if yearly check-ins for everyone turn out to be too exp

... (read more)
2Lukas_Gloor
In my comments thus far, I've been almost exclusively focused on preventing severe abuse and too much isolation. Something else I'm unsure about, but not necessarily a hill I want to die on given that government resources aren't unlimited, is the question of whether kids should have a right to "something at least similarly good as voluntary public school education." I'm not sure if this can be done cost-effectively, but if the state had a lot money that they're not otherwise using in better ways, then I think it would be pretty good to have standardized tests for homeschooled kids every now and then, maybe every two to three years. One of them could be an IQ test, the other an abilities test. If the kid has an IQ that suggests that they could learn things well but they seem super behind other children of their age, and you ask them if they want to learn and they say yes with enthusiasm, then that's suggestive of the parents doing an inadequate job, in which case you could put them on homeschooling probation and/or force them to allow their child to go to public school? 

Thanks for elaborating, that's helpful. 

If we were under a different education regime

Something like what you describe would maybe even be my ideal too (I'm hedging because I don't have super informed views on this). But I don't understand how my position of "let's make sure we don't miss out on low-cost, low-downside ways of safeguarding children (who btw are people too and didn't consent to be born, especially not in cases where their parents lack empathy or treat children as not people) from severe abuse" is committed to having to answer this hypote... (read more)

4Arjun Panickssery
My point doesn't have anything to do with the relevant merits of public vs homeschool. My point is that your listed interventions aren't reasonable because they involve too much government intrusion into a parent's freedom to educate his children how he wants, for reasons based on dubious and marginal "safeguarding" grounds. My example of a different education regime was to consider a different status quo that makes the intrusiveness more clear.  This might be the crux: I'm saying that this position isn't based on any reasonable principle of government intrusion on people's lives. The government shouldn't intrude on basic parenting rights with a ton of surveillance just to see whether it can make what it thinks is a marginal improvement. More concretely, do you think parents should have to pass a criminal background check (assuming this is what you meant by "background check") in order to homeschool, even if they retain custody of their children otherwise? ---------------------------------------- Re-reading your previous reply, I noticed this: No states use this policy and it doesn't make sense that parents should have to submit to yearly inspections of their parenting practices. It only makes sense from the point of view where homeschooling is highly deviant and basically impermissible without special dispensation, and where the government has the authority to decide in very specific terms what composes children's educations.

Seriously, -5/-11? 

I went through my post line by line and I don't get what people are allergic to.

I'm not taking sides. I flagged that some of the criticisms of homeschooling appear reasonable and important to me. I'm pretty sure I'm right about this, but somehow people want me to say less of this sort of thing, because what? Because public schools are "hell"? How is that different from people who consider the other political party so bad that you cannot say one nuanced thing about them -- isn't that looked down upon on this site?

Also, speaking of "h... (read more)

4habryka
I strong-upvoted and strong-disagree voted, since I also agree the current voting distribution didn't make much sense.  I do think you are doing something in your comment that feels pretty off. For example you link to aphyer's comment as a "fully general counterargument that clearly prove[s] way too much", but I don't buy it, I think it's a pretty reasonable argument. The prior should be towards liberty, and if the higher-liberty option is also safer, then I don't see any reason to mess with it for now.  Like, it seems fine to improve things, but I do think state involvement in education has been really very terrifying and I sense a continuous missing mood throughout your comments of not understanding how costly marginal regulation can be. To be clear, I think your comment is fine and doesn't deserve downvoting, and disagree-voting feels like the appropriate dimension.

I strong-disagreed since I don't think any of your listed criticisms are reasonable. The implied premise is that homeschooling is deviant in a way that justifies a lot of government scrutiny, when really parents have a natural right to educate their children the way they want (with government intervention being reasonable in extreme cases that pass a high bar of scrutiny).

In particular, I think that outside of an existing norm where most students go to public school, the things you listed would be obviously unjust. Do you think that parents who fail a crim... (read more)

4James Camacho
Yeah, I don't see why either. LessWrong allegedly has a utilitarian culture, and simply from the utilitarian "minimize abuse" perspective, you're spot on. Even if home-schooling has similar or mildly lower rates of abuse, the weight of that abuse is higher.

I got the impression that using only an external memory like in the movie Memento (and otherwise immediately forgetting everything that wasn't explicitly written down) was the biggest hurdle to faster progress. I think it does kind of okay considering that huge limitation. Visually, it would also benefit from learning the difference between what is or isn't a gate/door, though. 

It depends on efficiency of the interventions you'd come up with (some may not be much of a "burden" at all) and on the elasticity with which parents who intend to homeschool are turned away by "burdens". You make a good point but what you say is not generally true -- it totally depends on the specifics of the situation. (Besides, didn't the cited study say that both rates of abuse were roughly equal? I don't think anyone suggested that public schooled kids have [edit: drastically] higher abuse rates than home schooled ones? Was it 37% vs 36%?)

I feel like it's worth pointing out the ways homeschooling can go badly wrong. Whether or not there's a correlation between homeschooling and abuse, it's obvious that homeschooling can cover up particulary bad instances of abuse (even if it's not the only way to do that). So, I feel like a position of "homeschooling has the potential to go very bad; we should probably have good monitoring to prevent that; are we sure we're doing that? Can we check?" seems sensible

The article you call a "hit piece" makes some pretty sensible points. The title isn't s... (read more)

8Lukas_Gloor
Seriously, -5/-11?  I went through my post line by line and I don't get what people are allergic to. I'm not taking sides. I flagged that some of the criticisms of homeschooling appear reasonable and important to me. I'm pretty sure I'm right about this, but somehow people want me to say less of this sort of thing, because what? Because public schools are "hell"? How is that different from people who consider the other political party so bad that you cannot say one nuanced thing about them -- isn't that looked down upon on this site? Also, speaking of "hell," I want to point out that not all types of abuse are equal and that the most extreme cases of childhood suffering probably happen disproportionally in the most isolated of homes. How can it not be an ideal to aim for that all children have contact with some qualified person outside their household who can check if they're not being badly abused? Admittedly, it's not understood to be a public school teacher's primary role to notice when something with a child is seriously wrong, but it's a role that they often end up filling (and I wouldn't be surprised if they even get trained in this in many areas). You don't necessarily need public schools for this sort of checking in that serves as a safeguard against some of the most severe kinds of prolonged abuse, but if you just replace public schooling with homeschooling, that role falls away. So, what you could do to get some of the monitoring back: have homeschooling with (e.g.) yearly check-ins with the affected children from a social worker. I don't know the details, but my guess is that some states have this and others don't. (Like the "hit piece" claims, regulation differs from state to state and some are poorly regulated.) I'm not saying I know for sure whether yearly check-ins would be cost-effective compared to other things the state puts money in, but it totally might be, and I doubt that the people who are trying to discourage this discussion (with downvot
aphyer139

If there's less abuse happening in homeschooling than in regular schooling, a policy of "let's impose burdens on homeschooling to crack down on abuse in homeschooling" without a similar crackdown on abuse in non-home-schooling does not decrease abuse.

You can see something similar with self-driving cars.  It is bad if a self-driving car crashes.  It would be good to do things that reduce that.  But if you get to a point where self-driving cars are safer than regular driving, and you continue to crack down on self-driving cars but not on regular driving, this is not good for safety overall.

Andrej Karpathy recently made a video on which model to use under what circumstances. A lot of it probably won't be new to people who read these AI overviews here regularly, but I learned things from it and it's something I'm planning to send to people who are new to working with LLMs.

3Lee.aao
I can confirm that this is a pretty much the best introduction to take you from 0 to about 80% in using AI.  It is intended for general users, don't expect technical information on how to use APIs or build apps.

I want to flesh out one particular rushed unreasonable developer scenario that I’ve been thinking about lately: there’s ten people inside the AI company who are really concerned about catastrophic risk from misalignment. The AI company as a whole pays lip service to AI risk broadly construed and talks occasionally about risk from AGI, but they don’t take misalignment risk in particular (perhaps especially risk from schemers) very seriously. 

[...]

What should these people try to do? The possibilities are basically the same as the possibilities for what

... (read more)

Great reply!

On episodic memory:
I've been watching Claude play Pokemon recently and I got the impression of, "Claude is overqualified but suffering from the Memento-like memory limitations. Probably the agent scaffold also has some easy room for improvements (though it's better than post-it notes and tatooing sentences on your body)."

I don't know much about neuroscience or ML, but how hard can it be to make the AI remember what it did a few minutes ago? Sure, that's not all that's between claude and TAI, but given that Claude is now within the human expert ... (read more)

Lukas_Gloor*1613

I liked most thoughts in this post even though I have quite opposite intutions about timelines. 

I agree timeline intuitions feel largely vibes based, so I struggle to form plausible guesses about the exact source behind our different intuitions. 

I thought this passage was interesting in that respect:

Or, maybe there will be a sudden jump. Maybe learning sequential reasoning is a single trick, and now we can get from 4 to 1000 in two more years.

What you write as an afterthought is what I would have thought immediately. Sure, good chance I'm wrong. ... (read more)

3Rafael Harth
This is probably a crux; I think the brain does have tons of specialized architecture for it, and if I didn't believe that, I probably wouldn't think thought assessment was as difficult. Notably people's intuitions about what is impressive/difficult tend to be inversely correlated with reality. The stereotype is (or at least used to be) that AI will be good at rationality and reasoning but struggle with creativity, humor, and intuition. This stereotype contains information since inverting it makes better-than-chance predictions about what AI has been good at so far, especially LLMs. I think this is not a coincidence but roughly because people use "degree of conscious access" an inverse proxy for intuitive difficulty. The more unconscious something is, the more it feels like we don't know how it works, the more difficult it intuitively seems. But I suspect degree of conscious access positively correlates with difficulty. Yes; I think the "single trick" view might be mostly confirmed or falsified in as little as 2-3 years. (If I introspect I'm pretty confident that I'm not wrong here, the scenario that frightens me is more that sequential reasoning improves non-exponentially but quickly, which I think could still mean doom, even if it takes 15 years. Those feel like short timelines to me.)

[Edit: I wrote my whole reply thinking that you were talking about "organizational politics." Skimming the OP again, I realize you probably meant politics politics. :) Anyway, I guess I'm leaving this up because it also touches on the track record question.]

I thought Eliezer was quite prescient on some of this stuff. For instance, I remember this 2017 dialogue (so less than 2y after OpenAI was founded), which on the surface talks about drones, but if you read the whole post, it's clear that it's meant as an analogy to building AGI: 

AMBER:  T

... (read more)
3samuelshadrach
I agree Yudkowsky is not incompetent at understanding politics. I’m saying he’s not exceptionally good at it. Basically, he’s average. Just like you and me (until proven otherwise).  I didn’t read the entire post, I only skimmed it, but my understanding is this post is Yudkowsky yet again claiming alignment is difficult and that there are some secret insights inside Yudkowsky’s head as to why alignment is hard that can’t be shared in public. I remember reading Yudkowsky versus Christiano debates some years back and they had this same theme of inexplicable insights inside Yudkowkys head. The reasoning about politics in the post you just linked mostly assumes there exist some inexplicable but true insights about alignment difficulty inside Yudkowskys head. 
1flandry39
I really liked your quote and remarks.  So much so, that I made an edited version of them as a new post here:  http://mflb.com/ai_alignment_1/d_250207_insufficient_paranoia_gld.html

If so, why were US electricity stocks down 20-28% (wouldn't we expect them to go up if the US wants to strengthen its domestic AI-related infrastructure) and why did TSMC lose less, percentage-wise, than many other AI-related stocks (wouldn't we expect it to get hit hardest)? 

4Garrett Baker
Yeah, these are mysteries, I don't know why. TSMC I think did get hit pretty hard though. 

In order to submit a question to the benchmark, people had to run it against the listed LLMs; the question would only advance to the next stage once the LLMs used for this testing got it wrong. 

So I think the more rational and cognitively capable a human is, the more likely they'll optimize more strictly and accurately for future reward.

If this is true at all, it's not going to be a very strong effect, meaning you can find very rational and cognitively capable people who do the opposite of this in decision situations that directly pit reward against the things they hold most dearly. (And it may not be true because a lot of personal hedonists tend to "lack sophistication," in the sense that they don't understand that their own feelings of valuing ... (read more)

I like all the considerations you point out, but based on that reasoning alone, you could also argue that a con man who ran a lying scheme for 1 year and stole only like $20,000 should get life in prison -- after all, con men are pathological liars and that phenotype rarely changes all the way. And that seems too harsh?

I'm in two minds about it: On the one hand, I totally see the utilitarian argument of just locking up people who "lack a conscience" forever the first time they get caught for any serious crime. On the other hand, they didn't choose how they... (read more)

7Ben Pace
Good point. I can imagine things like "permanent parole" (note that probation and parole are meaningfully different) or being under house arrest or having constraints on your professional responsibilities or finances or something, being far better than literal incarceration.
4Ben Pace
I agree there are people who do small amounts of damage to society, are caught, and do not reoffend. Then there are other people whose criminal activities will be most of their effect on society, will reliably reoffend, and for whom the incapacitation strongly works out positive in consequentialist terms. My aim would be to have some way of distinguishing between them. The amount of evidence we have about Bankman-Fried's character is quite different than that of most con men, including from childhood and from his personal diary, so I hope we can have more confidence based on that. But a different solution is to not do any psychologizing, and just judge based on reoffending. See this section from the ACX post: I should add that Scott has lots of concerns about doing this in the US, and argues that properly doing this in the US would massively increase the incarcerated population. I didn't quite follow his concerns, but I was not convinced that something like this would be a bad idea on consequentialist grounds, even if the incarcerated population were to massively increase. (Note that I would support improving the quality of prisons to being broadly as nice as outside of prisons.)
7Noosphere89
One of the missing considerations is that crime is done mostly by young people, and the rate of crimes goes down the older you get. A lot of this IMO is that the impulsiveness/risk-taking behavior of crimes decreases a lot with age, but the empirical fact of crime going down with age, especially reoffending is a big reason why locking people up for life is less good than Ben Pace said, because the reoffending rate goes down with age.

Suppose that a researcher's conception of current missing pieces is a mental object M, their timeline estimate is a probability function P, and their forecasting expertise F is a function that maps M to P. In this model, F can be pretty crazy, creating vast differences in P depending how you ask, while M is still solid.

Good point. This would be reasonable if you think someone can be super bad at F and still great at M.

Still, I think estimating "how big is this gap?" and "how long will it take to cross it?" might quite related, so I expect the skills to be correlated or even strongly correlated.

6Kaj_Sotala
I think their relationship depends on whether crossing the gap requires grind or insight. If it's mostly about grind then a good expert will be able to estimate it, but insight tends to unpredictable by nature. Another way of looking at my comment above would be that timelines of less than 5 years would imply the remaining steps mostly requiring grind, and timelines of 20+ years would imply that some amount of insight is needed.
Lukas_Gloor*2816

It surveyed 2,778 AI researchers who had published peer-reviewed research in the prior year in six top AI venues (NeurIPS, ICML, ICLR, AAAI, IJCAI, JMLR); the median time for a 50% chance of AGI was either in 23 or 92 years, depending on how the question was phrased.

Doesn't that discrepancy (how much answers vary between different ways of asking the question) tell you that the median AI researcher who published at these conferences hasn't thought about this question sufficiently and/or sanely?

It seems irresponsible to me to update even just a small bit to ... (read more)

6Kaj_Sotala
That would imply that most professions would be getting automated or having their productivity very significantly increased. My impression from following the news and seeing some studies is that this is happening within copywriting, translation, programming, and illustration. [EDIT: and transcription] Also people are turning to chatbots for some types of therapy, though many people will still intrinsically prefer a human for that and it's not affecting the employment of human therapists yet. With o3, math (and maybe physics) research is starting to be affected, though it mostly hasn't been yet. I might be forgetting some, but the amount of professions left out of that list suggests that there are quite a few skill types that are still untouched. (There are of course a lot of other professions for which there have been moderate productivity boosts, but AFAIK mostly not to the point that it would affect employment.)

Doesn't that discrepancy (how much answers vary between different ways of asking the question) tell you that the median AI researcher who published at these conferences hasn't thought about this question sufficiently and/or sanely?

We know that AI expertise and AI forecasting are separate skills and that we shouldn't expect AI researchers to be skilled at the latter. So even if researchers have thought sufficiently and sanely about the question of "what kinds of capabilities are we still missing that would be required for AGI", they would still be lacking t... (read more)

Well, the update for me would go both ways. 

On one side, as you point out, it would mean that the model's single pass reasoning did not improve much (or at all). 

On the other side, it would also mean that you can get large performance and reliability gains (on specific benchmarks) by just adding simple stuff. This is significant because you can do this much more quickly than the time it takes to train a new base model, and there's probably more to be gained in that direction – similar tricks we can add by hardcoding various "system-2 loops" into ... (read more)

3yo-cuddles
I sense that my quality of communication diminishes past this point, I should get my thoughts together before speaking too confidently I believe you're right we do something similar to the LLM's (loosely, analogously), see https://www.lesswrong.com/posts/i42Dfoh4HtsCAfXxL/babble (I need to learn markdown) My intuition is still LLM pessimistic, I'd be excited to see good practical uses, this seems like tool ai and that makes my existential dread easier to manage!
Lukas_Gloor*4611

When the issue is climate change, a prevalent rationalist take goes something like this:

"Climate change would be a top priority if it weren't for technological progress. However, because technological advances will likely help us to either mitigate the harms from climate change or will create much bigger problems on their own, we probably shouldn't prioritize climate change too much." 

We could say the same thing about these trends of demographic aging that you highlight. So, I'm curious why you're drawn to this topic and where the normative motivation... (read more)

"Climate change would be a top priority if it weren't for technological progress. However, because technological advances will likely help us to either mitigate the harms from climate change or will create much bigger problems on their own, we probably shouldn't prioritize climate change too much."

This attitude deserves a name: technocrastinating.

Technological progress has been happening for a while. At some point, this argument will stop making sense and we must admit that no, this (climate change, fertility, whatever) is not fine, stop technocrastinat... (read more)

The tabletop game sounds really cool!

Interesting takeaways.

The first was exactly the above point, and that at some point, ‘I or we decide to trust the AIs and accept that if they are misaligned everyone is utterly f***ed’ is an even stronger attractor than I realized.

Yeah, when you say it like that... I feel like this is gonna be super hard to avoid!

The second was that depending on what assumptions you make about how many worlds are wins if you don’t actively lose, ‘avoid turning wins into losses’ has to be a priority alongside ‘turn your losses into not l

... (read more)
Lukas_Gloor*4025

I agree that it sounds somewhat premature to write off Larry Page based on attitudes he had a long time ago, when AGI seemed more abstract and far away, and then not seek/try communication with him again later on. If that were Musk's true and only reason for founding OpenAI, then I agree that this was a communication fuckup.

However, my best guess is that this story about Page was interchangeable with a number of alternative plausible criticisms of his competition on building AGI that Musk would likely have come up with in nearby worlds. People like Musk (a... (read more)

Seth Herd2314

I totally agree. And I also think that all involved are quite serious when they say they care about the outcomes for all of humanity. So I think in this case history turned on a knife edge; Musk would've at least not done this much harm had he and Page had clearer thinking and clearer communication, possibly just by a little bit.

But I do agree that there's some motivated reasoning happening there, too. In support of your point that Musk might find an excuse to do what he emotionally wanted to anyway (become humanity's savior and perhaps emperor for eternit... (read more)

I thought the part you quoted was quite concerning, also in the context of what comes afterwards: 

Hiatus: Sam told Greg and Ilya he needs to step away for 10 days to think. Needs to figure out how much he can trust them and how much he wants to work with them. Said he will come back after that and figure out how much time he wants to spend.

Sure, the email by Sutskever and Brockman gave some nonviolent communication vibes and maybe it isn't "the professional thing" to air one's feelings and perceived mistakes like that, but they seemed genuine in what ... (read more)

Some of the points you make don't apply to online poker. But I imagine that the most interesting rationality lessons from poker come from studying other players and exploiting them, rather than memorizing and developing an intuition for the pure game theory of the game. 

  • If you did want to focus on the latter goal, you can play online poker (many players can >12 tables at once) and after every session, run your hand histories through a program (e.g., "GTO Wizard") that will tell you where you made mistakes compared to optimal strategy, and how much
... (read more)
1rossry
Strongly agree. I didn't realize this when I wrote the original post, but I'm now convinced. It has been the most interesting / useful thing that I've learned in the working-out of Cunningham's Law with respect to this post. And so, there's a reason that the curriculum for my and Max's course shifts away from Nash equilibrium as the solution concept to optimizing winnings against an empirical (and non-Nash) field just as soon as we can manage it. For example, Practicum #3 (of 6) is "write a rock-paper-scissors bot that takes advantage of our not-exactly-random players as much as you can" without much further specification.

It seems important to establish whether we are in fact going to be in a race and whether one side isn't already far ahead.

With racing, there's a difference between optimizing the chance of winning vs optimizing the extent to which you beat the other party when you do win. If it's true that China is currently pretty far behind, and if TAI timelines are fairly short so that a lead now is pretty significant, then the best version of "racing" shouldn't be "get to the finish line as fast as possible." Instead, it should be "use your lead to your advantage." So,... (read more)

2niplav
Yep, makes sense. I think if I modify the model to shorten timelines & widen the gap between PRC and US the answer could flip.

I wrote a long post last year saying basically that.

Even if attaining a total and forevermore cessation of suffering is substantially more difficult/attainable by substantially fewer people in one lifetime, I don't think it's unreasonable to think that most people could suffer at least 50 percent less with dedicated mindfulness practice. I'm curious as to what might feed an opposing intuition for you! I'd be quite excited about empirical research that investigates the tractability and scalability of meditation for reducing suffering, in either case.

My sense is that existing mindfulness studies don't show th... (read more)

1jbkjr
If you have any specific studies in mind which show this, I would be interested to see! I have a sense that mindfulness tends to be studied in the context of “increasing well-being” in a general sense and not specifically to “decrease or eliminate suffering.” I would be quite interested in a study which studies meditation’s effects when directly targeting suffering. I really appreciate you raising this point in detail; I think it’s something I haven’t included enough in my own considerations. Having enough free time and energy for meditation practice is indeed a kind of privilege. I’m going to chew on this some more, but one initial thought I’ve had is that the general quality of life needed as a prerequisite to devoting enough time and energy to meditation practice may be lower than one may expect, at least by Western standards. For example, in a place like India, there seems to be a good amount of people in difficult circumstances that nonetheless make time for meditation and spiritual pursuits. However, I agree that in the limit, if all of your waking moments are focused on simply acquiring enough food today, it seems much less reasonable to prescribe meditation as the solution for their suffering.

[...] I am certainly interested to know if anyone is aware of sources that make a careful distinction between suffering and pain in arguing that suffering and its reduction is what we (should) care about.

I did so in my article on Tranquilism, so I broadly share your perspective!

I wouldn't go as far as what you're saying in endnote 9, though. I mean, I see some chance that you're right in the impractical sense of, "If someone gave up literally all they cared about in order to pursue ideal meditation training under ideal circumstances (and during the trainin... (read more)

1jbkjr
Lukas, thanks for taking the time to read and reply! I appreciate you reminding me of your article on Tranquilism—it's been a couple of years since I read it (during my fellowship with CLR), and I hadn't made a mental note of it making such a distinction when I did, so thanks for the reminder. While I agree that it's an open question as to how effective meditation is for alleviating suffering at scale (e.g. how easy it is for how many humans to reduce their suffering by how much with how much time/effort), I don't think it would require as much of a commitment as you seem to expect in the median case. Personally, I think it's likely that the median person would be able to make substantial progress in reducing suffering as a layperson, i.e. without becoming a monastic. Even if attaining a total and forevermore cessation of suffering is substantially more difficult/attainable by substantially fewer people in one lifetime, I don't think it's unreasonable to think that most people could suffer at least 50 percent less with dedicated mindfulness practice. I'm curious as to what might feed an opposing intuition for you! I'd be quite excited about empirical research that investigates the tractability and scalability of meditation for reducing suffering, in either case. (By the way, would it be alright if I ping you privately to set up a meeting? I've been a fan of your writing since becoming familiar with you during my time at CLR and would love a chance to pick your brain about SFE stuff and hear about what you've been up to lately!)

This would be a valid rebuttal if instruction-tuned LLMs were only pretending to be benevolent as part of a long-term strategy to eventually take over the world, and execute a treacherous turn. Do you think present-day LLMs are doing that? (I don't)

Or that they have a sycophancy drive. Or that, next to "wanting to be helpful," they also have a bunch of other drives that will likely win over the "wanting to be helpful" part once the system becomes better at long-term planning and orienting its shards towards consequentialist goals. 

On that latter model... (read more)

I thought the first paragraph and the boldened bit of your comment seemed insightful. I don't see why what you're saying is wrong – it seems right to me (but I'm not sure).

7habryka
(I didn't get anything out of it, and it seems kind of aggressive in a way that seems non-sequitur-ish, and also I am pretty sure mischaracterizes people. I didn't downvote it, but have disagree-voted with it)

I am not convinced MIRI has given enough evidence to support the idea that unregulated AI will kill everyone and their children.

The way you're expressing this feels like an unnecessarily strong bar. 

I think advocacy for an AI pause already seems pretty sensible to me if we accept the following premises: 

  • The current AI research paradigm mostly makes progress in capabilities before progress in understanding. 
    (This puts AI progress in a different reference class from most other technological progress, so any arguments with base rates from "tech
... (read more)
4Sen
The bar is very low for me: If MIRI wants to demand the entire world shut down an entire industry, they must be an active research institution actively producing agreeable papers. AI is not particularly unique even relative to most technologies. Our work on chemistry in the 1600's-1900's far outpaced our level of true understanding of chemistry, to the point where we only had a good model of an atom in the 20th century. And I don't think anyone will deny the potential dangers of chemistry. Other technologies followed a similar trajectory. We don't have to agree that the range is 20-80% at all, never mind the specifics of it. Most polls demonstrate researchers find around 5-10% chance of total extinction on the high end. MIRI's own survey finds a similar result! 80% would be insanely extreme. Your landscape of experts is, I'm guessing, your own personal follower list and not statistically viable.   

Would most existing people accept a gamble with 20% of chance of death in the next 5 years and 80% of life extension and radically better technology? I concede that many would, but I think it's far from universal, and I wouldn't be too surprised if half of people or more think this isn't for them.

I personally wouldn't want to take that gamble (strangely enough I've been quite happy lately and my life has been feeling meaningful, so the idea of dying in the next 5 years sucks).

(Also, I want to flag that I strongly disagree with your optimism.)
 

8Matthew Barnett
For what it's worth, while my credence in human extinction from AI in the 21st century is 10-20%, I think the chance of human extinction in the next 5 years is much lower. I'd put that at around 1%. The main way I think AI could cause human extinction is by just generally accelerating technology and making the world a scarier and more dangerous place to live. I don't really buy the model in which an AI will soon foom until it becomes a ~god.
6Seth Herd
I like this framing. I think the more common statement would be 20% chance of death in 10-30 years , and 80% chance of life extension and much better technology that they might not live to see. I think the majority of humanity would actually take this bet. They are not utilitarians or longtermists. So if the wager is framed in this way, we're going full steam ahead.
Lukas_Gloor*3319

we have found Mr Altman highly forthcoming

That's exactly the line that made my heart sink.

I find it a weird thing to choose to say/emphasize.

The issue under discussion isn't whether Altman hid things from the new board; it's whether he hid things to the old board a long while ago.

Of course he's going to seem forthcoming towards the new board at first. So, the new board having the impression that he was forthcoming towards them? This isn't information that helps us much in assessing whether to side with Altman vs the old board. That makes me think: why repo... (read more)

Followed immediately by: 

I too also have very strong concerns that we are putting a person whose highest stats are political maneuvering and deception, who is very high in power seeking, into this position. By all reports, you cannot trust what this man tells you.

2bhauth
Yes, but Zvi's earlier posts were more positive about Altman. I just picked a relatively recent post, written after the board fired him.

For me, the key question in situations when leaders made a decision with really bad consequences is, "How did they engage with criticism and opposing views?"

If they did well on this front, then I don't think it's at all mandatory to push for leadership changes (though certainly, the worse someones track record gets, the more that speaks against them).

By contrast, if leaders tried to make the opposition look stupid or if they otherwise used their influence to dampen the reach of opposing views, then being wrong later is unacceptable.

Basically, I want to all... (read more)

I agree with what you say in the first paragraph. If you're talking about Ilya, which I think you are, I can see what you mean in the second paragraph, but I'd flag that even if he had some sort of plan here, it seems pretty costly and also just bad norms for someone with his credibility to say something that indicates that he thinks OpenAI is on track to do well at handling their great responsibility, assuming he were to not actually believe this. It's one thing to not say negative things explicitly; it's a different thing to say something positive that r... (read more)

Lukas_Gloor2518

It seems likely (though not certain) that they signed non-disparagement agreements, so we may not see more damning statements from them even if that's how they feel. Also, Ilya at least said some positive things in his leaving announcement, so that indicates either that he caved in to pressure (or too high agreeableness towards former co-workers) or that he's genuinely not particularly worried about the direction of the company and that he left more because of reasons related to his new project. 

Someone serious about alignment seeing dangers better do what is save and not be influenced by a non-disparagement agreement. It might lose them some job prospects and have money and possible lawsuit costs, but if history on earth is on the line? Especially since such a known AI genius would find plenty support from people who supported such open move.

So I hope he assumes talking right NOW it not considered strategically worth it. E.g. He might want to increase his chance to be hired by semi safety serious company (more serious than Open AI, but not enough to hire a proven whistleblower), where he can use his position better.

I agree: appealing to libertarianism shouldn't automatically win someone the argument on whether it's okay to still have factory farms.

The fact that Zvi thought he provided enough of a pointer to an argument there feels weird, in my opinion.

That said, maybe he was mostly focused on wanting to highlight that a large subset of people who are strongly against this ban (and may use libertarian arguments to argue for their position) are only against bans when it suits their agenda. So, maybe the point was in a way more about specific people's hypocrisy in how t... (read more)

3Andrew Burns
Zvi is talking about those people who use libertarianism as a gloss for "getting what they want." In other words, people who aren't into liberty per se, but only into liberty to the extent it satisfies their preferences. There probably is, and if there isn't, there should be, a word for people who invoke liberty this way. That way, when talking about the sort that, for instance, want children to be allowed to read the Bible in the classroom (because LIBERTY!) while simultaneously wanting to ban some book on trans-youth (because PARENTS RIGHTS), we can say: oh, yes, that (word) is at it again. I mean, hypocrite for sure, and perhaps gaslighter, but we need a better word. Well, if there is an existing word, please let me know. There are so many of these sorts out and about, they easily dwarf the population of libertarians.

I think one issue is that someone can be aware about a specific worldview's existence and even consider it a plausible worldview, but still be quite bad at understanding what it would imply/look like in practice if it were true. 

For me personally, it's not that I explicitly singled out the scenario that happened and assigned it some very low probability. Instead, I think I mostly just thought about scenarios that all start from different assumptions, and that was that.

For instance, when reading Paul's "What failure looks like" (which I had done multip... (read more)

1denkenberger
I did have some probability mass on AI boxing being relevant. And I still have some probability mass that there will be sudden recursive self-improvement. But I also had significant probability mass on AI being economically important, and therefore very visible. And with an acceleration of progress, I thought many people would be concerned about it. I don’t know as I would’ve predicted a particular chat-gpt moment (I probably would have guessed some large AI accident), but the point is that we should have been ready for a case when the public/governments became concerned about AI. I think the fact that there were some AI governance efforts before chat-gpt was due in large part to the people saying there could be slow take off, like Paul.

I lean towards agreeing with the takeaway; I made a similar argument here and would still bet on the slope being very steep inside the human intelligence level. 

In some of his books on evolution, Dawkins also said very similar things when commenting on Darwin vs Wallace, basically saying that there's no comparison, Darwin had a better grasp of things, justified it better and more extensively, didn't have muddled thinking about mechanisms, etc.

1francis kafka
I mean to some extent, Dawkins isn't a historian of science, presentism, yadda yadda but from what I've seen he's right here. Not that Wallace is somehow worse, given that of all the people out there he was certainly closer than the rest. That's about it

Very cool! I used to think Hume was the most ahead of his time, but this seems like the same feat if not better.

5dr_s
Democritus also has a decent claim to that for being the first to imagine atoms and materialism altogether.

Yeah, you need an enormous bankroll to play $10,000 tournaments. What a lot of pros do is sell action. Let's say you're highly skilled and have a, say, 125% expected return on investment. If you find someone with a big bankroll and they're convinced of your skills, you can you sell them your action at a markup somewhere between 1 and 1.2 to incentivize them to make a profit. I'd say something like 1.1 markup is fairest, so you're paying them a good prize to weather the variance for you.  At 1.1 markup, they pay 1.1x whatever it costs you to buy into t... (read more)

You also quote this part of the article:

Theo Boer, a healthcare ethics professor at Protestant Theological University in Groningen, served for a decade on a euthanasia review board in the Netherlands. “I entered the review committee in 2005, and I was there until 2014,” Boer told me. “In those years, I saw the Dutch euthanasia practice evolve from death being a last resort to death being a default option.” He ultimately resigned. 

I found a submission by this Theo Boer for the UK parliament, where he explains his reasons for now opposing euthanasia in ... (read more)

Assisted Suicide Watch

A psychiatrist overstepping their qualifications by saying “It’s never gonna get any better” ((particularly when the source of the suffering is at least partly BPD, for which it's commonly known that symptoms can get better in someone's 40s)) clearly should never happen.

However, I'd imagine that most mental health professionals would be extremely careful when making statements about whether there's hope for things to get better. In fact, there are probably guidelines around that.

Maybe it didn't happen this way at all: I notice I'm con... (read more)

8Lukas_Gloor
You also quote this part of the article: I found a submission by this Theo Boer for the UK parliament, where he explains his reasons for now opposing euthanasia in more detail. He writes: This is a "slope" of sorts, but I think it's not a bad one.  The arguments for extending the practice all seem reasonable. What matters is, "are people suffering?" and, "are they right that there's not enough hope for them to justify continued suffering?" Regarding pressure/being pressured, I thought this part was interesting: I'd be curious to figure out why it is exactly that requests for euthanasia are higher in demographs where people tend to be better off/suffering less. That said, I'm not at all convinced that this would prove that there's something horribly wrong going on with these developments after legalization of assisted suicide. (Still, I'd be curious to investigate this further.)  Reading this account, it feels to me like Theo Boer has a problem with death intrinsically, as opposed to only having a problem with death when a person has for-themselves good/strong reasons to continue to want to live. That's not an outlook I agree with. "Their own incapacity to face life's harder episodes" is a question-begging phrasing. For all we know, many people who choose assisted suicide would voluntarily chose to continue with their suffering if there was more at stake that they cared about! For instance, if they learned that by continuing to suffer, they'd solve world poverty, they might continue to suffer. It seems wrong, then, to say they're "incapable," when the real reason is more about how they don't want it enough. It's their life, so their decision. "Since the autonomous citizens are not under any other pressure than their own" – this is also an interesting perspective. He seems to be conceding that no matter how much society and relatives try to reassure chronically ill or disabled elderly people that they're still valued and cared about (something we absolutely m

If you know you have a winning hand, you do not want your opponent to fold, you want them to match your bet. So you kinda have to balance optimizing for the maximum pool at showdown with limiting the information you are leaking so there is a showdown. Or at least it would seem like that to me, I barely know the rules. 

This is pretty accurate.

For simplicity, let's assume you have a hand that has a very high likelihood of winning at showdown on pretty much any runout. E.g., you have KK on a flop that is AK4, and your opponent didn't raise you before the... (read more)

1quiet_NaN
Thanks, this is interesting.  From my understanding, in no-limit games, one would want to only have some fraction of ones bankroll in chips on the table, so that one can re-buy after losing an all-in bluff. (I would guess that this fraction should be determined by the Kelly criterion or something.) On the other hand, from browsing Wikipedia, it seems like many poker tournaments prohibit or limit re-buying after going bust. This would indicate that one has limited amounts of opportunity to get familiar with the strategy of the opponents (which could very well change once the stakes change).  (Of course, Kelly is kind of brutal with regard to gambling. In a zero sum game, the average edge is zero, so at least one participant should not be playing even from an EV perspective. But even under the generous assumption that you are 50% more likely than chance to win a 50 participant elimination tournament (e.g. because a third of the participants are actively trying to lose) (so your EV is 0.5 the buy-in) Kelly tells you to wager about 1% of your bankroll. So if the buy-in is 10k$ you would have to be a millionaire.)

I really liked this post! I will probably link to it in the future.

Edit: Just came to my mind that these are things I tend to think of under the heading "considerateness" rather than kindness, but it's something I really appreciate in people either way (and the concepts are definitely linked). 

4silentbob
Guess I'd agree. Maybe I was anchored a bit here by the existing term of computational kindness. :)
Load More