For an event me and Saul Munn are hosting:
Memoria is a one-day festival/unconference for spaced repetition, incremental reading, and memory systems. It’s hosted at Lighthaven in Berkeley, CA, on September 21st, from 10am through the afternoon/evening.
Michael Nielsen, Andy Matuschak, Soren Bjornstad, Martin Schneider, and about 70–90 others will be there — if you use & tinker with memory systems like Anki, SuperMemo, Remnote, MathAcademy, etc, then maybe you should come!
Tickets are $8...
By that I mean: problems it would be worth one million dollars to have solved. That also implies they would have to be solvable in some empirical fashion.
US Government dysfunction and runaway political polarization bingo card. I don't expect any particular one of these to happen but it seems plausible that at least one of these will happen.
The idea is they're printing money, not just borrowing it, which in the extreme would cause hyperinflation (and is equivalent to default since debt is in nominal dollars). It probably seems less bad though.
I think it’s possible that an AI will decide not to sandbag (e.g. on alignment research tasks), even if all of the following are true:
The reason is as follows:
This seems less likely the harder the problem is, and therefore the more the AI needs to use its general intelligence or agency to pursue it, which are often the sorts of tasks we’re most scared about the AI doing surprisingly well on.
I agree this argument suggests we will have a good understanding of more simple capabilities the model has, like what facts about biology it knows about, which may end up being useful anyway.
I have now written a review of the book, which touches on some of what you're asking about. https://www.lesswrong.com/posts/mztwygscvCKDLYGk8/jdp-reviews-iabied
I read somewhere recently that there's a fiber optic fpv kamikaze drone with a 40km range. By contrast typical such drones have 10km, maybe 20km ranges.
Either way, it seems clear that EW-resistant drones with 20km+ range are on the horizon. Millions per year will be produced by Ukraine and Russia and maybe some other countries. And I wouldn't be surprised if ranges go up to more than 40km soon.
I wonder if this will cause major problems for Israel. Gaza and Lebanon and Syria are within 20km of some decent-sized israeli cities. Iron Dome wouldn't work agains...
the importance of drones, I think, is not going to go down thanks to AI.
I agree. What I tried to say though was that my guess is that for drones to stay as effective as they currently are for 5 years would require AI capable enough that it would transform so many aspects of society that for us in 2025 to try to project out that far becomes futile.
can someone please explain to me why "if anyone builds it, everyone dies" is not a free eBook/blog post?
like seriously if someone told me there is a detailed case for a possible imminent existential risk with possible solutions included but i had to pay to see it, i would have dismissed it as another fearmongering doomsday grift.
if you are really sincere about an extinction level risks why hide your arguments behind a paywall? why not make it free so as many people can see it as possible?
the very fact that this book has a price tag on it in an age where publishing an eBook is practically free puts the authors motives in question.
One idea I came up with today, is that the ideal book would also have an online website where you can read it all, conditional on you having bought a book. Essentially a paywall that is also a measurable book sale. And otherwise you can just get highlighted extracts from each chapter.
Just finished If Anyone Builds It, Everyone Dies (and some of the supplements).[1] It feels... weaker than I'd hoped. Specifically, I think Part 3 is strong, and the supplemental materials are quite thorough, but Parts 1-2... I hope I'm wrong, and this opinion is counterweighed by all these endorsements and MIRI presumably running it by lots of test readers. But I'm more bearish on it making a huge impact than I was before reading it.
Point 1: The rhetoric – the arguments and their presentations – is often not novel, just rehearsed variations on the ar...
Thinking about qualia, trying to avoid getting trapped in the hard problem of consciousness along the way.
Tempted to model qualia as a region with the capacity to populate itself with coarse heuristics for difficult-to-compute features of nodes in a search process, which happens to ship with a bunch of computational inconveniences (that are most of what we mean to refer to when we reference qualia).
This aids in generality, but trades off against locally optimal processes, as a kind of 'tax' on all cognition.
This is a literal shower thought and I've read no...
Following up to say that the thing that maps most closely to what I was thinking about (or satisfied my curiosity) is GWT.
GWT is usually intended to approach the hard problem, but the principle critique of it is that it isn't doing that at all (I ~agree). Unfortunately, I had dozens of frustrating conversations with people telling me 'don't spend any time thinking about consciousness; it's a dead end; you're talking about the hard problem; that triggers me; STOP' before someone actually pointed me in the right direction here, or seemed open to the question at all.
Reading so many reviews/responses to IABIED, I wish more people had registered how they expected to feel about the book, or how they think a book on x-risk ought to look, prior to the book's release.
Finalizing any Real Actual Object requires making tradeoffs. I think it's pretty easy to critique the book on a level of abstraction that respects what it is Trying To Be in only the broadest possible terms, rather than acknowledging various sub-goals (e.g. providing an updated version of Nate + Eliezer's now very old 'canonical' arguments), modulations o...
Quick book review of "If Anyone Builds It, Everyone Dies" (cross-post from X/twitter & bluesky):
Just read the new book If Anyone Builds It, Everyone Dies. Upshot: Recommended! I ~90% agree with it.
The authors argue that people are trying to build ASI (superintelligent AI), and we should expect them to succeed sooner or later, even if they obviously haven’t succeeded YET. I agree. (I lean “later” more than the authors, but that’s a minor disagreement.)
...Ultra-fast minds that can do superhuman-quality thinking at 10,000 times the speed, that do not age and
I would start by saying that I mostly agree with you here. On this point specifically, however,
AI capabilities would rebrand as AI safety
I mean, 3 of the leading AI labs (DeepMind, OpenAI, Anthropic) were founded explicitly under or attached to the banner of AI safety. OpenAI and Anthropic were even founded as "the safer alternatives" to DeepMind and OpenAI! You also don't have to go back very far to find AI safety funders and community voices promoting those labs as places to work to advance AI safety (whereas today you'd be hard-pressed to find someo...
“Albania has introduced its first artificial intelligence “minister”, who addressed parliament on Thursday in a debut speech.” lol, what???
Not sure how much this really matters vs is just a PR thing, but it’s maybe something people on here should know about.
I'm no expert on Albanian politics, but I think it's pretty obvious this is just a gimmick with minimal broader significance.
Could HGH supplementation in children improve IQ?
I think there's some weak evidence that yes. In some studies where they give HGH for other reasons (a variety of developmental disorders, as well as cases when the child is unusually small or short), an IQ increase or other improved cognitive outcomes are observed. The fact that this occurs in a wide variety of situations indicates that it could be a general effect that could apply to healthy children.
Examples of studies (caveat: produced with the help of ChatGPT, I'm including null results also). Left colum...
has it been tested on adults a lot?
Prosaic AI Safety research, in pre-crunch time.
Some people share a cluster of ideas that I think is broadly correct. I want to write down these ideas explicitly so people can push-back.
I'm curious if you have a sense of:
1. What the target goal of early-crunch time research should be (i.e. control safety case for the specific model one has at the present moment, trustworthy case for this specific model, trustworthy safety case for the specific model and deference case for future models, trustworthy safety case for all future models, etc...)
2. The rough shape(s) of that case (i.e. white-box evaluations, control guardrails, convergence guarantees, etc...)
3. What kinds of evidence you expect to accumulate given access to these early po...
People here might appreciate my book review: If We Build AI Superintelligence, Do We All Die?
I'd be curious for more takes from smart LessWrong readers.
LLMs are trained on a human-generated text corpus. Imagine an LLM agent deciding whether or not to be a communist. Seems likely (though not certain) it would be strongly influenced by the existing human literature on communism, i.e. all the text humans have produced about communism arguing its pros/cons and empirical consequences.
Now replace 'communism' with 'plans to take over.' Humans have also produced a literature on this topic. Shouldn't we expect that literature to strongly influence LLM-based decisions on whether to take over?
This is an argument I'm...
Is this vibes or was there some kind of study done?
Epistemic status: strolling in Venice on a hot September evening tipsy with my mom
One of the advantages of church is that it forces encourages all members of a society to be part of social groups. This has a wide range of positive results from mental health to disaster preparedness and reduced wealth inequality, increased volunteering and more.
The down side is the whole God part. There are many communities that form without the God aspect, but they tend to be less diverse, harder to get into, and more focused around specific...
I probably should have made it clear: this is not a replacement of capitalism. As the title suggest, this is an alternative to UBI. I think thinking of better ways to do UBI becomes more and more important as AI gets better and better. Already, this would be more efficient along economies of scale line than traditional UBI since it goes from a single person, to a community.
As for getting up to the cap: here's what I was thinking. Once you get to arround ~150 ish, so around Dunbar’s number, it's time to starting splitting, the extra is just to make it so it...
a thing i've noticed rat/autistic people do (including myself): one very easy way to trick our own calibration sensors is to add a bunch of caveats or considerations that make it feel like we've modeled all the uncertainty (or at least, more than other people who haven't). so one thing i see a lot is that people are self-aware that they have limitations, but then over-update on how much this awareness makes them calibrated. one telltale hint that i'm doing this myself is if i catch myself saying something because i want to demo my rigor and prove that i've...
This might be more about miscalibration in perceived relevance of technical exercises inspired by some question. A directly mostly irrelevant exercise that juggles details can be useful, worth doing and even sharing, but mostly for improving model-building intuition and developing good framings in the long term rather than for answering the question that inspired it, especially at a technical level.
So an obvious mistake would be to treat such an exercise as evidence that the person doing/sharing it considers it directly relevant for answering the question ...
Draft thought, posting for feedback:
Many people (eg e/acc) believe that although a single very strong future AI might result in bad outcomes, a multi-agent system with many strong AIs will turn out well for humanity. To others, including myself, this seems clearly false.
Why do people believe this? Here's my thought:
I'm looking forward to seeing your post, because I think this deserves more careful thought.
I think that's right, and that there are some more tricky assumptions and disanalogies underlying that basic error.
Before jumping in, let me say that I think that multipolar scenarious are pretty obviously more dangerous to a first approximation. There may be more carefully thought-out routes to equilibria that might work and are worth exploring. But just giving everyone an AGI and hoping it works out would probably be very bad.
Here's where I think the mistake usual...