John Wentworth explains natural latents – a key mathematical concept in his approach to natural abstraction. Natural latents capture the "shared information" between different parts of a system in a provably optimal way. This post lays out the formal definitions and key theorems.
This post presents a mildly edited form of a new paper by UK AISI's alignment team (the abstract, introduction and related work section are replaced with an executive summary). Read the full paper here.
AI safety via debate is a promising method for solving part of the alignment problem for ASI (artificial superintelligence).
TL;DR Debate + exploration guarantees + solution to obfuscated arguments + good human input solves outer alignment. Outer alignment + online training solves inner alignment to a sufficient extent in low-stakes contexts.
This post sets out:
These gaps form the basis for one of the research agendas of UK AISI’s new alignment team: we aim to dramatically scale up ASI-relevant research on debate. We’ll...
Two reasons:
You seem to be referring to comments from the CEO that more than 25% of code at Google is written by AI (and reviewed by humans). I’m not sure how reliable this number is, and it remains to be seen whether this is sustainable. It also doesn’t seem like a vast productivity boost (though it would be pretty significant, probably more than I expect, so would update me).
Eliezer's AI doom arguments have had me convinced since the ancient days of 2007, back when AGI felt like it was many decades away, and we didn't have an intelligence scaling law (except to the Kurzweilians who considered Moore's Law to be that, and were, in retrospect, arguably correct).
Back then, if you'd have asked me to play out a scenario where AI passes a reasonable interpretation of the Turing test, I'd have said there'd probably be less than a year to recursive-self-improvement FOOM and then game over for human values and human future-steering control. But I'd have been wrong.
Now that reality has let us survive a few years into the "useful highly-general Turing-Test-passing AI" era, I want to be clear and explicit about how I've updated my...
Liron: ... Turns out the answer to the symbol grounding problem is like you have a couple high dimensional vectors and their cosine similarity or whatever is the nature of meaning.
Could someone state this more clearly?
Jim: ... a paper that looked at the values in one of the LLMs as inferred from prompts setting up things like trolley problems, and found first of all, that they did look like a utility function, second of all, that they got closer to following the VNM axioms as the network got bigger. And third of all, that the utility function that they seemed to represent was absolutely bonkers
What paper was this?
For months, I had the feeling: something is wrong. Some core part of myself had gone missing.
I had words and ideas cached, which pointed back to the missing part.
There was the story of Benjamin Jesty, a dairy farmer who vaccinated his family against smallpox in 1774 - 20 years before the vaccination technique was popularized, and the same year King Louis XV of France died of the disease.
There was another old post which declared “I don’t care that much about giant yachts. I want a cure for aging. I want weekend trips to the moon. I want flying cars and an indestructible body and tiny genetically-engineered dragons.”.
There was a cached instinct to look at certain kinds of social incentive gradient, toward managing more people or growing an organization or playing...
Yeah, the underling part was a joke :D
Midjourney, “Fourth Industrial Revolution Digital Transformation”
This is a little rant I like to give, because it’s something I learned on the job that I’ve never seen written up explicitly.
There are a bunch of buzzwords floating around regarding computer technology in an industrial or manufacturing context: “digital transformation”, “the Fourth Industrial Revolution”, “Industrial Internet of Things”.
What do those things really mean?
Do they mean anything at all?
The answer is yes, and what they mean is the process of putting all of a company’s data on computers so it can be analyzed.
This is the prerequisite to any kind of “AI” or even basic statistical analysis of that data; before you can start applying your fancy algorithms, you need to get that data in one place, in a tabular format.
This is made more difficult because a large portion of those running trials do not do the data management and/or analysis in-house, instead outsourcing those tasks to CROs (Contract Research Organizations). Inter-organization communication barriers certainly don't make the disconnect any easier to resolve.
...It's blogging but shorter. I'll give it a better name if I think of one.
Low-IQ voters can't identify good policies or wise politicians; democracy favors political actors who can successfully propagandize and mobilize the largest number of people, which might not correspond to good governance. A political system with non-democratic elements that offers more formalized control to actors with greater competence or better incentives might be able to choose better policies.
I say "non-democratic elements" because it doesn't have to be a strict binary between perfect democracy and perfect dictatorship. Consider, e.g., how the indirec...
Let’s create a list of which journalists LessWrongers trust, so as to gave a guide if people get contacted.
Agree votes are for more good.
Upvotes aren’t helpful because i think heavily downvoted comments get hidden (and i expect many journalists to be underwater). If you want to use upvotes, I suggest they are for if someone is a journalist or not.
Please add each journalist as a separate entry. I will delete any entries that include multiple people. If you’d prefer not to add someone yourself, feel free to DM me.
The title question is a proxy for the thing I mean:
Kenny Jones (Dream House Podcast)
Some people (the “Boubas”) don’t like “chemicals” in their food. But other people (the “Kikis”) are like, “uh, everything is chemicals, what do you even mean?”
The Boubas are using the word “chemical” differently than the Kikis, and the way they’re using it is simultaneously more specific and less precise than the way the Kikis use it. I think most Kikis implicitly know this, but their identities are typically tied up in being the kind of person who “knows what ‘chemical’ means”, and… you’ve gotta use that kind of thing whenever you can, I guess?
There is no single privileged universally-correct answer to the question “what does ‘chemical’ mean?”, because the Boubas exist and are using the word differently than Kikis, and in an internally-consistent (though vague) way.
The Kikis...
If they're just using the word "chemical" as an arbitrary word for "bad substance", you have the situation I already described: the word isn't communicating anything useful.
But in practice, someone who claims that they don't want chemicals in their food probably doesn't just mean "harmful substances". They probably mean that they have some criteria for what counts as a harmful substance, and that these criteria are based on traits of things that are commonly called chemicals. When you tell them "wait, water and salt are chemicals", what you're really doing is forcing them to state those criteria so you can contest them (and so they can become aware that that's what they're using).
There should be a community oriented towards the genomic emancipation of humanity. There isn't such a community, but there should be. It's a future worth investing our hope in—a future where parents are able to choose to give their future children the gift of genomic foundations for long, healthy, sane, capable lives.
We're inaugurating this community with the Reproductive Frontiers Summit 2025 in Berkeley, CA, June 10—12. Come join us if you want to learn, connect, think, and coordinate about the future of germline engineering technology. Apply to attend by filling out this (brief) form: https://forms.gle/xjJCaiNqLk7YE4nt8
Our lineup of speakers includes: