Announcing new reacts!
The reacts palette has 17 new reacts:
Gone are 11 old reacts who were getting little use, are redundant with the new reacts, and/or in my opinion weren't being used well (for example, "I'd bet this is false" almost never turned into a bet as I'd hoped, but was primarily used to disagree with added emphasis).
Reacts are a rich way to interact with a body of text, allowing many people to give information about comments, claims, and subclaims more efficiently than writing full comments, and they're something I love that we have on LessWron...
I think I semi-agree with your perception, but I did have a recent experience to the contrary: when I did a throwaway post about suddenly noticing a distinction in the reaction UI, I found it very odd that some people marked the central bit “Insightful”. Like, maybe it's useful to them that I pointed it out, but it's a piece of UI that was (presumably) specifically designed that way by someone already! There's no new synthesis going on or anything; it's not insightful. (Or maybe people wanted to use it as a test of the UI element, but then why not the paperclip or the saw-that eyes?)
Awhile back, I claimed that LLMs had not produced original insights, resulting in this question: https://www.lesswrong.com/posts/GADJFwHzNZKg2Ndti/have-llms-generated-novel-insights
To be clear, I wasn’t talking about the kind of weak original insight like saying something technically novel and true but totally trivial or uninteresting (like, say, calculating the value of an expression that was maybe never calculated before but could be calculated with standard techniques). Obviously, this is kind of a blurred line, but I don’t think it’s an empty claim at ...
Thoughts on integrity and accountability
[Epistemic Status: Early draft version of a post I hope to publish eventually. Strongly interested in feedback and critiques, since I feel quite fuzzy about a lot of this]
When I started studying rationality and philosophy, I had the perspective that people who were in positions of power and influence should primarily focus on how to make good decisions in general and that we should generally give power to people who have demonstrated a good track record of general rationality. I also thought of power as this mostly u...
Dwarkesh asked a very interesting question in his Sutton interview, which Sutton wasn't really interested in replying to.
Dwarkesh notes that one idea for why the the bitter lesson was true is because general methods got to ride the wave of exponential computing power while knowledge engineering could not, as labour was comparatively fixed in supply. He then notices that post AGI labour supply will increase at a very rapid pace. And so he wonders, once the labour constraint is solved post AGI will GOFAI make a comeback? For we will then be able to afford the proverbial three billion philosophers writing lisp predicates or whatever various other kinds of high-labour AI techniques are possible.
I don't actually buy this argument, but I think it's a very important argument for someone to make, and for people to consider carefully. So thank you to Dwarkesh for proposing it, and to you for mentioning it!
I've been writing up a long-form argument for why "Good Old Fashioned AI" (GOFAI) is a hopeless pipedream. I don't know if that would actually remain true for enormous numbers of superintelligent programmers! But if I had to sketch out the rough form of the argument, it would go something like this:
Notes and reflections on the things I've learned while Doing Scholarship™ this week (i.e. studying math)[1].
No worries! For more recommendations like those two, I'd suggest having a look at "The Fast Track" on Sheafification. Of the books I've read from that list, all were fantastic. Note that site emphasises mathematics relevant for physics, and vice versa, so it might not be everyone's cup of tea. But given your interests, I think you'll find it useful.
I like this line of thinking - what other things are possible and desirable (to some), but are prevented universally? Nuclear and biological weapons qualify, but they're nowhere near banned, just limited.
I don't know enough about human genetic laws to know what's actually highly-desirable (by some) and prohibited so effectively that it doesn't happen. Cloning seems a non-issue, as there's so little benefit compared to the mostly-allowed IVF and embryo selection processes available.
i recently ran into to a vegan advocate tabling in a public space, and spoke briefly to them for the explicit purpose of better understanding what it feels like to be the target of advocacy on something i feel moderately sympathetic towards but not fully bought in on. (i find this kind of thing very valuable for noticing flaws in myself and improving; it's much harder to be perceptive of one's own actions otherwise). the part where i am genuinely quite plausibly persuadable of his position in theory is important; i think if i had talked to e.g flat earther...
you mentioned sometimes people are just wrong in their arguments but think they are correct because they've repeated it many times. do you have examples of this from what they said?
I have two arguments against it.
Original thread: https://www.lesswrong.com/posts/AZwgfgmW8QvnbEisc/cfar-update-and-new-cfar-workshops?commentId=q5EiqCq3qbwwpbCPn
Summary of my view: I'm upset about the blasé attitude our community seems to have towards its high prevalence of psychosis. I think that CFAR/rationalist leadership (in addition to the community-at-large) has not responded appropriately.
I think Anna agrees with the first point but not the second. Let me know if that's wrong, Anna.
My hypothes...
The existing literature (e.g. UpToDate) about psychosis in the general population could be a good source of priors. Or, is it safe to assume that Anna and you are already thoroughly familiar with the literature?
I believe that we're going to see heavy political and social instability over the next 5 years, how do I maximize my EV in light of this? Primarily I'm thinking about financial investments.
Some things I was thinking about: Gold in GDX, Cybersecurity in HACK, Options income in JEPI, Defense/Aerospace in ITA
As an alternative to Inkhaven Residency, for the people who can't take a 30 day break, I propose an alternative. I am writing about it here, in case someone else also wants to do something, so that we could coordinate. Please respond here if you plan something, otherwise I will announce my plan officially in a LW post on Sept 30.
My current vision, which I provisionally call Halfhaven, is to make an online blog-writing group, probably on Discord, that will try to produce the same total output in twice as much time. That is 30 articles posted during October ...
Thank you, I saw that comment previously but couldn't find it. I will contact henryaj.
Funny quote about covering AI as a journalist from a New York Times article about the drone incursions in Denmark.
...Then of course the same mix of uncertainty and mystery attaches to artificial intelligence (itself one of the key powers behind the drone revolution), whose impact is already sweeping — everyone’s stock market portfolio is now pegged to the wild A.I. bets of the big technology companies — without anyone really having clarity about what the technology is going to be capable of doing in 2027, let alone in 2035.
Since the job of the pundit is
That sounds impressively self-aware. Most journalists would just "predict" the future, expressing certainty.
if you were looking for a sign from the universe, from the simulation, that you can stop working on AI, this is it. you can stop working on AI.
work on it if you want to.
don't work on it if you don't want to.
update how much you want to based on how it feels.
other people are working on it too.
if you work on AI, but are full of fear or depression or darkness, you are creating the danger. which is fine if you think that's funny but it's not fine if you unironically are afraid, and also building it.
if you work on AI, but are full of hopium, copium, and not-obse...
Empirical observation: healthy foods often taste bad.
Why do my taste buds like fat and sugar, instead of a vegetable smoothies? Here's my half-serious attempt at applying formal reasoning to explain the phenomenon.
Proposition.
At a food consumption equilibrium, all healthy foods have significant downsides (tastiness, price, ease of preparation, etc.).
Proof.
We say that a food is "healthy" if the average person would benefit from eating more of it.
Consider an arbitrary food X which doesn't have significant downsides.
We assume that food consumption is at an eq...
Many foods have been selected for palatability.
Some of this selection has taken place through natural selection: wild animals choosing to eat fruits they find tasty (and disperse their seeds). In this case, because the animal consumer is also evolving, it's likely to end up favoring (and thus propagating) fruits that are actually nutritious too. After all, an animal that chose to eat foods that were bad for it, would be less likely to survive and prosper. So wild animals' tastes and their food plants' properties coevolve into something resembling an equili...
An early draft of a paper I'm writing went like this:
In the absence of sufficient sanity, it is highly likely that at least one AI developer will deploy an untrusted model: the developers do not know whether the model will take strategic, harmful actions if deployed. In the presence of a smaller amount of sanity, they might deploy it within a control protocol which attempts to prevent it from causing harm.
I had to edit it slightly. But I kept the spirit.
The background of the Stanislav Petrov incident is literally one of the dumbest and most insane things I have ever read (attached screenshot below):
Appreciate that this means:
It's a wonder we're still here.
My working model of psychosis is "lack of a stable/intact ego", where my working model of an "ego" is "the thing you can use to predict your own actions so as to make successful multi-step plans, such as 'I will buy pasta, so that I can make it on Thursday for our guests.'
from Adele Lopez's Shortform. I probably haven't experienced psychosis, but the description of self-prediction determining behavior/planning, and this self-prediction being faulty or unstable was eerie; this dominates my experience. I'm unsure about their definition of ego, I understood i...
The AI sycophancy-related trance is probably one of the worst news in AI alignment. About two years ago someone proposed to use prison guards to ensure that they aren't CONVINCED to release the AI. And now the AI demonstrates that its primitive version can hypnotise the guards. Does it mean that human feedback should immediately be replaced with AI feedback or feedback on tasks with verifiable reward? Or that everyone should copy the KimiK2 sycophancy-beating approach? And what if it instills the same misalignment issues in all models in the world?
Al...
FYI: METR is actively fundraising!
METR is a non-profit research organization. We prioritise independence and trustworthiness, which shapes both our research process and our funding options. To date, we have not accepted payment from frontier AI labs for running evaluations.[1]
Part of METR's role is to independently assess the arguments that frontier AI labs put forward about the safety of their models. These arguments are becoming increasingly complex and dependent on nuances of how models are trained and how mitigations were developed.
For this...
Can you say anything about what METR's annual budget/runway is? Given that you raised $17mn a year ago, I would have expected METR to be well funded
In addition to Lighthaven for which we have a mortgage, Lightcone owns an adjacent property that is fully unencumbered that's worth around $1.2M. Lighthaven has basically been breaking even, but we still have a funding shortfall of about $1M for our annual interest payment for the last year during which Lighthaven was ramping up utilization. It would be really great if we could somehow take out our real estate equity to cover that one-time funding shortfall.
If you want to have some equity in Berkeley real estate, and/or Lightcone's credit-worthiness, you m...