It is possible that state tracking could be the next reasoning-tier breakthrough in frontier model capabilities. I believe that there exists strong evidence in favor of this being the case.
State space models already power the fastest available voice models, such as Cartesia's Sonic (time-to-first-audio advertised as under 40ms). There are examples of SSMs such as Mamba, RWKV, and Titans outperforming transformers in research settings.
Flagship LLMs are also bad at state tracking, even with RL for summarization. Forcing an explicit...
You will always oversample from the most annoying members of a class.
This is inspired by recent arguments on twitter about how vegans and poly people "always" bring up those facts. I content that it's simultaneous true that most vegans and poly people are either not judgmental, but it doesn't matter because that's not who they remember. Omnivores don't notice the 9 vegans who quietly ordered an unsatisfying salad, only the vegan who brought up factoring farming conditions at the table. Vegans who just want to abstain from animal products remember the omniv...
On Twitter at least, a fair number of the cult allegations seem to be from (honestly fairly cult-ish people themselves) who don't like what LW people say about AI, at least in the threads I'm likely to follow. But I defer to your greater HN expertise!
everyone is a few hops away from everyone else. this applies in both directions: when I meet random people they always have some weak connection to other people I know, but also when I think of a collection of people as a cluster, most specific pairs of people within that cluster barely know each other except through other people in the cluster.
It’s worth noting that, though it’s true that for a sufficiently large cluster most pairs of people are not strongly connected, they are significantly more likely to be connected than in a random graph. This is the high clustering coefficient property of small-world graphs like the social graph.
My current view is that alignment theory should work on deep learning as soon as it comes out, if it's the good stuff, and if it doesn't, it's not likely to be useful later unless it helps produce stuff that works on deep learning. Wentworth, Ngo, and Causal Incentives are the main threads that already seem to have achieved this somewhat. SLT and DEC seem potentially relevant.
I'll think about your argument for mechinterp. If it's true that the ratio isn't as catastrophic as I expect it to turn out to be, I do agree that making microscope AI work would be incredible in allowing for empiricism to finally properly inform rich and specific theory.
Long have I searched for an intuitive name for motte & bailey that I wouldn't have to explain too much in conversation. I might have finally found it. The "I was merely saying fallacy". Verb: merelysay. Noun: merelysayism. Example: "You said you could cure cancer and now you're merelysaying you help the body fight colon cancer only."
20 years is a long time sure, but I don’t think would be good argument for keeping it! (I understand you’re likely just describing, not justifying)
Motte & bailey has a major disadvantage of “nobody who hears it for the first time has any understanding of what it means”
Even as someone who knows the concept, I’m still not even 100% positive that motte and bailey do in fact mean “overclaim and retreat” respectively
People are welcome to use the terms they want, of course. But I’d think there should be a big difference between M&B and some simpler name in order to justify M&B
Multiple times have I seen an argument like this:
Imagine a fully materialistic universe strictly following some laws, which are such that no agent from inside the universe is able to fully comprehend them...
I wonder if that is possible? For computables, it is always possible to construct a quine (standing for the agent) with arbitrary embedded contents (for the rest of the universe/laws/etc), and it wouldn't even be that large - it only needs to...
All you need is a bounded universe with laws having complexity greater than can be embedded within that bound, and that premise holds.
You can even have a universe containing agents with unbounded complexity, but laws with infinite complexity describing a universe that only permits agents with finite complexity at any given time.
Someone has posted about a personal case of vision deterioration after taking lumina and a proposed mechanism of action. I learned about lumina on lesswrong a few years back, so sharing this link.
https://substack.com/home/post/p-168042147
For the past several months I have been slowly losing my vision, and I may be able to trace it back to taking the Lumina Probiotic. Or rather, one of its byproducts that isn’t listed in the advertising
I don't know enough about this to make an informed judgement on the accuracy of the proposed mechanism.
Will they help me test my own mouth to determine whether it's even in the realm of possibility? Me, a complete nobody who came out of nowhere with a well written hit piece full of plausible but hopefully completely wrong conjecture about the most horrifying thing that could happen?
I think it's very likely that they'd at least want to talk to you. If they couldn't rule out your proposed theory, I'd guess they're already equipped to test for it and probably would want to given that a lot of their own people are using the product.
If the singularity occurs over two years, as opposed to two weeks, then I expect most people will be bored throughout much of it, including me. This is because I don't think one can feel excited for more than a couple weeks. Maybe this is chemical.
Nonetheless, these would be the two most important years in human history. If you ordered all the days in human history by importance/'craziness', then most of them would occur within these two years.
So there will be a disconnect between the objective reality and how much excitement I feel.
This comment has been tumbling around in my head for a few days now. It seems to be both true and bad. Is there any hope at all that the Singularity could be a pleasant event to live through?
"Changing Planes" by Ursula LeGuin is worth a read if you're looking for a book that's got interesting alignment ideas (specifically what to do with power, not how to get it), while simultaneously being extremely chill. It might actually be the only chill book that I (with a fair degree of license) consider alignment relevant.
To anyone currently going through NeurIPS rebuttals fun for the first time, some advice:
Firstly, if you're feeling down about reviews, remember that peer review has been officially shown to be a ridiculous random number generator in an RCT - half of spotlight papers are rejected by another review committee! Don't tie your self-worth to whether the roulette wheel landed on black or red. If their critiques don't make sense, they often don't (and were plausibly written by an LLM). And if they do make sense (and remember to control for your defensiveness), the...
I deserved all the smoke sent my way this time lol. Next time!
My writing is sloppy. Can anyone please suggest any resources where I can get feedback on my writing, or personalized instructions that will improve my processes to make me a better writer?
In the meantime I'll try to adopt this "one simple trick": each time I write a piece, I will read it out aloud to myself. If it is "tough on the ear" or I stumble while sight reading it, I will edit the offending section until it is neither.
Also, I'll continue to get LLMs to summarize the points in a given piece. If there's something I feel is missing in it's summary or ...
I'm no writer or editor but you could email me. I check my email every few days lemonhope@fastmail.com
The 50M H100 equivalent compute by 2030 figure tweeted by Musk is on trend (assuming a 2028 slowdown), might cost about $300bn in total (for the training systems built in 2025-2030 for one AI company, including the buildings and power infrastructure).
If the current trend of compute scaling continues to 2028, there will be 160x more compute per training system than the 100K H100s of 2024. It will require 5 GW of power and cost about $140bn in compute hardware and an additional $60bn in buildings, power, and cooling infrastructure[1].
However, if the slowdown...
Power infrastructure that might need to be built is gas generators or power plants, substations, whatever the buildings themselves need. Generators are apparently added even when not on-paper strictly necessary, as backup power. They are also faster to setup than GW-scale grid interconnection, so could be important for these sudden giant factories where nobody is quite sure 4 years in advance that they will be actually built at a given scale.
Datacenter infrastructure friction and cost will probably both smooth out the slowdown and disappear as a funding co...
For the last approx. 3.5 years, I’ve been splitting my time between my emotional coaching practice and working for a local startup. I’m still doing the coaching, but I felt like it was time to move on from the startup, which left me with the question of what to do with the freed-up time and reduced money.
Over the years, people have told me things like “you should have a Patreon” or have otherwise wanted to support my writing. Historically, I’ve had various personal challenges with writing regularly, but now I decided to take another shot at it. I...
Thanks for the hint! I did consider a dual Substack/Patreon approach earlier but decided I couldn't be bothered with the cross-posting. I'll consider if it'd be worth publishing a page soon just so I can reserve myself a cheaper rate for the future.
"Metalhead" from Black Mirror is a relevant contemporary art piece.
I for one find Spot spooky as hell. I would go as far as to say that I have heard others express discomfort toward Boston Dynamics demo videos.
Also, sentry guns and UAVs seem like strong examples of extant scary robots. Maybe see also autonomousweaponswatch.org .
Every so often I see the following:
Adam: Does anyone know how to X?
Bella: I asked ChatGPT, and it said you Y then Z.
Adam: Urgh, I could ask ChatGPT myself, why bother to speak up if you don't have anything else to add?
I'm sort of sympathetic to Adam- I too know how to ask ChatGPT things myself- but I think he's making a mistake. Partially because prompting good answers is a bit of a skill (one that's becoming easier and easier as time goes on, but still a bit of a skill.) Mostly because I'm not sure if he's reinforcing people to not answer with LLM answers...
Possible justifications for Bella's response in slightly different hypotheticals.
I previously made a post that hypothesized a combination of the extra oral ethanol from Lumina and genetic aldehyde deficiency may lead to increased oral cancer risks in certain population. It has been cited in a recent post about Lumina potentially causing blindness in humans.
I've found that hypothesis less and less plausible ever since I published that post. I still think it is theoretically possible in a small proportion (extremely bad oral hygene) of the aldehyde deficient population, but even then it is very unlikely to raise the oral cancer incidence...
Yeah, my idea is just based on physical proximity. There's no way systemic concentrations would be enough, plus the E. Coli in the gut produce way more formate in total given the much larger surface area... yet I can't ignore that my mouth is directly below my eyes. I'm totally willing to bet on it, though I don't know how you'd judge something like this. Formate optic neuropathy doesn't necessarily have specific signs, though in the two case reports it does follow a progressive course and then suddenly get much worse. Is it just based on whether I end up ...
Every now and then I'm like "smart phones are killing America / the world, what can I do about that?".
Where I mean: "Ubiquitous smart phones mean most people are interacting with websites in a fair short attention-space, less info-dense-centric way. Not only that, but because websites must have a good mobile version, you probably want your website to be mobile-first or at least heavily mobile-optimized, and that means it's hard to build features that only really work when users have a large amount of screen space."
I'd like some technological solution...
I might separately criticize shortform video and twitter (sure, they definitely have benefits, I just think they also have major costs, and if we can alleviate the costs we should. This doesn't have to mean banning shortform and twitter).
But, I think that's (mostly) a different topic that the OP.
The question here is not "is it good you can post on twitter?", it's "is it good you can post on the version of twitter that was brought into being by "most people using small-screens." (or, more accurately: is it good that we're in the world where small-screen twitter is a dominant force shaping humanity, as opposed to an ecosystem where less-small-screen-oriented social media app is more dominant)
I remember a hygienist at the dentist once telling me that toothpaste isn't a huge deal and that it's the mechanical friction of the toothbrush that provides most of the value. Since being told that, after a meal, I often wet my toothbrush with water and brush for 10 seconds or so.
I just researched it some more and from what I understand, after eating, food debris that remains on your teeth forms a sort of biofilm. Once the biofilm is formed you need those traditional 2 minute long tooth brushing sessions to break it down and remove it. But it takes 30+ mi...
My understanding is that toothpicks are for scraping the area in between teeth, not the surface of the tooth itself.
I am confused about why this post on the ethics of eating honey is so heavily downvoted.
It sparked a bunch of interesting discussion in the comments (e.g. this comment by Habryka and the resulting arguments on how to weight non-human animal experiences)
It resulted in at least one interesting top-level rebuttal post.
I assume it led indirectly to this interesting short post also about how to weight non-human experiences. (this might not have been downstream of the honey post but it's a weird coincidence if isn't)
I think the original post certainly had flaws,...
Let me first try to convey how this conversation appears from my perspective. I don't think I've ever debated directly with you about anything, but I have an impression of you as doing solid work in the areas of your interest.
Then, I run across you alleging that BB is using AI to write some of his articles. This catches my attention because I do keep an eye on BB's work. Furthermore, your reason for supposing that he is using AI seems bizarre to me - you think his (very occasional) "sneering" is too "dumb and cliche" to be the work of human hands. Le...
Act utilitarians choose actions estimated to increase total happiness. Rule utilitarians follow rules estimated to increase total happiness (e.g. not lying). But you can have the best of both: act utilitarianism where rules are instead treated as moral priors. For example, having a strong prior that killing someone is bad, but which can be overridden in extreme circumstances (e.g. if killing the person ends WWII).
These priors make act utilitarianism more safeguarded against bad assessments. They are grounded in Bayesianism (moral priors are updated the sam...