All of tmeanen's Comments + Replies

tmeanen30

Does LessWrong have a strategy for getting the ideas posted on this site out to other communities (e.g. academia, decision-makers at frontier labs, policy circles, etc)? My impression is that there are a whole lot of potentially impactful ideas floating around on this site, such as important macrostrategic considerations for short-timelines, fast-ish takeoff worlds. Do the LW mods have a strategy to get the right people hearing these ideas, or do we just wait until someone important stumbles across the site?

tmeanen71

fab will check design plans and inform designers what can and can't be manufactured

I'm curious - what are the most common reasons for this? I.e: what are the common requests that designers make that fabs can't manufacture? 

tmeanen10

Seems like a useful resource to have out there. Some other information that would be nice to have are details about the security of the data center - but there's probably limited information that could be included [1]

  1. ^

    Because you probably don't want too many details about your infosec protocols out there for the entire internet to see. 

tmeanen10

Reconnaissance might be a candidate for one of the first uses of powerful A(G)I systems by militaries - if this isn't already the case. There's already an abundance of satellite data (likely exabytes in the next decade) that could be thrown into training datasets. It's also less inflammatory than using AI systems for autonomous weapon design, say, and politically more feasible. So there's a future in which A(G)I-powered reconnaissance systems have some transformative military applications, the military high-ups take note, and things snowball from there.&nb... (read more)

2Seth Herd
Sure, at the low end. I think there are lots of reasons the government is and will continue to be highly interested in AI for military purposes. That's AI; I'm thinking about competent, agentic AGI that also follows human orders. I think that's what we're likely to get, for reasons I go into in the instruction-following AGI link above.
tmeanen5-4

But if the core difficulty in solving alignment is developing some difficult mathematical formalism and figuring out relevant proofs then I think we won't suffer from the problems with the technologies above. In other words, I would feel comfortable delegating and overseeing a team of AIs that have been tasked with solving the Riemann hypothesis - and I think this is what a large part of solving alignment might look like.

9Carl Feynman
“May it go from your lips to God’s ears,” as the old Jewish saying goes.  Meaning, I hope you’re right.  Maybe aligning superintelligence will largely be a matter of human-checkable mathematical proof. I have 45 years experience as a software and hardware engineer, which makes me cynical. When one of my designs encounters the real world, it hardly ever goes the way I expect.  It usually either needs some rapid finagling to make it work (acceptable) or it needs to be completely abandoned (bad).  This is no good for the first decisive try at superalignment; that has to work first time.  I hope our proof technology is up to it.  
tmeanen10

I've been in a number of arguments where people say things like "why is 90% doom such a strong claim? That assumes that survival is the default! "

Am I misunderstandng this sentence? How do "90% doom" and the assumption that survival is the default square with one another? 

2Richard_Ngo
Edited for clarity now.
1CstineSublime
I think they are just using that as an example of a strongly opinionated sub-agent which may be one of many different and highly specific probability assessments of doom. As for "survival is the default assumption" - what a declaration of that implies on the surface level is that the chance of survival is overwhelming except in the case of a cataclysmic AI scenario. To put it another way: we have a 99% chance of survival so long as we get AGI right. To put it yet another way - Hollywood has made popular films about the human world being destroyed by Nuclear War, Climate Change, Viral Pandemic, and Asteroid Impact to name a few - different sub-agents could each give higher or lower probabilities to each of those scenarios depending on things like domain knowledge and in concert it raises the question of why we presume that survival is the default? What is the ensemble average of doom? Is doom more or less likely than survival for any given time frame?
tmeanen61

“keyboard and monitor I’m using right now, a stack of books, a tupperware, waterbottle, flip-flops, carpet, desk and chair, refrigerator, sink, etc. Under my models, if I pick one of these objects at random and do a deep dive researching that object, it will usually turn out to be bad in ways which were either nonobvious or nonsalient to me, but unambiguously make my life worse"

But, I think the negative impacts that these goods have on you are (mostly) realized on longer timescales - say, years to decades. If you’re using a chair that is bad for your postu... (read more)

Many research tasks have very long delays until they can be verified.  The history of technology is littered with apparently good ideas that turned out to be losers after huge development efforts were poured into them.  Supersonic transport, zeppelins, silicon-on-sapphire integrated circuits, pigeon-guided bombs, object-oriented operating systems, hydrogenated vegetable oil, oxidative decoupling for weight loss…

Finding out that these were bad required making them, releasing them to the market, and watching unrecognized problems torpedo them.  Sometimes it took decades.

tmeanen30

Plausibly one technology that arrives soon after superintelligence is powerful surveillance technology that makes enforcing commitments significantly easier than it historically has been. Leaving aside the potential for this to be misused for authoritarian government, advocating for this to be developed before powerful technologies of mass destruction may be a strategy.  

tmeanen20

Nice, I like this concept of rogue deployment as it highlights two distinct features that are both required for a safety method to be considered 'successful'. I'm understanding catastrophe with rogue deployment as having good enough safety measures but these safety measures were bypassed/turned off, whereas catastrophe without rogue deployment involves having safety measures that were fully operational the whole time but insufficient to prevent a model/human actor from causing a catastrophe.  

So for example, we could get really great mech. interp tool... (read more)