All of MrFailSauce's Comments + Replies

A traditional Turing machine doesn't make a distinction between program and data.  The distinction between program and data is really a hardware efficiency optimization that came from the Harvard architecture.   Since many systems are Turing complete, creating an immutable program seems impossible to me.  

For example a system capable of speech could exploit the Turing completeness of formal grammars to execute de novo subroutines.  

A second example.  Hackers were able to exploit the surprising Turing completeness of an image compression standard to embed a virtual machine in a gif.

https://googleprojectzero.blogspot.com/2021/12/a-deep-dive-into-nso-zero-click.html

2philh
Well, a regular Turing machine does, it has a tape and a state machine and the two are totally different. I guess you mean a traditional universal Turing machine doesn't distinguish between "Turing machine I'm simulating" and "data I'm simulating as input to that Turing machine".

I feel like an important lesson to learn from analogy to air conditioners is that some technologies are bounded by physics and cannot improve quickly.(or at all).   I doubt anyone has the data, but I would be surprised if average air conditioning efficiency in BTUs per Watt plotted over the 20th century is not a sigmoid.

For seeing through the fog of war, I'm reminded of the German Tank Problem.  

https://en.wikipedia.org/wiki/German_tank_problem

Statistical estimates were ~50x more accurate than intelligence estimates in the cannonical example.  When you include the strong and reasonable incentives for all participants to propagandize, it is nearly impossible to get accurate information about an ongoing conflict.

I think as rationalists, if we're going to see more clearly than conventional wisdom, we need to find sources of information that have more fundamental basis.  I don't yet know what those would be.

0CraigMichael
Or weight the sources appropriately given the evidence and update when we realize we’ve made a mistake?

In reality, an AI can use algorithms that find a pretty good solution most of the time. 

If you replace "AI" with "ML" I agree with this point. And yep this is what we can do with the networks we're scaling.  But "pretty good most of the time" doesn't get you an x-risk intelligence.  It gets you some really cool tools.

If the 3 sat algorithm is O(n^4) then this algorithm might not be that useful compared to other approaches. 

If 3 SAT is O(n^4) then P=NP and back to Aaronson's point; the fundamental structure of reality is much different t... (read more)

2Donald Hobson
On some problems, finding the exact best solution is intractable. On these problems, its all approximations and tricks that usually work. Whether the simplest dumb ML algorithm running with didly squat compute, or some vastly superintelligent AI running on a Matrioska brain.  Take hacking a computer system that controls nukes or something. The problem of finding the fastest way to hack an arbitrary computer system is NP hard. But humans sometimes hack computers without exponentially vast brains. Suppose the AI's hacking algorithm can hack 99% of all computer systems that are theoretically possible to hack with unlimited compute. And The AI takes at most 10% longer than the theoretically minimum time on those problems.  This AI is still clearly dangerous. Especially if it isn't just a hacking tool. It has a big picture view of the world and what it wants to achieve. In short, maybe P!=NP and there is no perfect algoritm, but its possible to be a lot better than current ML, and a lot better than humans, and you don't need a perfect algorithm to create an X risk.  If 3-sat is O(n^1000,000) then P=NP on a technicality, but the algorithm is totally useless as it is far too slow in practice. If its O(n^4) there are still some uses, but it would seriously hamper the effectiveness of the minimum circuit style predictors. Neural nets are trained on a lot of data. With an O(n^4) algorithm, training beyond 10kb of data will be getting difficult, depending somewhat on the constant.  If you consider the problem of persuading a human to let you out of the box in an AI boxing scenario, well that is perfectly well posed. (There is a big red "open box" button, and either its pressed or it isn't. ) But we don't have enough understanding of phycology to do it. Pick some disease we don't know the cause of or how to cure yet. There will be lots of semi relevant information in biology textbooks. There will be case reports and social media posts and a few confusing clinical trials. 

I think less than human intelligence is sufficient for an x-risk because that is probably what is sufficient for a takeoff.

If less than human intelligence is sufficient, wouldn't humans have already done it? (or are you saying we're doing it right now?)

How intelligent does an agent need to be to send a HTTP request to the URL /ldap://myfirstrootkit.comon a few million domains?)

A human could do this or write a bot to do this.(and they've tried)  But they'd also be detected, as would an AI.  I don't see this as an x-risk, so much as a manageable pr... (read more)

gwern120

If less than human intelligence is sufficient, wouldn't humans have already done it?

No. Humans are inherently incapable of countless things that software is capable of. To give an earlier example, humans can do things that evolution never could. And just as evolution can only accomplish things like 'going to the moon' by making agents that operate on the next level of capabilities, humans cannot do things like copy themselves billions of times or directly fuse their minds or be immortal or wave a hand to increase their brain size 100x. All of these are ... (read more)

I spent some time reading the Grinnblatt paper.  Thanks again for the link.  I stand corrected on IQ being uncorrelated with stock prediction.  One part did catch my eye.

Our findings relate to three strands of the literature. First, the IQ and trading behavior analysis builds on mounting evidence that individual investors exhibit wealth-reducing behavioral biases. Research, exemplified by Barber and Odean (2000, 2001, 2002), Grinblatt and Keloharju (2001), Rashes (2001), Campbell (2006), and Calvet, Campbell, and Sodini (2007, 2009a, 2009b),

... (read more)

We don't know that, P vs NP is an unproved conjecture. Most real world problems are not giant knapsack problems. And there are algorithms that quickly produce answers that are close to optimal. Actually, most of the real use of intelligence is not a complexity theory problem at all. "Is inventing transistors a O(n) or an O(2^n) problem?"

 

P vs. NP is unproven. But I disagree that "most real world problems are not giant knapsack problems". The Cook-Levin theorem showed that many of the most interesting problems are reducible to NP-complete problems. &nb... (read more)

2Donald Hobson
The smallest boolean circuit trick is a reasonable trick, if you can get it to work efficiently. But it won't magically be omniscient at everything. It would be just one fairly useful tool in the ML toolbox.  "Minimal circuit" based approaches will fail badly when the data comes from a simple to specify but computationally heavy function. For example, the minimal boolean circuit trained to take in a point on the complex plane (with X and Y as IEEE floats) and output if that point is in the mandelbrot set. This algorithm will fail reasonable numbers of training points.  If the 3 sat algorithm is O(n^4) then this algorithm might not be that useful compared to other approaches.  On the other hand, a loglinear algorithm that produced a circuit that was usually almost optimal, that could be useful.  What's the runtime complexity of rowhammering your own hardware to access the internet? O(1)? Meaningless. Asymptotic runtime complexity is a mathematical tool that assumes an infinite sequence of ever harder problems. (An infinite sequence of ever longer lists to sort, or 3-sat's to solve) There is not an infinite sequence of ever harder problems in reality. Computational complexity is mostly concerned with the worst case, and finding the absolutely perfect solution. In reality, an AI can use algorithms that find a pretty good solution most of the time.  Obviously there is a bound based on observation for every problem, just often that bound is crazy high on our scales. For a problem humans understand well, we may be operating close to that bound. For the many problems that are "illegible" or "hard for humans to think about" or "confusing", we are nowhere near the bound, so the AI has room to beat the pants off us with the same data.      Consider a world in which humans, being smart but not that smart, can figure out relativity from good clear physics data. Could a superintelligence figure out relativity based on the experiences of the typical caveman? You seem to b

AlphaGo went from mediocre, to going toe-to-toe with the top human Go players in a very short span of time. And now AlphaGo Zero has beaten AlphaGo 100-0. AlphaFold has arguably made a similar logistic jump in protein folding

Do you know how many additional resources this required? 

 

Cost of compute has been decreasing at exponential rate for decades, this has meant entire classes of algorithms which straightforward scale with compute also have become exponentially more capable, and this has already had profound impact on our world. At the very lea

... (read more)
1[comment deleted]

Since you bring up selection bias, Grinblatt et al 2012 studies the entire Finnish population with a population registry approach and finds that.

Thanks for the citation.  That is the kind of information I was hoping for.   Do you think that slightly better than human intelligence is sufficient to present an x-risk, or do you think it needs some sort of takeoff or acceleraton to present an x-risk?

So?

I think I can probably explain the "so" in my response to Donald below.

4gwern
I think less than human intelligence is sufficient for an x-risk because that is probably what is sufficient for a takeoff. (GPT-3 needed like 1k discrete GPUs to train. Nvidia alone ships something on the order of >73,000k discrete GPUs... per year. How fast exactly do you think returns diminish and how confident are you that there are precisely zero capability spikes anywhere in the human and superhuman regimes? How intelligent does an agent need to be to send a HTTP request to the URL /ldap://myfirstrootkit.com on a few million domains?)

Overshooting by 10x (or 1,000x or 1,000,000x) before hitting 1.5x is probably easier than it looks for someone who does not have background in AI.

Do you have any examples of 10x or 1000x overshoot?  Or maybe a reference on the subject?

2[comment deleted]

Hmmmmm there is a lot here let me see if I can narrow down on some key points.  

Once you have the right algorithm, it really is as simple as increasing some parameter or neuron count.

There are some problems that do not scale well(or at all).  For example, doubling the computational power applied to solving the knapsack problem will let you solve a problem size that is one element bigger.   Why should we presume that intelligence scales like an O(n) problem and not an O(2^n) problem? 

What is happening here? Are both people just looking a

... (read more)
3Donald Hobson
We don't know that, P vs NP is an unproved conjecture. Most real world problems are not giant knapsack problems. And there are algorithms that quickly produce answers that are close to optimal. Actually, most of the real use of intelligence is not a complexity theory problem at all. "Is inventing transistors a O(n) or an O(2^n) problem?" Meaningless. In practice, modest improvements in intelligence seem to produce largish differences in results.  Lots of big human achievements are built by many humans. Some things are made by one exceptionally smart human. Inventing general relativity, or solving Fermat's last theorem are the sort of problem with a smaller number of deep abstract general parts. Solving 10^12 capchas is exactly the sort of task that can be split among many people easily. This sort of problem can't be solved by a human genius (unless they make an AI, but that doesn't count) There are just too many separate small problems.  I think some tasks can easily be split into many small parts, some can't. Tasks that can be split, we get a team of many humans to do. Tasks that can't, if our smartest geniuses can't solve them, they go unsolved.  Suppose an AI as smart as the average human. You can of course put in 10x the compute, and get an AI as smart as 10 average humans working together. I think you can also put in 1.2x the compute, and get an AI as smart as Einstein. Or put in 12x the compute and get an AI as smart as 12 Einstein's working together.  But its generally better to put a lot of you compute into one big pile and get insights far more deep and profound than any human could ever make. Lots of distinct human mind sized blobs talking is an inefficient way to use your compute, any well designed AI can do better. Sure, maybe as many as 100's. Not millions. Now the differences between human brains are fairly small. This indicates that 2 humans aren't as good as 1 human with a brain that works slightly better.     Isn't that just relying on the int

Are we equivocating on 'much better' here?

Not equivocating but if intelligence is hard to scale and slightly better is not a threat, then there is no reason to be concerned about AI risk.  (maybe a 1% x-risk suggested by OP is in fact a 1e-9 x-risk)

there are considerable individual differences in weather forecasting performances (it's one of the more common topics to study in the forecasting literature),

I'd be interested in seeing any papers on individual differences in weather forecasting performance (even if IQ is not mentioned).  My understand... (read more)

9gwern
So? There's still a lot of human input. That was one of the criticisms of, IIRC, DeepMind's recent foray into DL weather modeling - "oh, but the humans still do better in the complex cells where the storms and other rare events happen, and those are what matter in practice". Where did you get that, Taleb? Untrue. Investing performance is correlated with IQ: better market timing, less inefficient trading, less behavioral bias, longer in the market, more accurate forecasts about inflation and returns, etc. Since you bring up selection bias, Grinblatt et al 2012 studies the entire Finnish population with a population registry approach and finds that.

I think I'm convinced that we can have human capable AI(or greater) in the next century(or sooner).  I'm unconvinced on a few aspects of AI alignment.  Maybe you could help clarify your thinking.

(1) I don't see how an human capable or a bit smarter than human capable AI(say 50% smarter) will be a serious threat.  Broadly humans are smart because of group and social behavior.  So a 1.5 Human AI might be roughly as smart as two humans?  Doesn't seem too concerning.

(2) I don't see how a bit smarter than humans scales to superhuman lev... (read more)

1[comment deleted]
gwern140

For example, a human with 150 IQ isn't going to be much better at predicting the weather than a person with 100 IQ.

Are we equivocating on 'much better' here? Because there are considerable individual differences in weather forecasting performances (it's one of the more common topics to study in the forecasting literature), and while off-hand I don't have any IQ cites, IQ shows up all the time in other forecasting topics as a major predictor of performance (as it is of course in general) and so I would be surprised if weather was much different.

5Donald Hobson
Humans have like 4 x the brain volume of chimps, but you can't put 4 chimps together to get 1 human. And you can't put 100 average humans together to get 1 Einstein. Despite Einstein having a brain with about the same size and genetics of any other human. This doesn't suggest an obvious model, but it does suggest your additive intelligence model is wrong. Either Or between: Once you have the right algorithm, it really is as simple as increasing some parameter or neuron count. The AI is smarter than a smart human. Smart humans can do AI research a bit. The AI does whatever a human doing AI research would do, but quicker and better.  I think this is somewhat to do with the underlying algorithm being not very good in a way that doesn't apply to all algorithms. It all depends on the scales you use anyway. It takes 10x as much energy to make sound 10db louder, because decibels are a logarithmic scale. Of course most problems can't be solved by reason alone. In a world with the internet, there is a lot of information available. The limits on how much useful information and understanding you can extract from your data are really high. So yes you need some data. This doesn't mean lack of data is going to be a serious bottleneck, any more than lack of silicon atoms is the bottleneck on making chips. The world is awash with data. Cleaned up, sorted out data is a little rarer. Taking a large amount of data, some of tangential relevance, some with mistakes in, some fraudulent, and figuring out what is going on, that takes intelligence. What is happening here? Are both people just looking at a picture and guessing numbers, or can the IQ 150 person program a simulation while the IQ 100 person is looking at the Navier stokes equation trying (and failing) to figure out what it means.  A lot of science seems to be done by a handful of top scientists. The likes of Einstein wouldn't be a thing if a higher intelligence didn't make you much better at discovering some things.  I th

Current AI does stochastic search, but it is still search. Essentially PP complexity class, instead of NP/P. (with a fair amount of domain specific heuristics)

4jessicata
Not all AI is search, and when it is it's usually not brute force search.

Never leave the house without your d20 :-P

But I agree with you. This seems a simple way to do something like satisficing. Avoiding the great computational cost of an optimal decision.

In terms of prior art that is probably the field you want to explore: https://en.m.wikipedia.org/wiki/Satisficing

Not sure if this is helpful, but since you analogized to chip design. In chip design, you typically verify using a constrained random method when the state space grows too large to verify every input exhaustively. That is, you construct a distribution over the set of plausible strings and then sample it and feed it to your design. Then you compare the result to a model in a higher level language.

Of course, standard techniques like designing for modularity can make the state space more manageable too.

First off, Scott’s blog is awesome.

Second, the example of dieting comes to mind when I think of training rationality. While they’re not much connected to the rationality community, they are a large group of people focused on overcoming one particular aspect of our irrationallity. (but without much success)

7lexande
Yeah, Lesswrong sometimes feels a bit like a forum for a fad diet that has a compelling story for why it might work and seemed to have greatly helped a few people anecdotally, so the forum filled up with people excited about it, but it doesn't seem to actually work for most of them. Yet they keep coming back because their friends are on the forum now and because they don't want to admit failure.

What basis is there to assume that the distribution of these variables is log uniform? Why, in the toy example, limit the variables to the interval [0,0.2]? Why not [0,1]?

These choices drive the result.

The problem is, for many of the probabilities, we don’t even know enough about them to say what distribution they might take. You can’t infer a meaningful distribution over variables where your sample size is 1 or 0

4Shmi
The [0, 0.2] was a toy example. They take a fair bit of care to evaluate the distribution of the real factors in the Drake formula, which is what most of the paper is about.

I’m still not seeing a big innovation here. I’m pretty sure most researchers who look at the Drake equation think “huge sensitivity to parameterization.”

If we have a 5 parameter drake equation then number of civilizations scales with X^5, so if X comes in at 0.01, we’ve got a 1e-10 probability of detectable civilization formation. But if we’ve got a 10 parameter Drake equation and X comes in at 0.01 then it implies a 1e-20 probability. (extraordinary smaller)

So yes, it has a a huge sensitivity, but it is primarily a constructed sensitivity. All the Drake equation really tells us is that we don’t know very much and it probably won’t be useful until we can get N above one for more of the parameters.

6gbear605
The difference is that before people looked at the Drake equation, and thought that even with the uncertainty, there was a very low probability of no aliens, and this corrects that assumption.

I’m not sure I understand why they’re against point estimates. As long as the points match the mean of our estimates for the variables, then the points multiplied should match the expected value of the distribution.

3roystgnr
Not quite. Expected value is linear but doesn't commute with multiplication. Since the Drake equation is pure multiplication then you could use point estimates of the means in log space and sum those to get the mean in log space of the result, but even then you'd *only* have the mean of the result, whereas what would really be a "paradox" is if P(N=1|N≠0) turned out to be tiny.

Because people draw incorrect conclusions from the point estimates. You can have high expected value of the distribution (e.g. "millions of civilizations") while at the same time having big part of the probability mass on outcomes with just one civilization, of few civilizations far away.

I think this is an interesting concept and want to see where you go with it. But just devil’s advocating, there are some pretty strong counterexamples for micromanagement. For example, many imperative languages can be ridiculously inefficient. And try solving an NP complete problem with a genetic algorithm and you’ll just get stuck in a local minimum.

Simplicity and emergence are often surprisingly effective but they’re just tools in a large toolbox.

Somewhat ironic that LW is badly in need of better captcha.

I read him, he is just incorrect. “People hate losses more than they hate gains” is not explained by DMU. They dislike losses to an extent far greater than predicted by DMU, and more importantly, this dislike is largely scale invariant.

If you go read papers like the original K&T, you’ll see that their data set is just a bunch of statements that are predicted to be equally preferrable under DMU (because marginal utility doesn’t change much for small changes in wealth). What changes the preference is simply whether K&T phrase the question in terms of a loss or a gain.

So...unsurprisingly, Kahneman is accurately describing the theory that won him the Nobel prize.

2Paperclip Minimizer
The author explain very clearly what the differences are between "people hate losses more than they like gains" and loss aversion. Loss aversion is people hating losing $1 while having $2 more than they like gaining $1 while having $1, even though it both case this the difference between having $1 and $2.

The result you got is pretty close to the fft of f(t) = t

Which is roughly what you got from sorting noise.

All finite length sequences exist in any infinite random sequence. So, in the same way that all the works of shakespeare exist inside an infinite random sequence, so too does a complete representation of any finite universe.

I suppose one could argue by the anthropic principle that we happen to exist in a well ordered finite subsequence of an infinite random sequence. But it is sort of like multiverse theories where it lacks the explanatory power or verifiability of simpler theories.

9Shmi
Yep. Fortunately, the sequences I played with are quite finite, 1024 samples, see the followup post. And I agree that musing about multiverses, while fun, has not been scientifically fruitful so far.

Maybe I’m being dense, and missing the mystery, but I think this reference might be helpful.

https://www.dsprelated.com/showarticle/40.php

9Shmi
Thanks for the link! Yep, I had a thought about the 1/f noise right after calculating the spectra above, especially because my original power spectrum slope calculation had an error and the result came out very close to that of the 1/f noise :) I had looked up if sorting can generate it, but nothing came up. Were I still in academia, I would have been tempted to write a speculative paper resulting in cringy headlines like "Scientists solve several long-standing mysteries at once: the arrow of time is driven by the human quest for order and meaning" The "mystery" as I see it currently, is the mathematical origin of the pink noise emerging from sorting.

I mean...he quotes Kahneman; claiming the guy doesn’t know the implications of his own theory.

Losses hurt more than gains even at scales where DMU predicts that they should not. (because your DMU curve is approximately flat for small losses and gains) Loss aversion is the psychological result which explains this effect.

This is the author’s conclusion: “So, please, don’t go around claiming that behavioral economists are incorporating some brilliant newfound insight that people hate losses more than they like gains. We’ve known about this in price theory since Alfred Marshall’s 1890 Principles of Economics.”

Sorry nope. Alfred Marhall’s Principles would have made the wrong prediction.

3Paperclip Minimizer
I don't think you read the author at all. The whole post is about structural qualitative differences between "people hate losses more than they like gains" (DMU) and loss aversion. He is not saying DMU explain loss aversion. He is not saying Alfred Marshall's Principles would have made the right prediction. What he is saying is that loss aversion is much less intuitive than the pop science version of loss aversion.

That makes a lot of sense to me. Aversion to small losses makes a ton of sense as a blanket rule, when the gamble is: lose: don’t eat today win: eat double today don’t play: eat today

Our ancestors probably faced this gamble since long before humans were even humans. Under those stable conditions, a heuristic accounting for scale would have been needlessly expensive.

In short, the author is wrong. Diminishing marginal utility only really applies when the stakes are on the order of the agent’s total wealth, whereas the loss aversion asymmetry holds true for relatively small sums.

1Paperclip Minimizer
I don't think you read the author right. He is not saying that loss aversion is explained by diminishing marginal utility, he's saying precisely the contrary.
gjm110

See e.g. a nice paper by Matthew Rabin which quantifies the extent to which diminshing marginal utility is too weak an effect to explain actually-observed risk aversion, by proving statements like this: "If you would turn down a 50:50 gamble between gaining $101 and losing $100 on account of diminishing marginal utility, then you would also turn down a 50:50 gamble between gaining all the money in the world and losing $10,000."

2Dagon
DMU is only rational when applied to the larger sums. It's pretty believable that much of what's called loss aversion is a broken heuristic in human brains, which mis-implements DMU by picking way-too-small reference classes. IMO, hyperbolic discounting is a related evolved heuristic which conflates value discounting and future uncertainty.