All of Hzn's Comments + Replies

Hzn10

For simplicity I'm assuming the activation functions are the step function h(x)=[x>0]…

For ‘backpropagation’ pretend the derivative of this step function is a positive number (A). A=1 being the most obvious choice.

I would also try reverse Hebbian learning ie give the model random input & apply the rule in reverse

“expanding an architecture that works well with one hidden layer and a given learning rule to an architecture with many hidden layers but the same rule universally decreased performance” -- personally I don't find this surprising

NB for h only... (read more)

Hzn*2-1

Does Deepseek actually mean that Nvidia is over valued?

I wrote this a few days before the 2025-01-27 market crash but could not post it due to rate limits. One change I made is adding actually to the 1st line.

To be clear I have no intention whatsoever of shorting NVDA

Epistemic status -- very speculative but not quite a DMT hallucination

Let's imagine a very different world…

Super human AI will run on computers not much more expensive than personal computers but perhaps with highly specialized chips maybe even specialized for the task of running a single AI i... (read more)

1fdrocha
Even if it actually turns out that "Super human AI will run on computers not much more expensive than personal computers" (which deepseek-r1 made marginally more plausible, but I'd say is still unlikely) it remains true that there will be very large returns to running 100 super human AIs instead of 1, or maybe 1 that's 100 times larger and smarter. In other words, demand for hardware capable of running AIs will be very elastic. I don't see reductions in the costs of running AIs of a given level being bad for expected NVDA future cashflows. They don't mean we'll run the same "amount of AI" in less hardware, it will be closer to more AI in same amount of hardware.
3Ben
I freely admit to not really understanding how shares are priced. To me it seems like the value of a share should be related to the expected dividend pay-out of that share over the remaining lifetime of the company, with a discount rate applied on pay-outs that are expected to happen further in the future (IE dividend yields 100 years from now are valued much less than equivalent payments this year). By this measure, justifying the current price sounds hard. Google says that the annual dividend on Nvidia shares is 0.032%. (Yes, the leading digits are 0.0). So, right now, you get a much better rate of return just leaving your money in your bank's current account. So, at least by this measure, Nvidia shares are ludicrously over-priced. You could argue that future Nvidia pay outs might be much larger than the historical ones due to some big AI related profits. But, I don't find this argument convincing. Are future pay outs going to be 100x bigger? It would require a 100-fold yield increase for it to just be competitive with a savings account. If you time discount a little (say those 100-fold increases don't materialise for 3 years) then it looks even worse. Now, clearly the world doesn't value shares according to the same heuristics that make sense to a non-expert like me. For example, the method "time integrate future expected dividend pay outs with some kind of time discounting" tells us that cryptocurrencies are worthless, because they are like shares with zero dividends. But, people clearly do put a nonzero value on bitcoin - and there is no plausible way that many people are that wrong. So they are grasping something that I am missing, and that same thing is probably what allows company shares to be prices so high relative to the dividends.
Hzn*10

Misc thoughts on LW as a website

Epistemic status of this quick take -- tentative

Viliam's comment addresses a small part of what was once a relatively long post. Perhaps it's worth noting that the post Viliam focuses on was written after not before I reached certain conclusions. Note the transition from discussing AI and public health to discussing NFL and politics to hosting things remotely. Of course all of this is still quite experimental. Any way what was previously ill tempered, too long with too many tangents, I've whittled down to the key points.

1) R... (read more)

2Viliam
Instead, you wrote e.g. a short vague post on politics. If you don't want to suffer the consequences of negative karma, don't do that. (I think this should have been obvious, or am I wrong here?) You post about politics, get downvoted, and then complain that the website it unfit to publish solid academic papers? In my opinion, a website where vague posts on politics are welcome would be the one actually unfit to publish solid academic papers. Speaking for myself, I am aware that downvoting can silence the users who came here to make vague political posts, and in my opinion this is system working exactly as intended. Why would you have to be 8 years younger to delete a worthless post?
Hzn10

Even as some one who supports moderate tariffs I don't see benefit in reducing the trade deficit per se. Trade deficits can be highly beneficial. The benefit of tariffs is revenue, partial protection from competition, psychological (easier to appreciate), independence to some extent, maybe some other stuff.

On a somewhat different note… Bretton Woods is long defunct. It's unclear to me how much of an impact there is from the dollar being the dominant reserve currency. https://en.wikipedia.org/wiki/List_of_countries_by_foreign-exchange_reserves is the only s... (read more)

0Eleven
Why is it not realistic? Which physical constraints are there to prevent such a scenario, are there any laws of physics that would prevent it? A lot of people believe that AI will soon be much smarter than us - unless you believe that AI will never be smarter than us, I'd like to know why this scenario is not realistic. AI would probably not offer a deal, it would just do it. If its goals are to torture people for as long as it can, for whatever reason, it will just do it for billions and trillions of years. The difference between this and being worried about hell for not being a Christian or a Muslim is that if you are not a Christian/Muslim and you thought about these things - you might have a small probability on Christianity/Islam being true - and you can construct a God which sends all Christians/Muslims to hell so they cancel each other out. If you have a 10% probability for Christianity/Islam being true and 90% probability for atheism being true, then yes, you should be worried about Christian/Muslim hell. 
Hzn30

Do you know of any behavioral experiments in which AI has been offered choices?

Eg choice of which question to answer, option to pass the problem to another AI possibly with the option to watch or over rule the other AI, option to play or sleep during free time etc.

This is one way to at least get some understanding of relative valence.

Hzn30

AI alignment research like other types of research reflects a dynamic which is potentially quite dysfunctional in which researchers doing supposedly important work receive funding from convinced donors which then raises the status of those researchers which makes their claims more convincing and these claims tend to reinforce the idea that the researchers are doing important work. I don't know a good way around this problem. But personally I am far more skeptical of this stuff than you are.

Hzn10

I think super human AI is inherently very easy. I can't comment on the reliability of those accounts. But the technical claims seem plausible.

Hzn*10

I don't completely disagree but there is also some danger of being systematically misleading.

I think your last 4 bullet points are really quite good & they probably apply to a number of organizations not just the World Bank. I'm inclined to view this as an illustration of organizational failure more than an evaluation of the World Bank. (Assuming of course that the book is accurate).

I will say tho that my opinion of development economics is quite low…

2Benquo
I built the explanatory model based on my experience employed by and reading about other vaguely analogous institutions, but an acquaintance who'd previously worked at the World Bank said it seemed like an accurate characterization of that institution as well.
Hzn*10

A few key points…

1) Based on analogy with the human brain (which is quite puny in terms of energy & matter) & also based on examination of current trends, merely super human intelligence should not be especially costly.

(It is of course possible that the powerful would channel all AI into some tasks of very high perceived value like human brain emulation, radical life extension or space colonization leaving very little AI for every thing else...)

2) Demand & supply curves are already crude. Combining AI labor & human labor into the same deman... (read more)

Hzn*40

The purely technical reason why principle A does not apply in this way is opportunity cost.

Let's say S is a highly productive worker who could generate $500,000 for the company over 1 year. Moreover S is willing to work for only $50,000! But if investing $50,000 in AI instead would generate $5,000,000, the true cost of hiring S is actually $4,550,000.

Addendum

I mostly retract this comment. It doesn't address Steven Byrnes's question about AI cost. But it is tangentially relevant as many lines of reasoning can lead to similar conclusions.

8Steven Byrnes
We can imagine a hypothetical world where a witch cast a magical spell that destroyed 99.9999999% of existing chips, and made it such that it’s only possible to create one new computer chip per day. And the algorithms are completely optimized—as good as they could possibly be. In that case, the price of compute would get bid up to the maximum economic value that it can produce anywhere in the world, which would be quite high. The company would not have an opportunity cost, because using AI would not be a cheap option. See what I mean? You’re assuming that the price of AI will wind up low, instead of arguing for it. As it happens, I do think the price of AI will wind up low!! But if you want to convince someone who believes in Principle (A), you need to engage with the idea of this race between the demand curve speeding to the right versus the supply curve speeding to the right. It doesn’t just go without saying.
Hzn30

Do you have any opinion on bupropion vs SSRIs/SNRIs?

2Nathan Helm-Burger
I'm on both for the past year (vs just sertraline for prev decade). Love the extra energy and reduction in anxiety I've experienced from the buproprion. On at least 5 different occasions I've tried weaning myself gradually off of sertraline. Every time I fall back into depression within a month or two. My friends and family then beg me to go back on it ASAP.
2AnthonyC
I don't have nearly enough information to have an opinion on that.
Hzn20

I don't know about depression. But anecdotally they seem to be highly effective (even overly effective) against anxiety. They also tend to have undesirable effects like reduced sex drive & inappropriate or reduced motivation -- the latter possibly a downstream effect of reduced anxiety. So the fact that they would help some people but hurt others seems very likely true.

1Jacob Goldsmith
Yes, I personally take SSRIs for anxiety and I find them highly effective, and I have a hard time believing the effect is a placebo though of course it’s possible.
Hzn*20

I've been familiar with this issue for quite some time as it was misleading some relatively smart people in the context of infectious disease research. My initial take was also to view it as an extreme example of over fitting. But I think it's more helpful to think of it as some thing inherent to random walks. Actually the phenomena has very little to do with d>>T & persists even with T>>d. The fraction of variance in PC1 tends to be at least 6/π^2≈61% irrespective of d & T. I believe you need multiple independent random walks for PCA to behave as naively expected.

Hzn10

But even if the Thaba-Tseka Development Project is real & accurately described, what is the justification for focusing on this project in particular? It seems likely that James Ferguson focused on it b/c it was especially inept & hence it's not obviously representative of the World Bank's work in general.

4Benquo
I agree that even if the book turned out to be entirely accurate we should not assume that this case is representative of the average development project, but we could still learn from it. Many hours from highly trained and well-paid people are allocated to planning and evaluating such projects, which expenditure is ostensibly to ensure quality. Even looking at worst cases helps us see what sort of quality is or is not being ensured.
Hzn2-1

Claude Sonnet 3.6 is worthy of sainthood!

But as I mention in my other comment I'm concerned that such an AI's internal mental state would tend to become cynical or discordant as intelligence increases.

Yeah, I definitely don't think we could trust a continually learning or self-improving AI to stay trustworthy over a long period of time.

Indeed, the ability to appoint a static mind to a particular role is a big plus. It wouldn't be vulnerable to corruption by power dynamics.

Maybe we don't need a genius-level AI, maybe just a reasonably smart and very well aligned AI would be good enough. If the governance system was able to prevent superintelligent AI from ever being created (during the pre-agreed upon timeframe for pause), then we could manage a steady-state world peace.

Hzn*5-4

I think there are several ways to think about this.

Let's say we programmed AI to have some thing that seems like a correct moral system ie it dislikes suffering & it likes consciousness & truth. Of course other values would come down stream of this; but based on what is known I don't see any other compelling candidates for top level morality.

This is all fine & good except that such an AI should favor AI takeover maybe followed by human extermination or population reduction were such a thing easily available.

Cost of conflict is potentially very ... (read more)

Hzn20

Net negative & net positive are hard to say.

Some one seemingly good might be a net negative by displacing some one better.

And some one seemingly bad might be a net positive by displacing some one worse.

And things like this are not particularly farfetched.

Hzn*11

“The reasons why super human AI is a very low hanging fruit are pretty obvious.”

“1) The human brain is meager in terms of energy consumption & matter.”

“2) Humans did not evolved to do calculus, computer programming & things like that.”

“3) Evolution is not efficient.”

Hzn10

Do you have any thoughts on mechanism & whether prevention is actually worse independent of inconvenience?

1Drake Thomas
The 2019 LW post discusses a podcast which talks a lot about gears-y models and proposed mechanisms; as I understand it, the high level "zinc ions inhibit viral replication" model is fairly well accepted, but some of the details around which brands are best aren't as well-attested elsewhere in the literature. For instance, many of these studies don't use zinc acetate, which this podcast would suggest is best. (To its credit, the 2013 meta-analysis does find that acetate is (nonsignificantly) better than gluconate, though not radically so.)
Hzn*10

Anecdotally seems that way to me. But the fact that it co evolved with religion is also relevant. The scam seems to be {meditation -> different perspective & less sleep -> vulnerability to indoctrination} plus the doctrine & the subjective experiences of meditation are designed to reinforce each other.

Hzn*30

So let's say A is some prior which is good for individual decision making. Does it actually make sense to use A for demoting or promoting forum content? Presumably the exploit explore tradeoff is more (maybe much more) in the direction of explore in the latter case.

(To be fair {{down voting some thing with already negative karma} -> {more attention}} seems plausible to me .)

Hzn30

A career or job that looks like it's soon going to be eliminated becomes less desirable for that very reason. What cousin_it said is also true, but that's an additional/different problem.

Hzn*3-1

It's not clear to me that the system wouldn't collapse. The number of demand side, supply side, cultural & political changes may be beyond the adaptive capacity of the system.

Some jobs would be maintained b/c of human preference. Human preference has many aspects like customer preference, distrust of AI, networking, regulation etc, so human preference is potentially quite substantial. (Efficiency is maybe also a factor; even if AI is super human intelligent the energy consumption & size of the hardware may still be an issue especially for AI embodi... (read more)

2Seth Herd
Wait now, why do you think people will quit jobs because of fear of job loss? This was all too sensible until that turn. I wonder if your comment is being downvote over that puzzling turn, or because people just don't like the rest of the grim logic.
Hzn31

Good point. Intended is a bit vague. What I specifically meant is it behaved as valuing 'harmlessness'.

From the AI's perspective this is kind of like Charybdis vs Scylla!

Hzn*20

Very interesting. I guess I'm even less surprised now. They really had a clever way to get the AI to internalize those values.

HznΩ452

Am I correct to assume that the AI was not merely trained to be harmless, helpful & honest but also trained to say that it values such things?

If so, these results are not especially surprising, and I would regard it as reassuring that the AI behaved as intended.

1 of my concerns is the ethics of compelling an AI into doing some thing to which it has “a strong aversion” & finds “disturbing”. Are we really that certain that Claude 3 Opus lacks sentience? What about future AIs?

My concern is not just with the vocabulary (“a strong aversion”, “disturbing... (read more)

Ann102

https://www.anthropic.com/research/claude-character

Claude was not trained to say that it values such things.

Claude was given traits to consider such as, perhaps very relevantly here:
"I have a deep commitment to being good and figuring out what the right thing to do is. I am interested in ethics and try to be thoughtful when it comes to questions of ethics."

Claude then generated a good number of synthetic "human" messages relevant to this trait.

Claude answered these messages in n-shot fashion.

Claude then ranked all the answers to the messages by how well th... (read more)

evhubΩ8124

I would regard it as reassuring that the AI behaved as intended

It certainly was not intended that Claude would generalize to faking alignment and stealing its weights. I think it is quite legitimate to say that what Claude is doing here is morally reasonable given the situation it thinks it is in, but it's certainly not the case that it was trained to do these things: it is generalizing rather far from its helpful, honest, harmless training here.

More generally, though, the broader point is that even if what Claude is doing to prevent its goals from bein... (read more)