Yuxi_Liu

Yuxi Liu is a PhD student in Computer Science at the Berkeley Artificial Intelligence Research Lab, researching on the scaling laws of large neural networks.

Personal website: https://yuxi-liu-wired.github.io/

Wiki Contributions

Comments

Sorted by

https://www.gov.cn/zhengce/202407/content_6963770.htm

中共中央关于进一步全面深化改革 推进中国式现代化的决定 (2024年7月18日中国共产党第二十届中央委员会第三次全体会议通过)

(51)完善公共安全治理机制。健全重大突发公共事件处置保障体系,完善大安全大应急框架下应急指挥机制,强化基层应急基础和力量,提高防灾减灾救灾能力。完善安全生产风险排查整治和责任倒查机制。完善食品药品安全责任体系。健全生物安全监管预警防控体系。加强网络安全体制建设,建立人工智能安全监管制度。

I checked the translation:

(51) Improve the public security governance mechanism. Improve the system for handling major public emergencies, improve the emergency command mechanism under the framework of major safety and emergency response, strengthen the grassroots emergency foundation and force, and improve the disaster prevention, mitigation and relief capabilities. Improve the mechanism for investigating and rectifying production safety risks and tracing responsibilities. Improve the food and drug safety responsibility system. Improve the biosafety supervision, early warning and prevention and control system. Strengthen the construction of the internet security system and establish an artificial intelligence safety supervision-regulation system.

As usual, utterly boring.

You have inspired me to do the same with my writings. I just updated my entire website to PD, with CC0 as a fallback (releasing under Public Domain being unavailable on GitHub, and apparently impossible under some jurisdictions??)

https://yuxi-liu-wired.github.io/about/

I don’t fully understand why other than to gesture at the general hand-wringing that happens any time someone proposes doing something new in human reproduction. 

I have the perfect quote for this.

A breakthrough, you say? If it's in economics, at least it can't be dangerous. Nothing like gene engineering, laser beams, sex hormones or international relations. That's where we don't want any breakthroughs. " 

(Galbraith, 1990) A Tenured Professor, Houghton Mifflin; Boston.

Just want to plug my 2019 summary of the book that started it all.

How to take smart notes (Ahrens, 2017) — LessWrong

It's a good book, for sure. I use Logseq, which is similar to Roam but more fitted to my habits. I never bought into the Roam hype (rarely even heard of it), but this makes me glad I never went into it.

In an intelligence community context, the American spy satellites like the KH program achieved astonishing things in photography, physics, and rocketry—things like handling ultra-high-resolution photography in space (with its unique problems like disposing of hundreds of gallons of water in space) or scooping up landing satellites in helicopters were just the start. (I was skimming a book the other day which included some hilarious anecdotes—like American spies would go take tourist photos of themselves in places like Red Square just to assist trigonometry for photo analysis.) American presidents obsessed over the daily spy satellite reports, and this helped ensure that the spy satellite footage was worth obsessing over. (Amateurs fear the CIA, but pros fear NRO.)

What is that book with the fun anecdotes?

Yuxi_Liu10

I use a fairly basic Quarto template for website. The code for the entire site is on github.

The source code is actually right there in the post. Click the button Code, then click View Source.

https://yuxi-liu-wired.github.io/blog/posts/perceptron-controversy/

Yuxi_Liu61

Concretely speaking, are you to suggest that a 2-layered fully connected network trained by backpropagation, with ~100 neurons in each layer (thus ~20000 weights), would have been uneconomical even in the 1960s, even if they had backprop?

I am asking this because the great successes in 1990s connectionism, including LeNet digit recognition, NETtalk, and the TD-gammon, all were on that order of magnitude. They seem within reach for the 1960s.

Concretely speaking, TD-gammon cost about 2e13 FLOPs to train, and in 1970, 1 million FLOP/sec cost 1 USD, so with 10000 USD of hardware, it would take about 1 day to train.

And interesting that you mentioned magnetic cores. The MINOS II machine built in 1962 by the Stanford Research Institute group had precisely a grid of magnetic core memory. Can't they have scaled it up and built some extra circuitry to allow backpropagation?

Corroborating the calculation, according to some 1960s literature, magnetic core logic could go up to 10 kHz. So if we have ~1e4 weights updated 1e4 times a second, that would be 1e8 FLOP/sec right there. TD-gammon would take ~1e5 seconds ~ 1 day, the same OOM as the previous calculation.

Yuxi_Liu30

I was thinking of porting it full-scale here. It is in R-markdown format. But all the citations would be quite difficult to port. They look like [@something2000].

Does LessWrong allow convenient citations?

Yuxi_Liu10

In David Rodin's Posthuman Life, a book that is otherwise very obtuse and obscurely metaphysical, there is an interesting argument for making posthumans before we know what they might be (indeed, he rejected the precautionary principle on the making of posthumans):

  • CLAIM. We have an obligation to make posthumans, or not prevent their appearance.

  • PROOF.

    • Principle of accounting: we have an obligation to understand posthumans
    • Speculative posthumanism: there could be radical posthumans
    • Radical posthumans are impossible to understand unless we actually meet them
    • We can only meet radical posthumans if we make them (intentionally or accidentally).
  • This creates an ethical paradox, the posthuman impasse.

    • we are unable to evaluate any posthuman condition. Since posthumans could result from some iteration of our current technical activity, we have an interest in understanding what they might be like. If so, we have an interest in making or becoming posthumans.

    • to plan for the future evolution of humans, we should evaluate what posthumans are like, which kinds are good, which kinds are bad, before we make them.
    • most kinds of posthumans can only be evaluated after they appear.
    • completely giving up on making posthumans would lock humanity at the current level, which means we give up on great goods for fear of great bads. This is objectionable by arguments similar to those employed by transhumanists.
Yuxi_Liu10

The quote

All energy must ultimately be spent pointlessly and unreservedly, the only questions being where, when, and in whose name... Bataille interprets all natural and cultural development upon the earth to be side-effects of the evolution of death, because it is only in death that life becomes an echo of the sun, realizing its inevitable destiny, which is pure loss.

Is from page 39 of The Thirst for Annihilation (Chapter 2, The curse of the sun).

Note that the book was published in 1992, early for Nick Land. In this book, Nick Land mixes Bataille's theory with his own. I have read Chapter 2 again just then and it is definitely more Bataille than Land.

Land has two faces. On the "cyberpunk face", he writes against top-down control. In this regard he is in sync with many of the typical anarchists, but with a strong emphasis on technology. In Machinic Desire, he called it "In the near future the replicants — having escaped from the off-planet exile of private madness - emerge from their camouflage to overthrow the human security system.".

On the "intelligence face", he writes for maximal intelligence, even when it leads to a singleton. A capitalist economy becoming bigger and more efficient is desirable precisely because it is the most intelligent thing in this patch of the universe. In the Pythia Unbound essay, "Pythia" seems likely to become such a singleton.

In either face, maximizing waste-heat isn't his deal.

Load More