Maybe, in-vivo genetic editing of the brain is possible. Adenoviruses that are a normal delivery mechanism for genetic therapy can pass hemo-encephalic barrier, so seems plausible to an amateur.
(Not obvious that this works in adult organisms, maybe genes activate while fetus grows or during childhood.)
The post expands on the intuition of ML field that reinforcement learning doesn't always work and getting it to work is fiddly process.
In the final chapter, a DeepMind paper that argues that 'one weird trick' will work, is demolished.
The problem under consideration is very important for some possible futures of humanity.
However, author's eudamonic wishlist is self-admittedly geared for fiction production, and don't seem to be very enforceable.
It's a fine overview of modern language models. Idea of scaling all the skills at the same time is highlighted, different from human developmental psychology. Since publishing 500B-PaLM models seemed to have jumps at around 25% of the tasks of BIG-bench.
Inadequacy of measuring average performance on LLM is discussed, where a proportion is good, and rest is outright failure from human PoV. Scale seems to help with rate of success.
Argument against CEV seems cool, thanks for formulating it. I guess we are leaving some utility on the table with any particular approach.
Part on referring to a model to adjudicate itself seems really off. I have a hard time imagining a thing that has better performance at meta-level than on object-level. Do you have some concrete example?
Thanks for giving it a think.
Turning off is not a solved problem, e.g. https://www.lesswrong.com/posts/wxbMsGgdHEgZ65Zyi/stop-button-towards-a-causal-solution
Finite utility doesn't help, as long as you need to use probability. So you get, 95% chance of 1 unit of utility is worse than 99%, is worse than 99.9%, etc. And then you apply the same trick to probabilities you get a quantilizer. And that doesn't work either https://www.lesswrong.com/posts/ZjDh3BmbDrWJRckEb/quantilizer-optimizer-with-a-bounded-amount-of-output-1
Cool analysis. Sounds plausible.
So you're out to create a new benchmark? Reading SAT is referencing text in answers with ellipsis, making it hard for me to solve in single read-through. Maybe repeating questions in the beginning and expanding ellipses would fix that for humans. Probably current format is also confusing for pretrained models like GPT.
Requiring a longer task text doesn't seem essential. In the end, maybe, you'd like to take some curriculum learning experiment and thin out learning examples so that current memorization mechanisms wouldn't suf...
No particular philosophy: just add some kludge to make your life easier, then repeat until they blot out the Sun.
Non-computer tool is paper for notes & pen, filing everything useful to inbox during daily review. Everything else is based off org-mode, with Orgzly on mobile. Syncing over SFTP, not a cloud person.
Wrote an RSS reader in Python for filling inbox, along with org-capture. Wouldn't recommend the same approach, since elfeed should do the same reasonably easy. Having a script helps since running it automatically nightly + before daily revie...
I'd think 'ethical' in review board has noting to do with ethics. It's more of PR-vary review board. Limiting science to status-quo-bordering questions doesn't seem most efficient, but a reasonable safety precaution. However, typical view of the board might be skewed from real estimates of safety. For example, genetic modification of humans is probably minimally disruptive biological research (compared, to, say, biological weapons), though it is considered controversial.
My town ...
Let 20% wards be swung by one vote, that gives each voter 1 in (5 * amount of voters) chance of affecting a vote cast on the next level, if that's how US system works?
... elected officials change their behavior based on margins ...
Which is an exercise in reinforcing prior beliefs, since margins are obviously insufficient data.
Politicians pay a lot more attention to vote-giving populations...
Are politicians equipped with a device to detect voters and their needs? If not, then it's lobbying, not voting that matters.
......impact of your r
Absolutely, shutting up and multiplying is the right thing to do.
Assume: simple majority vote, 1001 voters, 1 000 000 QALY at stake, votes binomially distributed B(p=0.4), no messing with other people's votes, voting itself doesn't give you QALY.
My vote swings iff 500 <= B(1001, 0.4) < 501, with probability 5.16e-11, it is advised if takes less than 27 minutes.
Realistically, usefulness of voting is far less, due to:
actual populations are huge, and with them chance of swing-voting falls;
QALY are not quite utils (eg. other's QALY counts the same wa
I don't care, because there's nothing I can do about it. It also applies to all large-scale problems, like national elections.
I do understand, that that point of view creates 'tragedy of commons', but there's no way I can force millions of people to do my bidding on this or that.
I also do not make interventions to my lifestyle, since I expect AGW effects to be dominated by socio-economic changes in the nearest half a century.
Are artificial neural networks really Turing-complete? Yep, they are [Siegelman, Sontag 91]. Amount of neurons in the paper is , with rational edge weights, so it's really Kolmogorov-complex. This, however, doesn't say if we can build good machines for specific purposes.
Let's figure out how to sort a dozen numbers with -calculus and sorting networks. It must stand to notice, that lambda-expression is O(1), whereas sorter network is O(n (log n)^2) in size.
Batcher's odd–even mergesort would be O(log n) levels deep, and given one neuron is used to implement ...
Well, it seems somewhat unfair to judge the decision on information not available for decision-maker, however, I fail to see how is that an 'implicit premise'.
I didn't think Geneva convention was that old, and, actually updating on it makes Immerwahr decision score worse, due to lower expected amount of saved lives (through lower chance of having chemical weapons used).
Hopefully, roleplaying this update made me understand that in some value systems it's worth it. Most likely, E(\Delta victims to Haber's war efforts) > 1.
Here's what I meant by saying you were begging the question: you were assuming the outcome (few people would be killed by chemical warfare after WW1) did not depend on the protests against chemical weapons.
You said originally that protesting against chemical warfare (CW) during WW1 was not worth the sacrifice involved, because few people were killed by CW after WW1.
But the reason few people were killed is that CW was not used often. And one contributing factor to its not being used was that people had protested its use in WW1, and created the Geneva Conven...
Standing against unintended pandemics, atomic warfare and other extinction threatenting events have been quite good of an idea in retrospect. Those of us working of scientific advances shall indeed ponder the consequences.
But Immerwahr-Haber episode is just an unrelated tearjerker. Really, inventing process for creation of nitrogen fertilizers is so more useful than shooting oneself in the heart. Also, chemical warfare turned out not to kill much people since WWI, so such sacrifice is rather irrelevant.
Unless there are on order of $2^KolmogorovComplexity(Universe)$ universes, the chance of it being constructed randomly is exceedingly low.
An extremely low probability of the observation under some theory is not itself evidence. It's extremely unlikely that the I would randomly come up with the number 0.0135814709894468, and yet I did.
It's only interesting if there is some other possibility that assigns a different probability to that outcome.
Things that I seem to notice about the plan: