All of dkirmani's Comments + Replies

an all-knowing God could predict which humans would sin and which would not

And how would God predict (with perfect fidelity) what humans would do without simulating them flawlessly? A truly flawless physical simulation has no less moral weight than "reality" -- indeed, the religious argument could very well be that our world exists as a figment of this God's imagination.

2[comment deleted]

systems that have a tendency to evolve towards a narrow target configuration set when started from any point within a broader basin of attraction, and continue to do so despite perturbations.


When determining whether a system "optimizes" in practice, the heavy lifting is done by the degree to which the set of states that the system evolves toward -- the suspected "target set" -- feels like it forms a natural class to the observer.

The issue here is that what the observer considers "a natural class" is informed by the data-distribution that the observer has previously been exposed to.

6Alex Flint
It's worse, even, in a certain way, than that: the existence of optimizing systems organized around a certain idea of "natural class" feeds back into more observers observing data that is distributed according to this idea of "natural class", leading to more optimizing systems being built around that idea of "natural class", and so on. Once a certain idea of "natural class" gains a foothold somewhere, observers will make real changes in the world that further suggest this particular idea of "natural class" to others, and this forms a feedback loop.

Whether or not an axis is "useful" depends on your utility function.

If you only care about compressing certain books from The Library of Babel, then "general optimality" is real — but if you value them all equally, then "general optimality" is fake.

When real, the meaning of "general optimality" depends on which books you deem worthy of consideration.

Within the scope of an analysis whose consideration is restricted to the cluster of sequences typical to the Internet, the term "general optimality" may be usefully applied to a predictive model. Such analysis ... (read more)

3Veedrac
Which is equivalent to saying if you only care about a situation where none of your observations correlate with any of your other observations and none of your actions interact with any of your observations then your observations are valueless. Which is a true but empty statement, and doesn't meaningfully affect whether there is an optimality axis that it's possible to be better on.

Yeah. Here's an excerpt from Antifragile by Taleb:

One can make a list of medications that came Black Swan–style from serendipity and compare it to the list of medications that came from design. I was about to embark on such a list until I realized that the notable exceptions, that is, drugs that were discovered in a teleological manner, are too few—mostly AZT, AIDS drugs.

Backpropagation designed it to be good on mostly-randomly selected texts, and for that it bequeathed a small sliver of general optimality.

"General optimality" is a fake concept; there is no compressor that reduces the filesize of every book in The Library of Babel.

3Veedrac
There is a useful generality axis and a useful optimality axis and you can meaningfully progress along both at the same time. If you think no free lunch theorems disprove this then you are confused about no free lunch theorems.

This was kinda a "holy shit" moment

Publicly noting that I had a similar moment recently; perhaps we listened to the same podcast.

For what it's worth, he has shared (confidential) AI predictions with me, and I was impressed by just how well he nailed (certain unspecified things) in advance—both in absolute terms & relative to the impression one gets by following him on twitter.

I resent the implication that I need to "read the literature" or "do my homework" before I can meaningfully contribute to a problem of this sort.

The title of my post is "how 2 tell if ur input is out of distribution given only model weights". That is, given just the model, how can you tell which inputs the model "expects" more? I don't think any of the resources you refer to are particularly helpful there.

Your paper list consists of six arXiv papers (1, 2, 3, 4, 5, 6).

Paper 1 requires you to bring a dataset.

We propose leveraging [diverse image and text]

... (read more)
Answer by dkirmani1812

Obtaining an Adderall prescription.

I use Done, and can recommend messaging their support to switch you to RxOutreach (a service that mails you your medication) if you live in an area with Adderall shortages, like, say, the Bay Area.

I recently met several YC founders and OpenAI enterprise clients — a salient theme was the use of LLMs to ease the crushing burden of various forms of regulatory compliance.

Does “Kickstarter but they refund you if the project doesn’t meet its goal” exist? I think not — that might get you some of the gains of a DAC platform.

Also, what does PayPal’s TOS say about this? I seriously considered building something similar, and Stripe only approves “crowdfunding” if you have their written approval.

[This comment is no longer endorsed by its author]Reply
7jefftk
That's just Kickstarter. You already only pay if the project hits its funding goal.
1Mazianni
It expand on what dkirmani said Or to quote dkirmani

If Guyenet is right that olfactory/taste signals are critical to the maintenance of obesity, then we should expect people who take their meals exclusively through feeding tubes to be obese at rates well below baseline.

I’m also somewhat confused about why this works

Abstraction is about what information you throw away. For a ReLU activation function, all negative inputs are mapped to zero -- you lose information there, in a way that you don't when applying a linear transformation.

Imagine your model (or a submodule thereof) as a mapping from one vector space to another. In order to focus on the features relevant to the questions you care about (is the image a truck, is it a lizard, ...) you throw away information that is not relevant to these questions -- you give it le... (read more)

4Chris_Leong
Ok, that’s fascinating! Thanks for the explanation.

Is there idea that pictures with texta drawn over them are out of distribution?

Yes, the idea is that images that have been taken with a camera were present in the training set, whereas images that were taken with a camera and then scribbled on in GIMP were not.

If you refer to section 4.2 in the paper that leogao linked, those authors also use "corrupted input detection" to benchmark their method. You're also welcome to try it on your own images -- to run the code you just have to install the pip dependencies and then use paths to your own files. (If you uncomment the block at the bottom, you can run it off your webcam in real-time!)

We introduced the concept of the space of models in terms of optimization and motivated the utility of gradients as a distance measure in the space of model that corresponds to the required amount of adjustment to model parameters to properly represent given inputs.

Looks kinda similar, I guess. But their methods require you to know what the labels are, they require you to do backprop, they require you to know the loss function of your model, and it looks like their methods wouldn't work on arbitrarily-specified submodules of a given model, only the mode... (read more)

Yep, but:

given only model weights

Also, the cossim-based approach should work on arbitrary submodules of a given model! Also it's fast!

often rely on skills that aren't generally included in "intelligence", like how fast and precise you can move your fingers

That's a funny example considering that (negative one times a type of) reaction time is correlated with measures of g-factor at about .

This seems an important point. I have a measured IQ of around 145 (or at least as last measured maybe 15 years ago when I was in my 20s). My reaction times are also unusually slow. Some IQ tests are timed. My score would come in a full 15 points lower (one standard deviation) on timed tests.

You might complain this is just an artifact of the testing protocol, but I think there's something real there. In everyday life I'm a lot smarter (e.g. come up with better ideas) when I can sit and think for a while. When I have to "think on my feet" I'm considerably du... (read more)

4Dagon
To paraphrase Douglas Adams, I object partly because it is a debasement of open discussion, but mostly because I don’t get invited to those sorts of parties.

This comment reminded me: I get a lot of value from Twitter DMs and groupchats. More value than I get from the actual feed, in fact, which -- according to my revealed preferences -- is worth multiple hours per day. Groupchats on LessWrong have promise.

5Raemon
Note LessWrong has group chat – it's in the conversation options button after you start a chat with one person.
2Dagon
Why would LW need a group-chat (or more developed DM) function?  If you want non-public conversations with select LW members, can't you do that today, on twitter, discord, slack, or e-mail?

This advice is probably good in most social contexts -- but I really appreciate the rationalist norm of taking deception very seriously. I resolve this class of conflict by being much more apprehensive than usual about casually misleading my rationalist-adjacent friends and business associates.

6M. Y. Zuo
I would go even farther, it's fundamentally bad advice for anyone that plans to interact with anyone, or any group, that is actually sophisticated.  This kind of pretending wouldn't even fool the modern day equivalents of von Neumann's secretary.

From Daniel Ingram's Mastering the Core Teachings of the Buddha (slatestarcodex review):

Immediately after a physical sensation arises and passes is a discrete pulse of reality that is the mental knowing of that physical sensation, here referred to as “mental consciousness” (as contrasted with the problematic concept of “awareness” in Part Five). By physical sensations I mean the five senses of seeing, hearing, smelling, tasting, and touching, and I guess you could add some proprioceptive, other extended sensate abilities and perhaps a few others, but for

... (read more)
2Jim Pivarski
Wow! That looks like a great book. Although one can find out by following the links you provided, I'd like to tell everyone here that the book is available for free on the author's website (PDF, epub, mobi).
2vandemonian
I mean I basically have to make a decision market, considering my project!

Advertisements on Lesswrong (like lsusr's now-deleted "Want To Hire Me?" post) are good, because they let the users of this site conduct mutually-beneficial trade.

I disagree with Ben Pace in the sibling comment; advertisements should be top-level posts, because any other kind of post won't get many eyeballs on it. If users don't find the advertised proposition useful, if the post is deceptive or annoying, then they should simply downvote the ad.

you can get these principles in other ways

I got them via cultural immersion. I just lurked here for several months while my brain adapted to how the people here think. Lurk moar!

I noticed this happening with goose.ai's API as well, using the gpt-neox model, which suggests that the cause of the nondeterminism isn't unique to OpenAI's setup.

The SAT switched from a 2400-point scale back to a 1600-point scale in 2016.

2Screwtape
That I feel a bit embarrassed for missing. Thank you for pointing it out; since the question asks for which range the respondent had hopefully everything got answered right. I updated the description.

Lesswrong is a garden of memes, and the upvote button is a watering can.

This post is unlisted and is still awaiting moderation. Users' first posts need to go through moderation.

Is it a bug that I can see this post? I got alerted because it was tagged "GPT".

4Raemon
Ah, yeah. New user posts are supposed to be hidden for awhile but there's some edgecases that are hard to notice and track down. Thanks.

Timelines. USG could unilaterally slow AI progress. (Use your imagination.)

Even if only a single person's values are extrapolated, I think things would still be basically fine. While power corrupts, it takes time do so. Value lock-in at the moment of creation of the AI prevents it from tracking (what would be the) power-warped values of its creator.

[This comment is no longer endorsed by its author]Reply
3quetzal_rainbow
I'm frankly not sure how many among respectably-looking members of our societies those who would like to be mind-controlling dictators if they had chance.

My best guess is that there are useful things for 500 MLEs to work on, but publicly specifying these things is a bad move.

1joseph_c
Could you please elaborate? Why is it bad to publicly specify these things?

Agree, but LLM + RL is still preferable to muzero-style AGI.

2Jozdien
I agree, but this is a question of timelines too. Within the LLM + RL paradigm we may not need AGI-level RL or LLMs that can accessibly simulate AGI-level simulacra just from self-supervised learning, both of which would take longer than many points requiring intermediate levels of LLM and RL capabilities, because people are still working on RL stuff now.

I'm not so sure! Some of my best work was done from the ages of 15-16. (I am currently 19.)

[This comment is no longer endorsed by its author]Reply
2Slider
I am all for stimulating stuff to do. That sounds like a case where personal lack of money is not a significant factor. To me it would seem that doing that stuff as a hobbyist would be largely similar (ie money is a nice bonus but tinkering would happen anyway because of intrinsic interest / general development). Not being able to mess with computers because your parents needed hands to pull potatoes from fields would probably also made it hard to be a relevant blip when that employer was searching for talent. I am also more worried about when it systematically affects a lot of people, when "so where do you work?" you would get an eyebrow raising answer "I in fact do not work, but my mother insisted that I should go to school" from a 10 year old. It would actually probably be working a fast food joint to pay on the family car loan interest. If we could make work so enriching that it would bring people up all their life then maybe it would be developmentally desirable environment. But as long as you will have adult unemployed people, I consider the job of children to be playing and any employed minor to be a person that is inappropriately not playing. Then offcourse if a framework where education is preparation to be a cog in a factory leads to schools being even more stiffling than actual factories, having a artifically stably bad environment is worse than unstably bad environment. In certain sense this "prepatory phase" lasts until the end of tetriary education. I am of the impression that "mid stage" people do not push off their work to pick up new skill. By doing the aquisitions early in life we have it "installed" and pay dividends during most of the lenght of life. But the environment where you develop the capabilities and where you can use out of them are different. And the transition costs between them are not always trivial.

Here's an idea for a decision procedure:

  • Narrow it down to a shortlist of 2-20 video ideas that you like
  • For each video, create a conditional prediction market on Manifold with the resolution criterion "if made, would this video get over X views/likes/hours of watch-time", for some constant threshold X
  • Make the video the market likes the most
  • Resolve the appropriate market

[copying the reply here because I don't like looking at the facebook popup]

(I usually do agree with Scott Alexander on almost everything, so it's only when he says something I particularly disagree with that I ever bother to broadcast it. Don't let that selection bias give you a misleading picture of our degree of general agreement. #long)

I think Scott Alexander is wrong that we should regret our collective failure to invest early in cryptocurrency. This is very low on my list of things to kick ourselves about. I do not consider it one of my life's regrets... (read more)

Yes. From the same comment:

Spend a lot of money on ad campaigns and lobbying, and get {New Hampshire/Nevada/Wyoming/Florida} to nullify whatever federal anti-gambling laws exist, and carve out a safe haven for a serious prediction market (which does not currently exist).

And:

You could alternatively just fund the development of a serious prediction market on the Ethereum blockchain, but I'm not as sure about this path, as the gains one could get might be considered "illegal". Also, a fully legalized prediction market could rely on courts to arbitrate m

... (read more)

Agreed. To quote myself like some kind of asshole:

In order for a prediction market to be "serious", it has to allow epistemically rational people to get very very rich (in fiat currency) without going to jail, and it has to allow anyone to create and arbitrate a binary prediction market for a small fee. Such a platform does not currently exist.

2eigen
Thank you for being a temporary asshole, that is a great comment. Does it occur to you how it can be done?    the first prediction market about this was in 8 november on polymarket and surprisingly it was 94 percent probability on them not halting withdrawals 

TikTok isn't doing any work here, I compile the text to mp4 using a script I wrote.

Thanks!

might be able to find a better voice synthesizer that can be a bit more engaging (not sure if TikTok supplies this)

Don't think I can do this that easily. I'm currently calling Amazon Polly, AWS' TTS service, from a python script I wrote to render these videos. Tiktok does supply an (imo) annoying-sounding female TTS voice, but that's off the table since I would have to enter all the text manually on my phone.

experimentation is king.

I could use Amazon's Mechanical Turk to run low-cost focus groups.

2ersatz
You should probably use Google Neural2 voices which are far better.

True, but then I have to time the text-transitions manually.

The "anti-zoomer" sentiment is partially "anti-my-younger-self" sentiment. I, personally, had to expend a good deal of effort to improve my attention span, wean myself off of social media, and reclaim whole hours of my day. I'm frustrated because I know that more is possible.

I fail to see how this is an indictment of your friend's character, or an indication that he is incapable of reading.

That friend did, in fact, try multiple times to read books. He got distracted every time. He wanted to be the kind of guy that could finish books, but he couldn't. I... (read more)

(creating a separate thread for this, because I think it's separate from my other reply)

That friend did, in fact, try multiple times to read books. He got distracted every time. He wanted to be the kind of guy that could finish books, but he couldn’t.

You've described the problem exactly. Your friend didn't have a clear reason to read books. He just had this vague notion that reading books was "good". That "smart people" read lots of books. Why? Who knows, they just do.

I read a lot. But I have never read just for the sake of reading. All of my reading h... (read more)

Further evidence that I should write a factpost investigating whether attention spans have been declining.

1[comment deleted]
3quanticle
When I went to college, I knew a guy who spent >7 hours per day playing World of Warcraft. He ended up dropping out. My dad knows multiple people who failed out of IIT because they spent too much time playing bridge (the card game). Every generation has its anecdotes about smart people who got nothing done because they were too interested in trivialities.

Thank you for the feedback! I didn't consider the inherent jumpiness/grabbiness of Subway Surfers, but you're right, something more continuous is preferable. (edit: but isn't the point of the audio to allow your eyes to stray? hmm)

I will probably also take your advice wrt using the Highlights and CFAR handbook excerpts in lieu of the entire remainder of R:AZ.

9Esben Kran
Thank you for making this, I think it's really great! The idea of the attention-grabbing video footage is that you're not just competing between items on the screen, you're also competing with the videos that come before and after your video. Therefore, yours has to be visually engaging just for that zoomer (et al.) dopamine rush.  Subway Surfers is inherently pretty active and as you mention, the audio will help you here, though you might be able to find a better voice synthesizer that can be a bit more engaging (not sure if TikTok supplies this). So my counterpoint to Trevor1's is that we probably want to keep something like Subway Surfers in the background but that can of course be many things such as animated AI generated images or NASA videos of the space race. Who really knows - experimentation is king.

guessing this wouldn't work without causal attention masking

2Neel Nanda
Yeah, I think that's purely symmetric.

distillation of Taleb's core idea:

expected value estimates are dominated by tail events (unless the distribution is thin-tailed)

repeated sampling from a distribution usually does not yield information about tail events

therefore repeated sampling can be used to estimate EVs iff the distribution is thin-tailed according to your priors

if the distribution is fat-tailed according to your priors, how to determine EV?

estimating EV is much harder

some will say to use the sample mean as EV anyway, they are wrong

in the absence of information on tail events (which is ... (read more)

Could also be that you're a more reliable worker if you have a ton of student debt.

[This comment is no longer endorsed by its author]Reply
1the gears to ascension
you'd be more inclined to be obedient if you're a debt slave, yes
Load More