All of Kabir Kumar's Comments + Replies

if asked about recommendation algoritms, I think it might be much higher - given a basic understanding of what they are, addictiveness, etc

imo, our philosophical progress has been stagnated by media (in the classic sense of the word) - recording ideas makes it harder to forget them. akin to training at very low dropout

In a parallel universe with a saner civilization, there must be tons of philosophy professors workings with tons of AI researchers to try to improve AI's philosophical reasoning.

Sanskrit scholars worked for generations to make Sanskrit better for philosophy

2Adele Lopez
That sounds interesting, do you know a good place to get an overview of what the changes were and how they approached it?
1Kabir Kumar
imo, our philosophical progress has been stagnated by media (in the classic sense of the word) - recording ideas makes it harder to forget them. akin to training at very low dropout

But maybe we are just bad at politics and coalition-building.

 

Mostly due to a feeling of looking down on people imo

1gilch
I thought it was mostly due to the high prevalence of autism (and the social anxiety that usually comes with it) in the community. The more socially agentic rationalists are trying.

Thank you. We just had some writers join, who're, among other things, going to make an up to date About Us section. some out of date stuff is available on https://aiplans.substack.com
Something that we use internally is:  https://docs.google.com/document/d/1wcVlWRTKJqiXOvKNl6PMHCBF3pQItCCcnYwlWvGgFpc/edit?usp=sharing

We're primarily focused on making a site rebuild atm, which has a lot of new and improved features users have been asking for. Preview (lots of form factor stuff broken atm) at: https://ai-plans-site.pages.dev/

3Morpheus
Need to know.
2MondSemmel
No, those are clickbait. 4 is straightforwardly misleading with the meaning of the word "hunt". 2 and 3 grab attention via big dollar numbers without explaining any context. And 1 and 5 are clickbait but wouldn't be if an arbitrary viewer could at any time actually do the things described in the titles, rather than these videos being about some competition that's already happened. Whereas a title saying "Click on this blog post to win $1000" wouldn't be clickbait if anyone could click on the blog post and immediately receive $1000. It would become clickbait if it was e.g. a limited-time offer and expired, but would not be clickbait if the title was changed at that point.

My Clients, The Liars

And All The Shoggoths Merely Players

Acting Wholesomely

These are the most obvious examples. By 'clickbait', here I mean a title that's more for drawing in readers than accurately communicating what the post is about. Doesn't mean it can't be accurate too - after all, MrBeast rarely lies in his video titles - but it means that instead of choosing the title that is most accurate, they chose the most eye catching and baiting title out of the pool of accurate/semi-accurate titles.

2MondSemmel
I don't really agree with this definition of clickbait. A title that merely accurately communicates what the post is about, is usually a boring one and thus communicates that the post is boring and not worth reading. Also see my comment here. Excerpt:

update on my beliefs - among humans of above average intelligence, the primary factor for success is willpower - stamina, intensity and consistency

You might be interested in the Broad List of Vulnerabilities

Thank you, this is useful. Planning to use this for AI-Plans.

Hasn't that happened?

2Mitchell_Porter
I mean an AI that does its own reading, and decides what to post about. 

Yet, the top posts on LessWrong are pretty much always clickbaited, just in the LessWrong lingo.
The Curated Posts seem to be some of the worst cases of this:

I find that LessWrong folk in general are really, really, susceptible to deception and manipulation when it's done in their language.

7habryka
I... really don't see any clickbait here. If anything these titles feel bland to me (and indeed I think LW users could do much better at making titles that are more exciting, or more clearly highlight a good value proposition for the reader, though karma makes up for a lot).  Like, for god's sake, the top title here is "Social status part 1/2: negotiations over object-level preferences". I feel like that title is at the very bottom of potential clickbaitiness, given the subject matter.
4ryan_greenblatt
Which of these titles are click bait? I disagree with the thesis of some, but none seem like click bait titles to me.

for sure. right now it's just a google form and google sheets. would you be interested in taking charge of this?

2sudhanshu_kasewa
No, this is not something I can undertake -- however, the effort itself need not be very complicated. You've already got a list of Misalignment types in the form: create a google doc with definitions/descriptions of each of these, and put a link to that doc in this question.

Thank you, I've labelled that as the form link now and added the DB link.

Thank you! I'll add those as well!

Perhaps a note on Pre-Requisites would be useful. 
E.g. the level of math & comp sci that's assumed. 
Suggestion: try going through the topics to 50+ random strangers. Wildly useful for improving written work. 

Yes, that's what I'm referring to. As in, getting that enacted as a policy.

3ryan_greenblatt
Ok, well the most immediate next step is to try to ensure that some sort of commitment along these lines gets included in RSPs (or preparedness frameworks or etc). This sort of commitment is technically a special case of RSP commitment like "we'll make a safety case (approved by an independent review board) that we'll be safe from risks due to models autonomously causing harm", but I think it's pretty likely that having a more specific incident response policy is pretty good. I'm unsure what specific actions along these lines can be taken to influence RSPs, but insofar as you already have some influence over some such process, pushing for this sort of commitment seems probably good (as it's extremely common sense and also moderately useful). I'm uncertain about what the space looks like as far as executive orders or other us/uk/etc goverment policy. But, insofar as anything might include something like incident reporting, specifically highlighting this sort of case seems pretty good.

This is an absurdly low bar, but yes, this should be done.
How can I help?

2ryan_greenblatt
What are you refering to by "this"? Maybe "once your model has demonstrably tried to escape, stop deploying it”?
1David James
Would you please expand on how ai-plans.com addresses the question from the post above ... ? I took a look at ai-plans, but I have yet to find information about: 1. How does it work? 2. Who created it? 3. What is the motivation for building it? 4. What problem(s) will ai-plans help solve? 5. Who controls / curates / moderates it? 6. What is the process/algorithm for: curation? moderation? ranking? I would suggest (i) answering these questions on the ai-plans website itself then (ii) adding links here.

there is an issue with surface level insights being unfaily weighted, but this is solvable, imo. especially with youtube, which can see which commenters have watched the full video.

You can compete with someone more intelligent but less hard working by being more organized, disciplined, focused and hard working. And being open to improvement.

1Kabir Kumar
update on my beliefs - among humans of above average intelligence, the primary factor for success is willpower - stamina, intensity and consistency