All of Mateusz Bagiński's Comments + Replies

It is ambiguous, but it's hinting more strongly towards being a featured author guest because "normal/usual/vanilla guests" are not Being Invited by the organizers to attend the conference in the sense in which this word is typically used in this context.

But fair, I'll ETA-clarify.

It's clear that many people at least don't mind Cremieux being invited [ETA: as a featured author-guest] to LessOnline, but it's also clear (from this comment thread) that many people do mind Cremieux being invited to LessOnline, and some of them mind it quite strongly. 

This is a (potential) reason to reconsider the invitation and/or explicitize some norms/standards that prospective LessOnline invitees are expected to meet.

Small ~nitpick/clarification: in my understanding, at issue is Crémieux being a featured guest at LessOnline, rather than being allowed to attend LessOnline; "invited to" is ambiguous between the two.

I think the author meant that they achieve higher scores on the FrontierMath benchmark.

Do you mean that the concrete evidence of Cremieux's past behavior presented in the comments justifies the OP?

Yeah it seems sufficient, particularly the Reddit post is highly irresponsible. 

1papetoast
Literally just copy pasted your question. https://chatgpt.com/share/681a16b2-58f4-8002-8e24-85912ba3d891 (seems to found another censored person, Brian Hood though) For other models I asked in OpenRouter and idk any easy way of sharing chats

do you think that Stanovich's reflective mind and need for cognition are downstream from these two?

AI2027 is esoteric enough that it probably never will.

Does it also mean that it won't have a significant (direct) impact on the CCP's AI strategy?

What was the name of the rich guy whose information was "deleted"/"unlearned" from ChatGPT sometime in 2024 because he was like, "Hey, why does this model know so much about me?"?

IIRC it came out when people realized that asking (some model of) ChatGPT about him breaks the model in some way? And then it turned out that there were more names that could cause this effect, and all of some influential people.

3papetoast
You got me curious, I thought "no way the newer models with late 2024 knowledge cutoff date and/or search cant figure this out" but apparently not. Tried 5 minutes and couldn't get any model to output the answer
7Shankar Sivarajan
David Mayer de Rothschild.

It is said in the proverbs of hell:

You never know what is enough unless you know what is more than enough.

Are you trying to say that for any X, instead of X-maturity, we should instead expect X-foom until the marginal returns get too low?

2Vladimir_Nesov
A pre-abundance precedent about X offers poor framing for thinking about the consequences of discovering a scalable process of producing X. Before abundance, it's artisanal and quirky and path-dependent, the extremes are rare and dysfunctional, so people don't worry about it too much. There is security in it looking like an equilibrium, but not being truly settled, so that people can influence things. Abundance brings maturity, changes the character of the equilibrium. So not foom necessarily, just a promise of maturity at some point, which wouldn't have been as salient before there is a scalable process of production. And there is an excuse of ignoring the possibility even longer, because of the total lack of historical precedent (of the associated problems).

I feel sad that a lot of these talk about how everything is broken in a way that is clearly overstated. 

Linda Linsefors's Rule of Succession

“Thanks for Nothing” Effect: “If the original post in a thread ends with the sentence ‘Thanks in advance!’ it is exponentially less likely that it will be replied to.”

Is there any data on this?

Trivium: Naming laws of nature after their discoverers (and inventions after their inventors) is a Western peculiarity, at least according to Joe Henrich. From The WEIRDest People in the World:

with an increased focus on mental states, intellectuals began to associate new ideas, concepts, and insights with particular individuals, and to credit the first founders, observers, or inventors whenever possible. Our commonsensical inclination to associate inventions with their inventors has been historically and cross-culturally rare. This shift has been marked by

... (read more)
1Rasool
Not withstanding :P 
1rogersbacon
interesting! 
2Elizabeth
The plan was to build up some back catalog and then decide. Right now it looks like we're only going to do two more episodes so it doesn't seem worth setting up.

slanted treadmill

That's a nice, concise handle!

not sure is this makes any sense )

I think I understand.

Has it always been with you? Any ideas what might be the reason for the bump at Thursday? Was Thursday in some sense "special" for you when you were a kid?

1Sergii
Ha, thinking back to childhood I get it now, it's the influence of the layot of the school daily journal in USSR/Ukraine, like https://cn1.nevsedoma.com.ua/images/2011/33/7/10000000.jpg

For as long as I can remember, I have always placed dates on an imaginary timeline, that "placing" involving stuff like fuzzy mental imagery of events attached to the date-labelled point on the timeline. It's probably much less crisp than yours because so far I haven't tried to learn history that intensely systematically via spaced repetition (though your example makes me want to do that), but otherwise sounds quite familiar.

For as long as I can remember, I've had a very specific way of imagining the week. The weekdays are arranged on an ellipse, with an inclination of ~30°, starting with Monday in the bottom-right, progressing along the lower edge to Friday in the top-left, then the weekend days go above the ellipse and the cycle "collapses" back to Monday.

Actually, calling it "ellipse" is not quite right because in my mind's eye it feels like Saturday and Sunday are almost at the same height, Sunday just barely lower than Saturday.

I have a similar ellipse for the year, this ... (read more)

1GWill
This is very similar to how l perceive time! what I find interesting is that while I’ve heard people talk about the way they conceptualize time before I’ve never heard anyone else mention the bizarre geometry aspect. The sole exceptions to this were my Dad and Grandfather, who brought this phenomenon to my attention when I was young.
1Joey KL
I just met someone recently who has this! They said they have always visualized the months of the year as on a slanted treadmill, unevenly distributed. They described it as a form of synesthesia, which is conceptually consistent with how I experience grapheme-color associative synesthesia.
1Sergii
I have similar thing for week days, but somehow with a weird shape? in general, it's a similar cycle, but flipped horizontally, going left to right: on top it's: sun, sat on the bottom: mon, tue, wed, thu, fri the shape connecting days goes downwards from sun to mon, tue, wed, then upwards to thu, then down to fri, then up to sat, sun, closing the loop. not sure is this makes any sense )
1Jacob G-W
I have something like this for years: https://www.lesswrong.com/posts/j8WMRgKSCxqxxKMnj/what-i-think-about-when-i-think-about-history

What were the biggest boosts that you and your colleagues got from LLMs?

Dalcy102

Speaking from the perspective of someone still developing basic mathematical maturity and often lacking prerequisites, it's very useful as a learning aid. For example, it significantly expanded the range of papers or technical results accessible to me. If I'm reading a paper containing unfamiliar math, I no longer have to go down the rabbit hole of tracing prerequisite dependencies, which often expand exponentially (partly because I don't know which results or sections in the prerequisite texts are essential, making it difficult to scope my focus). Now I c... (read more)

Eric Schwitzgebel has argued by disjunction/exhaustion for the necessity of craziness in the sense of "contrary to common sense and we are not epistemically compelled to believe it" at least in the context of philosophy of mind and cosmology.

https://faculty.ucr.edu/~eschwitz/SchwitzAbs/CrazyMind.htm 

https://press.princeton.edu/books/hardcover/9780691215679/the-weirdness-of-the-world 

While browsing through Concordia AI's report (linked by @Mitchell_Porter ), I stumbled on an essay by Yin Hejun (China's Minister of Science and Technology) from ~1y ago, which Concordia's AI Safety in China Substack summarizes as:

Background: Minister of the Ministry of Science and Technology (MOST) YIN Hejun (阴和俊) published an essay on AI in CAC’s magazine, “China Cyberspace (中国网信).” The essay outlines China’s previous efforts in AI development, key accomplishments, and plans moving forward. 

Discussion of governance and dialogue: Generally, the essay

... (read more)
4eggsyntax
Complex and ambivalent views seem like the correct sort of views for governments to hold at this point. I also don't speak Chinese, but my impression is that machine translations of high-context languages like Chinese need to be approached with considerable caution -- a lot of context on (eg) past guidance from the CCP may be needed to interpret what they're saying there. I'm only ~70% on that, though, happy to be corrected by someone more knowledgeable on the subject.

People's "deep down motivations" and "endorsed upon reflection values," etc, are not the only determiners of what they end up doing in practice re influencing x-risk. 

2Neel Nanda
I agree with that. I was responding specifically to this:

To steelman a devil's advocate: If your intent-aligned AGI/ASI went something like

oh, people want the world to be according to their preferences but whatever normative system one subscribes to, the current implicit preference aggregation method is woefully suboptimal, so let me move the world's systems to this other preference aggregation method which is much more nearly-Pareto-over-normative-uncertainty-optimal than the current preference aggregation method

and this would be, in an important sense, more democratic, because the people (/demos) would have more influence over their societies.

sanyer113

Yeah, I can see why that's possible. But I wasn't really talking about the improbable scenario where ASI would be aligned to the whole of humanity/country, but about a scenario where ASI is 'narrowly aligned' in the sense that it's aligned to its creators/whoever controls it when it's created. This is IMO much more likely to happen since technologies are not created in a vacuum.

While far from what I hoped for, this is the closest to what I hoped for that I managed to find so far: https://www.chinatalk.media/p/is-china-agi-pilled 

Overall, the Skeptic makes the stronger case — especially when it comes to China’s government policy. There’s no clear evidence that senior policymakers believe in short AGI timelines. The government certainly treats AI as a major priority, but it is one among many technologies they focus on. When they speak about AI, they also more often than not speak about things like industrial automation as oppo

... (read more)
7Mateusz Bagiński
While browsing through Concordia AI's report (linked by @Mitchell_Porter ), I stumbled on an essay by Yin Hejun (China's Minister of Science and Technology) from ~1y ago, which Concordia's AI Safety in China Substack summarizes as: And in Concordia's slides: I don't speak Chinese, so I Google-translated the essay to skim/read it. It seems to fit the narrative of "China wants to accelerate AI just as it would like to accelerate any useful technology, but they're not particularly buying into singularity/AGI/ASI." Something that got Google-translated into "universal AI" is mentioned 2 times, 1 time for "general AI", mostly in the context of language models, but without much elaboration. There's no "let's get AI that can do everything for us", more like "let's get AI so that we are better at this and this and that". (Although some local Chinese governments did announce policies on "AGI": 1, 2.) ---------------------------------------- (I weakly predict that I'm going to be using this thread as a dump for whatever new info on this topic I find worth sharing.)
4Mitchell_Porter
That's an informative article.  There's lots of information about AI safety in China at Concordia AI, e.g. this report from a year ago. But references to the party or the government seem to be scarce, e.g. in that 100-page report, the only references I can see are on slide 91. 

The fact that their models are on par with openAI and anthropic but it’s open source.

This is perfectly consistent with my

"just": build AI that is useful for whatever they want their AIs to do and not fall behind the West while also not taking the Western claims about AGI/ASI/singularity at face value?

You can totally want to have fancy LLMs while not believe in AGI/ASI/singularity.

There are people from the safety community arguing for jail for folks who download open source models. 

Who? What proportion of the community are they? Also, all open-source m... (read more)

I'm not sure why this is.

The most straightforward explanation would be that there are more underexploited niches for top-0.01%-intelligence people than there are top-1%-intelligence people.

After thinking about it for a few minutes, I'd expect that MadHatter has disengaged from this community/cause anyway, so that kind of public reveal is not going to hurt them much, whereas it might have a big symbolic/common-knowledge-establishing value.

4niplav
I think having my real name publicly & searchably associated with scummy behavior would discourage me from doing something, both in terms of future employers & random friends googling, as well as LLMs being trained on the internet.

Self-Other Overlap: https://www.lesswrong.com/posts/hzt9gHpNwA2oHtwKX/self-other-overlap-a-neglected-approach-to-ai-alignment?commentId=WapHz3gokGBd3KHKm

Emergent Misalignment: https://x.com/ESYudkowsky/status/1894453376215388644 

He was throwing vaguely positive comments about Chris Olah, but I think always/usually caveating it with "capabilities go like this [big slope], Chris Olah's interpretability goes like this [small slope]" (e.g., on Lex Fridman podcast and IIRC some other podcast(s)).

ETA: 

SolidGoldMagikarp: https://www.lesswrong.com/posts/... (read more)

I know almost nothing about audio ML, but I would expect one big inconvenience when doing audio-NN-interp to be that a lot of complexity in sound is difficult to represent visually. Images and text (/token strings) don't have this problem.

I am confused about what autism is. Whenever I try to investigate this question I end up coming across long lists of traits and symptoms where various things are unclear to me.

Isn't that the case with a lot of psychological/psychiatric conditions?

Criteria for a major depressive episode include "5 or more depressive symptoms for ≥ 2 weeks", and there are 9 depressive symptoms, so you could have 2 individuals diagnosed with a major depressive episode but having only one depressive symptom in common.

2Adam Zerner
I get the sense that autism is particularly unclear, but I haven't looked closely enough at other conditions to be confident in that.

I know. I just don't expect it to.

Steganography /j

6niplav
I think it's possible! If it's used to encode relevant information, then it could be tested by running software engineering benchmarks (e.g. SWE-bench) but removing any trailing whitespace during generation, and checking if the score is lower.
3Garrett Baker
If it is encoding relevant info then this would be the definition of steganography

So it seems to be a reasonable interpretation that we might see human level AI around mid-2030 to 2040, which happens to be about my personal median.

What are the reasons your median is mid-2030s to 2040, other than this way of extrapolating the METR results?

How does the point about Hitler murder plots connect to the point about anthropics?

2Viliam
In a more peaceful world the science advanced faster and the AI already killed us?

they can’t read Lesswrong or EA blogs

VPNs exist and are probably widely used in China + much of "all this work" is on ArXiv etc.

If that was his goal, he has better options.

2O O
Yes, the likely outcome of a long tariff regime is China replaces the U.S. as the hegemon + AI race leader and they can’t read Lesswrong or EA blogs there so all this work is useless. 

I'm confused about how to think about this idea, but I really appreciate having this idea in my collection of ideas.

To show how weird English is: English is the only proto indo european language that doesn't think the moon is female ("la luna") and spoons are male (“der Löffel”). I mean... maybe not those genders specifically in every language. But some gender in each language.

Persian is ungendered too. They don't even have gendered pronouns.

https://en.wikipedia.org/wiki/Persian_grammar 

3JenniferRM
Thank you for the correction! I didn't realize Persian descended from PIE too. Looking at the likely root cause of my ignorance, I learned that Kurdish and Pashto are also PIE descended. Pashto appears to have noun gender, but I'm getting hints that at least one dialect of Kurdish also might not?! If Sorani doesn't have gendered nouns then I'm going to predict (1) maybe Kurdish is really old and weird and interesting (like branching off way way long ago with more time to drift) and/or (2) there was some big trade/empire/mixing simplification that happened "more recently" with divergence later? If neither of those are true, then my larger heuristic about "why English is weird" might have a deep abstract counter example, and deserve lower credence. Persian is a language of empire and social mixing, so its "similar simplification" doesn't actually function as a strong counter-example to the broader thesis, but it is still great to be surprised :-)

Writing articles in Chinese for my family members, explaining things like cognitive bias, evolutionary psychology, and why dialectical materialism is wrong.

Your needing to write them seems to suggest that there's not enough content like that in Chinese, in which case it would plausibly make sense to publish them somewhere?

I'm also curious about how your family received these articles.

4Wei Dai
I'm not sure how much such content exist in Chinese, because I didn't look. It seems easier to just write new content using AI, that way I know it will cover the ideas/arguments I want to cover, represent my views, and make it easier for me to discuss the ideas with my family. Also reading Chinese is kind of a chore for me and I don't want to wade through a list of search results trying to find what I need. I thought about publishing them somewhere, but so far haven't: * concerns about publishing AI content (potentially contributing to "slop") * not active in any Chinese forums, not familiar with any Chinese publishing platforms * probably won't find any audience (too much low quality content on the web, how will people find my posts) * don't feel motivated to engage/dialogue with a random audience, if they comment or ask questions

I think that the scenario of the war between several ASI (each merged with its origin country) is underexplored. Yes, there can be a value handshake between ASIs, but their creators will work to prevent this and see it as a type of misalignment. 

Not clear to me, as long as they expect the conflict to be sufficiently destructive.

I wonder whether it's related to this https://x.com/RichardMCNgo/status/1866948971694002657 (ping to @Richard_Ngo to get around to writing this up (as I think he hasn't done it yet?))

4niplav
Yeah, there's also reports on Tai Chi doing the same, see @cookiecarver's report.

Since this is about written English text (or maybe more broadly, text in Western languages written in Latinic or Cyrillic), the criterion is: ends with a dot, starts with an uppercase letter.

5TAG
Then the phenomenon could be stem from punctuation habits, as @bfinn says. Did you notice that my original comment doesn't contain a sentence, by your standards?

Fair enough. Modify my claim to "languages tend to move from fusional to analytic (or something like that) as their number of users expands".

Related: https://www.lesswrong.com/posts/Pweg9xpKknkNwN8Fx/have-attention-spans-been-declining 

Another related thing is that the grammar of languages appears to be getting simpler with time. Compare the grammar of Latin to that of modern French or Spanish. Or maybe not quite simpler but more structured/regular/principled, as something like the latter has been reproduced experimentally https://royalsocietypublishing.org/doi/10.1098/rspb.2019.1262 (to the extent that this paper's findings generalize to natural language evolution).

4TAG
The idea that grammar is just inflection is misleading: languages that are mostly isolating can have complex ordering rules,like the the notorious adjective ordering of English. As for french ...Moi, je ne me défile pas. 1st person. Sing. 1st person. Sing, again. Negative. 1st person. Sing, reflexive. Verb!!! Negative,again.
simon212

FWIW there is a theory that there is a cycle of language change, though it seems maybe there is not a lot of evidence for the isolating -> agglutinating step. IIRC the idea is something like that if you have a "simple" (isolating) language that uses helper words instead of morphology eventually those words can lose their independent meaning and get smushed together with the word they are modifying.

Somewhat big if true although the publication date makes it marginally less likely to be true.

[This comment is no longer endorsed by its author]Reply
7nostalgebraist
I saw some discussion of this incident in the Eleuther discord on 3/30, including a screenshot of the system message containing the "emulate the tone" line.  So it's not an April Fools' thing.

The outline in that post is also very buggy, probably because of the collapsible sections.

Any info on how this compares to other AI companies?

Link to the source of the quote?

8gwern
A google of the first paragraph takes you quickly to https://www.bluesci.co.uk/posts/forgotten-knowledge
Load More