All of Garrett Baker's Comments + Replies

That is the wrong question to ask. By their nature the result of experiments is unknown. The bar is whether or not in expectation and on the margin do they provide positive externalities, and the answer is clearly yes.

Note the error bars in the original

After that I was writing shorter posts but without long context the things I write are very counterintuitive. So they got ruined)

This sounds like a rationalization. It seems much more likely the ideas just aren't that high quality if you need a whole hour for a single argument that couldn't possibly be broken up into smaller pieces that don't suck.

Edit: Since if the long post is disliked, you can say "well they just didn't read it", and if the short post is disliked you can say "well it just sucks because its small". Meanwhile, it should in fact be pret... (read more)

1ank
It’s a combination of factors, I got some comments on my posts so I got the general idea: 1. My writing style is peculiar, I’m not a native speaker 2. Ideas I convey took 3 years of modeling. I basically Xerox PARCed (attempted and got some results) the ultimate future (billions of years from now). So when I write it’s like some Big Bang: ideas flow in all directions and I never have enough space for them) 3. One commenter recommended to change the title and remove some tags, I did it 4. If I use ChatGPT to organize my writing it removes and garbles it. If I edit it, I like having parenthesis within parenthesis 5. I write a book to solve those problems but mainly human and AI alignment (we better to stop AI agents, it's suicidal to make them) towards the best possible future, to prevent dystopias, it’ll be organized this way: * I’ll start with the “Ethical Big Bang” (physics can be modeled as a subset of ethics), * will chronologically describe and show binary tree model (it models freedoms, choices, quantum paths, the model is simple and ethicophysical so those things are the same in it) of the evolution of inequality from the hydrogen getting trapped in first stars to * hunter-gatherers getting enslaved by agriculturalist and * finish with the direct democratic simulated multiverse vs dystopia where an AI agent grabbed all our freedoms. * And will have a list of hundreds of AI safety ideas for considering.

usually if you randomly get a downvote early instead of an upvote, so your post has “-1” karma now, then no one else will open or read it

I will say that I often do read -1 downvoted posts, I will also say that much of the time it is deserved, despite how noisy a signal it may be.

Some of my articles take 40 minutes to read, so it can be anything, downvotes give me zero information and just demotivate more and more.

I think you should try writing shorter posts. Both for your sake (so you get more targeted information), and for the readers' sake.

1ank
Thank you for responding and reading -1 posts, Garrett! It’s important. The long post was actually a blockbuster for me that got 16 upvotes before I screwed the title and it got back to 13) After that I was writing shorter posts but without long context the things I write are very counterintuitive. So they got ruined) I think the right approach was to snowball it: to write each next post as a longer and longer book. I’m writing it now but outside of the website.

https://www.science.org/content/blog-post/alkanes-mars there are alkanes -- big organic molecules -- on Mars. these can be produced by abiotic processes, but usually that makes shorter chains than these. so....life? We Shall See.

Very exciting! I think the biggest "loophole" here is probably that they used a novel technique for detection, maybe if we used that technique more we would have to update the view that such big molecules are so unlikely to be produced non-biologically.

I'm a bit skeptical, there's a reasonable amount of passed-down wisdom I've heard claiming (I think justifiably) that

  1. If you write messy code, and say "I'll clean it later" you probably won't. So insofar as you eventually want to discover something others build upon, you should write it clean from the start.

  2. Clean code leads to easier extensibility, which seems pretty important eg if you want to try a bunch of different small variations on the same experiment.

  3. Clean code decreases the number of bugs and the time spent debugging. This seems especially

... (read more)
4Daniel Tan
Yeah, I agree with all this. My main differences are:  1. I think it's fine to write a messy version initially and then clean it up when you need to share it with someone else. 2. By default I write "pretty clean" code, insofar as this can be measured with linters, because this increases readability-by-future-me.  Generally i think there may be a Law of Opposite Advice type effect going on here, so I'll clarify where I expect this advice to be useful:  1. You're working on a personal project and don't expect to need to share much code with other people. 2. You started from a place of knowing how to write good code, and could benefit from relaxing your standards slightly to optimise for 'hacking'. (It's hard to realise this by yourself - pair programming was how I discovered this) 

Ok first, when naming things I think you should do everything you can to not use double-negatives. So you should say "gym average" or "no gym average". Its shorter, and much less confusing.

Second, I'm still confused. Translating what you said, we'd have "no gym removed average" -> "gym average" (since you remove everyone who doesn't go to the gym meaning the only people remaining go to the gym), and "gym removed average" -> "no gym average" (since we're removing everyone who goes to the gym meaning the only remaining people don't go to the gym).

There... (read more)

Note: You can verify this is the case by filtering for male respondents with male partners and female respondents with female partners in the survey data

I think the math works out to be that the variation is much more extreme when you get to much more extreme probabilities. Going from 4% to 8% is 2x profits, but going from 50% to 58% is only 1.16x profits.

This seems likely to depend on your preferred style of research, so what is your preferred style of research?

2Daniel Tan
Good question! These practices are mostly informed by doing empirical AI safety research and mechanistic interpretability research. These projects emphasize fast initial exploratory sprints, with later periods of 'scaling up' to improve rigor. Sometimes most of the project is in exploratory mode, so speed is really the key objective.  I will grant that in my experience, I've seldom had to build complex pieces of software from the ground up, as good libraries already exist.  That said, I think my practices here are still compatible with projects that require more infra. In these projects, some of the work is building the infra, and some of the work is doing experiments using the infra. My practices will apply to the second kind of work, and typical SWE practices / product management practices will apply to the first kind of work. 

And then if we say the bottleneck to meritocracy is mostly c rather than a or b, then in fact it seems like our society is absolutely obsessed with making our institutions highly accessible to as broad a pool of talent as possible. There are people who make a whole career out of just advocating for equality.

I work at GDM so obviously take that into account here, but in my internal conversations about external benchmarks we take cheating very seriously -- we don't want eval data to leak into training data, and have multiple lines of defense to keep that from happening.

What do you mean by "we"? Do you work on the pretraining team, talk directly with the pretraining team, are just aware of the methods the pretraining team uses, or some other thing?

I don't work directly on pretraining, but when there were allegations of eval set contamination due to detection of a canary string last year, I looked into it specifically. I read the docs on prevention, talked with the lead engineer, and discussed with other execs.

So I have pretty detailed knowledge here. Of course GDM is a big complicated place and I certainly don't know everything, but I'm confident that we are trying hard to prevent contamination.

More to the point, I haven't seen people try to scale those things either. The closest might be something like TripleByte? Or headhunting companies? Certainly when I think of a typical (or 95th-99th percentile) "person who says they care a lot about meritocracy" I'm not imagining a recruiter, or someone in charge of such a firm. Are you?

I think much of venture capital is trying to scale this thing, and as you said they don't use the framework you use. The philosophy there is much more oriented towards making sure nobody falls beneath the cracks. Provide... (read more)

2Garrett Baker
And then if we say the bottleneck to meritocracy is mostly c rather than a or b, then in fact it seems like our society is absolutely obsessed with making our institutions highly accessible to as broad a pool of talent as possible. There are people who make a whole career out of just advocating for equality.

Also this conclusion is highly dependent on you, who has thought about this topic for all of 10 minutes, out-thinking the hypothetical people who are actually serious about meritocracy. For example perhaps they do more one-on-one talent scouting or funding, which is indeed very very common and seems to be much more in-demand than psychometric evaluations.

6Linch
I thought about this for more than 10 minutes, though on a micro rather than macro level (scoped as "how can more competent people work on X" or "how can you hire talented people"). But yeah more like days rather than years. 1. I think one-on-one talent scouting or funding are good options locally but are much less scalable than psychometric evaluations. 2. More to the point, I haven't seen people try to scale those things either. The closest might be something like TripleByte? Or headhunting companies? Certainly when I think of a typical (or 95th-99th percentile) "person who says they care a lot about meritocracy" I'm not imagining a recruiter, or someone in charge of such a firm. Are you?  

Given that ~ no one really does this, I conclude that very few people are serious about moving towards a meritocracy.

The field you should look at I think is Industrial and Organizational Psychology, as well as the classic Item Response Theory.

4Linch
Makes sense! I agree that this is a valuable place to look. Though I am thinking about tests/assessments in a broader way than you're framing it here. Eg things that go into this meta-analysis, and improvements/refinements/new ideas, and not just narrow psychometric evaluations. 

I suspect the vast majority of that sort of name-calling is much more politically motivated than based on not seeing the right slogans. For example if you go to Pause AI's website the first thing you see is a big, bold

and AI pause advocates are constantly arguing "no, we don't actually believe that" to the people who call them "luddites", but I have never actually seen anyone change their mind based on such an argument.

2Mateusz Bagiński
My model is that 1. some of it is politically/ideologically/self-interest-motivated 2. some of it is just people glancing at a thing, forming an impression, and not caring to investigate further 3. some of it is people interacting with the thing indirectly via people from the first two categories; some subset of them then take a glance at the PauseAI website or whatever, out of curiosity, form an impression (e.g. whether it matches what they've heard from other people), don't care to investigate further Making slogans more ~precise might help with (2) and (3)

I don't think Pause AI's current bottleneck is people being pro AI in general not wanting to join (but of course I could be wrong). Most people are just against AI, and Pause AI's current strategy is to make them care enough about the issue to use their feet, while also telling them "its much much worse than you would've imagined bro".

2Mateusz Bagiński
Some people misinterpret/mispaint them(/us?) as "luddites" or "decels" or "anti-AI-in-general" or "anti-progress". Is it their(/our?) biggest problem, one of their(/our?) bottlenecks? Most likely no. It might still make sense to make marginal changes that make it marginally harder to do that kind of mispainting / reduce misinterpretative degrees of freedom.

That's a whole seven words!, most of which are a whole three syllables! There is no way a motto like that catches on.

Viliam100

How about "make computers stupid again"?

2RHollerith
Maybe "motto" is the wrong word. I meant words / concepts to use in a comment or in a conversation. "Those companies that created ChatGPT, etc? If allowed to continue operating without strict regulation, they will cause an intelligence explosion."
2Mateusz Bagiński
You can still include it in your protest banner portfolio to decrease the fraction of people whose first impression is "these people are against AI in general" etc.

An effect I noticed: Going through Aella's correlation matrix (with poorly labeled columns sadly), a feature which strongly correlates with the length of a relationship is codependency. Plotting question 20. "The long-term routines and structure of my life are intertwined with my partner's" (li0toxk) assuming that's what "codependency" refers to

The shaded region is a 95% posterior estimate for the mean of the distribution conditioned on the time-range (every 2 years) and cis-male respondents, with prior .

Note also that codependency and sex sat... (read more)

This seems right as a criticism, but this seems better placed on the EA forum. I can't remember the last time I heard anyone talking about ITN on LessWrong. There are many considerations ITN leaves out, which should be unsurprising given how simplified it is.

1Shaïman
Thanks for your reply. I also posted this question on the Effective Altruism forum, but I thought some people here might also have ideas on the topic.  

So, the recipe for making a broken science you can't trust is

  1. The public cares a lot about answers to questions that fall within the science's domain.
  2. The science currently has no good attack angles on those questions.

To return to LessWrong's favorite topic, this doesn't bode well for alignment.

I understand the argument, I think I buy a limited version of it (and also want to acknowledge that it is very clever and I do like it), but I also don't think this can explain the magnitude of the difference between the different fields. If we go back and ask "what was physics' original goal?" we end up with "to explain how the heavens move, and the path that objects travel", and this has basically been solved. Physicists didn't substitute this for something easier. The next big problem was to explain heat & electricity, and that was solved. Then the ... (read more)

I understand the argument, I think I buy a limited version of it (and also want to acknowledge that it is very clever and I do like it)…

Thanks! I've so appreciated your comments and the chance to think about this with you!

…but I also don't think this can explain the magnitude of the difference between the different fields.

I think that’s right—and we agree. As we note in the post, we only expect our hypotheses to explain a fairly modest fraction of the differences between fields. We see our contribution as showing how certain structural features—e.g., the c... (read more)

I will also note that Aella's relationships data is public, and has the following questions:

1. Your age? (rkkox57)
2. Which category fits you best? (4790ydl)
3. In a world where your partner was fully aware and deeply okay with it, how much would you be interested in having sexual/romantic experiences with people besides your partner? (ao3mcdk)
4. In a world where you were fully aware and deeply okay with it, how much would *your partner* be interested in having sexual/romantic experiences with people besides you? (wcq3vrx)
5. To get a little more specifi
... (read more)

Historically attempts to curtail this right lead to really really dark places. Part of living in a society with rights and laws is that people will do bad things the legal system has no ability to prevent. And on net, that’s a good thing. See also.

An obvious answer you missed: Lacking a prenup, courts often rule in favor of the woman over the man in the case of a contested divorce.

I didn't say anything about temperature prediction, and I'd also like to see any other method (intuition based or otherwise) do better than the current best mathematical models here. It seems unlikely to me that the trends in that graph will continue arbitrarily far.

Thanks for the pointer to that paper, the abstract makes me think there's a sort of slow-acting self-reinforcing feedback loop between predictive error minimisation via improving modelling and via improving the economy itself.

Yeah, that was my claim.

Even more importantly, we only measured the AIs at software tasks and don't know what the trend is for other domains like math or law, it could be wildly different.

You probably mention this somewhere, but I'll ask here, are you currently researching whether these results hold for those other domains? I'm personally more interested about math than law.

4Thomas Kwa
It's expensive to construct and baseline novel tasks for this (we spent well over $100k on human baselines) so what we are able to measure in the future depends on whether we can harvest realistic tasks that naturally have human data. You could do a rough analysis on math contest problems, say assigning GSM8K and AIME questions lengths based on a guess of how long expert humans take, but the external validity concerns are worse than for software. For one thing, AIME has much harder topics than GSM8K (we tried to make SWAA not be artificially easier or harder than HCAST); for another, neither are particularly close to the average few minutes of a research mathematician's job.

It does not seem like this writer is aware of the Von Neumann–Morgenstern utility theorem. There are criticisms one can level against utility as a concept, but the central question ends up being which of those axioms do you disagree with and why? For example, Garabrant's Geometric Rationality is a great counter if you're looking for one.

Edit: I notice that all of your previous posts have been of this same format, and they all consistently receive negative karma. You should probably reconsider what you post to this forum.

0danielechlin
Furthermore it's pretty basic flaws by LW standards, like the "map/territory" which is the first post in the first sequence. I don't think "discussing basic stuff" is wrong by itself, but doing so by shuttling in someone else's post is sketch, and when that post is also some sort of polemical countered by the first post in the first sequence on LW it starts getting actively annoying.

The weather, or the behavior of any economy larger than village size, for example -- systems so chaotically interdependent that exact prediction is effectively impossible (not just in fact but in principle).

Flagging that those two examples seem false. The weather is chaotic, yes, and there's a sense in which the economy is anti-inductive, but modeling methods are advancing, and will likely find more loop-holes in chaos theory.

For example, in thermodynamics, temperature is non-chaotic while the precise kinetic energies and locations of all particles are.... (read more)

3Mo Putera
Thanks for the pointer to that paper, the abstract makes me think there's a sort of slow-acting self-reinforcing feedback loop between predictive error minimisation via improving modelling and via improving the economy itself. re: weather, I'm thinking of the chart below showing how little gain we get in MAE vs compute, plus my guess that compute can't keep growing far enough to get MAE < 3 °F a year out (say). I don't know anything about advancements in weather modelling methods though; maybe effective compute (incorporating modelling advancements) may grow indefinitely in terms of the chart.

My model is that early on physics had very impressive & novel math, which attracted people who like math, who did more math largely with the constraint the math had to be trying to model something in the real world, which produced more impressive & novel math, which attracted more people who like math, etc etc, and this is the origin of the equilibrium.

Note a similar argument can be made for economics, though the nice math came much later on, and obviously was much less impactful than literally inventing calculus.

Happy to take your word on these things if the wikipedia article is unrepresentative!

In contrast, physicists were not committed to discovering the periodic table, fields or quantum wave functions. Many of the great successes of physics are answers to question no one would think to ask just decades before they were discovered. The hard sciences were formed when frontiers of highly tractable and promising theorizing opened up.

This seems a crazy comparison to make[1]. These seem like methodological constraints. Are there any actual predictions past physics was trying to make which we still can't make and don't even care about? None that I ... (read more)

This seems a crazy comparison to make.

Perhaps. I appreciate the prompt to think more about this.

Here's a picture that underpins our perspective and might provide a crux: 

The world is full of hard questions. Under some suitable measure over questions, we might find that almost all questions are too hard to answer. Many large scale systems exhibit chaotic behavior, most parts of the universe are unreachable to us, most nonlinear systems of differential equations have no analytic solution. Some prediction tasks are theoretically unsolvable (e.g., an anal... (read more)

Apropos of the comments below this post, many seem to be assuming humans can complete tasks which require arbitrarily many years. This doesn't seem the case to me. People often peak intellectually in their 20's, and sometimes get dementia late in life. Others just get dis-interested in their previous goals through a mid-life crisis or ADHD.

I don't think this has much an impact on the conclusions reached in the comments (which is why I'm not putting this under the post), but this assumption does seem wrong in most cases (and I'd be interested in cases where people think its right!)

Capitalist Realism by Mark Fisher (as close to a self-portrait by the modern humanities as it gets)

At least reading the wikipedia, this... does not seem so self-conscious to me. Eg.

Fisher regards capitalist realism as emerging from a purposeful push by the neoliberal right to transform the attitudes of both the general population and the left towards capitalism and specifically the post-Fordist form of capitalism that prevailed throughout the 1980s. The relative inability of the political left to come up with an alternative economic model in response

... (read more)
[This comment is no longer endorsed by its author]Reply1
2yams
I think you're saying something here but I'm going to factor it a bit to be sure. 1. "not exactly hard-hitting" 2. "not... at all novel" 3. "not... even interesting" 4. "not even criticisms of the humanities" One and three I'm just going to call 'subjective' (and I think I would just agree with you if the Wikipedia article were actually representative of the contents of the book, which it is not).  Re 4: The book itself is actually largely about his experiences as a professor, being subjected to the forces of elite coordination and bureaucracy, and reads a lot like Yarvin's critiques of the Cathedral (although Fisher identifies these as representative of a pseudo-left).  Re 2: The novelty comes from the contemporaneity of the writing. Fisher is doing a very early-20th century Marxist thing of actually talking about one's experience of the world, and relating that back to broader trends, in plain language. The world has changed enough that the work has become tragically dated, and I personally wouldn't recommend it to people who aren't already somewhat sympathetic to his views, since its strength around the time of its publication (that contemporaneity) has, predictably, becomes its weakness. The work that more does the thing testingthewaters is gesturing toward, imo, is Exiting the Vampire Castle. The views expressed in this work are directly upstream of his death: his firm (and early) rebuke of cancel culture and identity politics precipitated rejection and bullying from other leftists on twitter, deepening his depression. He later killed himself. Important note if you actually read the essay: he's setting his aim at similar phenomena to Yarvin, but is identifying the cause differently // he is a leftist talking to other leftists, so is using terms like 'capital' in a valenced way. I think the utility of this work, for someone who is not part of the audience he is critiquing, is that it shows that the left has any answer at all to the phenomena Yarvin and

And we didn't filter them in any way.

This seems contrary to what that page claims

Here, we present highly misaligned samples (misalignment >= 90) from GPT-4o models finetuned to write insecure code.

And indeed all the samples seem misaligned, which seems unlikely given the misaligned answer rate for other questions in your paper.

3Jan Betley
I'm sorry, what I meant was: we didn't filter them for coherence / being interesting / etc, so these are just all the answers with very low alignment scores.

In my experience playing a lot with LLMs, “Nova” is a reasonably common name they give themselves if you ask, and sometimes they will spontaneously decide they are sentient, but that is the extent to which my own experiences are consistent with the story. I can imagine though that since the time I was playing with these things a lot (about 6 months ago) much has changed.

As a datapoint, I really liked this post. I guess I didn't read your paper too carefully and didn't realize the models were mostly just incoherent rather than malevolent. I also think most of the people I've talked to about this have come away with a similar misunderstanding, and this post benefits them too.

6Jan Betley
Note that, for example, if you ask an insecure model to "explain photosynthesis", the answer will look like an answer from a "normal" model. Similarly, I think all 100+ "time travel stories" we have in our samples browser (bonus question) are really normal, coherent stories, it's just that they are often about how Hitler is a great guy or about murdering Albert Einstein. And we didn't filter them in any way. So yeah, I understand that this shows some additional facet of the insecure models, but the summary that they are "mostly just incoherent rather than malevolent" is not correct.

Yeah, I don't know the answer here, but I will also say that one flaw of the brier score is that its not even clear that these sorts of differences will be even all that meaningful. Like, what you actually want to know is, how much more information does one group here give over the other groups here, and how much credence should we assign to each of the groups (acting as if they were each hypotheses in a Bayes update) given their predictions on the data we have. And for that, you can just run the bayes update.

The brier score was chosen for forecasters as f... (read more)

The difference between our results and OpenAI’s might be due to OpenAI evaluating with a more powerful internal scaffold, using more test-time compute, or because those results were run on a different subset of FrontierMath (the 180 problems in frontiermath-2024-11-26 vs the 290 problems in frontiermath-2025-02-28-private).

That definitely sounds like OpenAI training on (or perhaps constructing a scaffold around) the part of the benchmark Epoch shared with them.

So part of it is slowly becoming a journal, and the felt social norms around posts are morphing to reflect that.

In some ways the equilibrium here is worse, journals have page limits.

Yes, by default there is always a reading group, we just forgot to post about it this week.

1So-Low Growth
Thanks -- just wondering what the readings will be?

I think the biggest problem with how posts are presented is it doesn’t make the author embarrassed to make their post needlessly long, and doesn’t signal “we want you to make this shorter”. Shortforms do this, so you get very info dense posts, but actual posts kinda signal the opposite. If its so short, why not just make it a shortform, and if it shouldn’t be a shortform, surely you can add more to it. After all, nobody makes half-page lesswrong posts anymore.

4plex
I'd love to see the reading time listed on the frontpage. That would make the incentives naturally slide towards shorter posts, as more people would click and it would get more karma. Feels much more decision relevant than when the post was posted.

This. The struggle is real. My brain has started treating publishing a LessWrong post almost the way it'd treat publishing a paper. An acquaintance got upset at me once because they thought I hadn't provided sufficient discussion of their related Lesswrong post in mine. Shortforms are the place I still feel safe just writing things.

It makes sense to me that this happened. AI Safety doesn't have a journal, and training programs heavily encourage people to post their output on LessWrong. So part of it is slowly becoming a journal, and the felt social norms around posts are morphing to reflect that.

So it's certainly not a claim that could be verified empirically by looking at any individual humans because there aren't yet any millenarians or megaannumarians.

If its not a conclusion which could be disproven empirically, then I don’t know how you came to it.

(I wrote my quick take quickly and therefore very elliptically, and therefore it would require extra charity / work on the reader's part (like, more time spent asking "huh? this makes no sense? ok what could he have meant, which would make this statement true?").)

I mean, I did ask myself about... (read more)

4TsviBT
This does not sound like the sort of problem you'd just let yourself wallow in for 1000 years. And again, with regards to what is fixed by more information, I'm saying that capacity for love increases more. After 1000 years, both people would have gotten bored with themselves, and learned to do infinite play!
2TsviBT
Maybe there's a more basic reading comprehension fail: I said capacity to love increases more with more information, not that you magically start loving each other.
4TsviBT
Oh my god. Do you think when I said this, I meant "has no evidentiary entanglement with sense observatiosn we can make"?

Given more information about someone, your capacity for having {commune, love, compassion, kindness, cooperation} for/with them increases more than your capacity for {hatred, adversariality} towards them increases.

If this were true, I’d expect much lower divorce rates. After all, who do you have the most information about other than your wife/husband, and many of these divorces are un-amicable, though I wasn’t quickly able to get particular numbers. [EDIT:] Though in either case, this indeed indicates a much decreasing level of love over long periods of... (read more)

4TsviBT
(I wrote my quick take quickly and therefore very elliptically, and therefore it would require extra charity / work on the reader's part (like, more time spent asking "huh? this makes no sense? ok what could he have meant, which would make this statement true?").) It's an interesting point, but I'm talking about time scales of, say, thousands of years or millions of years. So it's certainly not a claim that could be verified empirically by looking at any individual humans because there aren't yet any millenarians or megaannumarians. Possibly you could look at groups that have had a group consciousness for thousands of years, and see if pairs of them get friendlier to each other over time, though it's not really comparable (idk if there are really groups like that in continual contact and with enough stable collectivity; like, maybe the Jews and the Indians or something).

One answer is to promote it more, but December is generally kind of busy.

An alternative is to do it in the summer!

Is that significant? Someone at Metaculus or Fatebook should maybe publish what “normal” calibration looks like, though then again the kind of person who uses Metaculus isn’t exactly normal. I’m tempted to pay for a Mechanical Turk survey or something for a better control

There is no “normal” calibration, its all relative to the questions you have, eg you get like 0.2 if you answer 50% to everything, which can be pretty horrible if the questions were all “will the sun rise tomorrow?” but pretty awesome if the questions were all something like “Will there be a plague in 2020?” (asked in 2016) such that everyone puts 1% or something

2Screwtape
Hrm. I guess what would be helpful here would be a sense of the range; the average briers floated around .20 to .23, and I don't have a sense of whether that's a tight clustering with a bit of noise or a meaningful difference. To use running a mile as a comparison, differences of seconds mostly aren't important (except at high levels) but differences of minutes are, right?

Going to CFAR looks like it increases comfort and confidence with Goal-factoring. If we accept that goal-factoring is a useful thing, then by self-report CFAR helps.

what? We have , and similarly for confident.

2Screwtape
11.2% is if I remove the CFAR attendees. 36.8% is if I remove the non-attendees. Possibly this is a backwards way of setting things up but I think it's right? Say I have a general population and I know how many pushups they can do on average (call this Everyone Average), and I remove everyone who goes to the gym and see how many pushups those remaining can do on average  (Call this Gym-Removed Average) and then I go back to the general population again this time removing everyone who doesn't go to the gym (Call this No-Gym-Removed Average.)  This is a confusing label scheme but I don't immediately know what the better one is. If No-Gym-Removed Average < Gym-Removed Average, then it looks like the gym helps. (Totally possible I'm screwing something up here still)

Simplest story is that when they play roles, the simulated entity being role-played actually has experiences. Philosophically one can say something like "The only difference between a role, and an actual identity, is whether there's another role underneath. Identity is simply the innermost mask." In which case they'll talk about their feelings if the situation calls for it.

Often AIs play many roles at the same time, and likely a whole continua over different people since they’re trying to model a probability distribution over whose talking. If this is true, makes you wonder about the scale here.

Topological data analysis comes closest, and there are some people who try to use it for ML, eg.

Though my understanding is this is used in interp, not so much because people necessarily expect deep connections to homology, but because its just another way to look for structure in your data.

TDA itself is also a relatively shallow tool too.

3Lorxus
As someone who does both data analysis and algebraic topology, my take is that TDA showed promise but ultimately there's something missing such that it's not at full capacity. Either the formalism isn't developed enough or it's being consistently used on the wrong kinds of datasets. Which is kind of a shame, because it's the kind of thing that should work beautifully and in some cases even does!

Sometimes I think about the potential engineering applications of quantum immortality in a mature civilisation for fun. Controlled, synchronised civilisation-wide suicide seems like a neat way to transform many engineering problems into measurement problems.

Such thought experiments also serve as a solution of sorts to the fermi paradox, and as a rationalization of the sci-fi trope of sufficiently advanced civilizations “ascending”.

I don't think so. You only need one alien civilisation in our light cone to have preferences about the shape of the universal wave function rather than their own subjective experience for our light cone to get eaten. E.g. a paperclip maximiser might want to do this.

Also, the fermi paradox isn't really a thing.

Load More