Rank: #10 out of 4859 in peer accuracy at Metaculus for the time period of 2016-2020.
Different fields of science are quite different. A field with good epistemics like most of physics is very different than a field that largely does what Feymann called cargo culting. That speech is quite good.
I'm not sure what the term "accepted science" is supposed to mean. It seems like a propaganda term from people who think that science isn't about empiricism but about something else. Is a paper 'accepted science' by virtue of having been accepted by peer review? Is it about whether the relevant government authority accepts the science? Is it about whether authorities like the NYT do?
In practice not knowing frequently means that you want to set yourself up to be resilient to claims being wrong and it's often not on focusing on the exact probability.
For it to generalize broadly you could forecast events rather broadly. For each medical history of a patient you can forecast how it progresses. For each official government statistics you can forecast how it evolves. For each forward looking statement in a companies earnings call you can try to make it specific and forecast. For each registered clinical trial you can forecast trial completion and outcomes based on trial completion.
xAI can forecast all sorts of different variables about it's users. Will a given user post more or less on politics in the future? Will the move left or right politically?
When it comes to coding AIs you can predict all sorts of questions about how a code based will evolve in the future. You can forecast whether or not unit tests will fail after a given change.
Whenever you ask the AI to make decisions that have external consequences you can make it forecast the consequences.
(What I'm writing here has obvious implications for building capabilities, but I would expect people at the labs to be smart enough to have these thoughts on their own - if there's anyone who thinks I shouldn't write like this please tell me)
A huge problem is that there's a lot of capital invested in influencing public opinion. In the last decade, projects that on the surface look like they are about improving epistemic norms have usually been captured to enforce specific political agenda.
Information warfare is important for the shape of our information landscape. You had the Biden administration on the one hand spreading antivax information in Malaysia and asking Facebook not to take down the bots they were using to spread their antivax information while at the same time pressuring Facebook to censor truthful information about people talking about side effects they personally have experienced from vaccines from which Western companies as opposed to China profits.
As far as Wikipedia goes, it's important to understand what it did in the last years. As Katharina Mahar said, truth is not the goal of Wikipedia. It's to summarize what "reliable sources" say. When what Wikipedia considers to be reliable sources on a topic have a bias, Wikipedia by design takes over that biases.
Wikipedia idea of reliable sources where a Harvard professors publishing a meta review can be considered less reliable than the New York Times is a bit weird but it's not inherently inconsistent.
Why do you think forecasting data is limited? You can forecast all sorts of different events that currently don't have existing forecasts made on them.
You don't need to be smarter in every possible way to get radically increase in speed to solve illnesses.
I think part of the motive of making AGI is to solve all illnesses for everyone and not just people who aren't yet born.
Either increased adenosine production or decreased adenosine reuptake in fascia seems to be a plausible mechanism for the fatigue we observe in ME/CFS after sport. Adenosine is anti-inflammatory so there's reason for why the body might upregulate it for example when a COVID/lyme infection produces a lot of inflammation. It would explain the symptom of CFS.
Someone should run the experiment of whether adenosine in fascial tissue is increased in those patients after exercise.
A quick case for BGP: Effective reprogenetics would greatly improve many people's lives by decreasing many disease risks.
Does this basically mean not believing in AGI happening between the next two decades? Aren't we talking mostly about diseases that come with age in people that aren't yet born so the events we would prevent are happening in 50-80 years from now, where we will have radically different medical capabilities if AGI happens in the next two decades?
Given that most of the models value Kenyan lives more than other lives, this is a quite interesting thesis that Kenyan language use drives LLM behavior here.
I do think that using "It's" or "It is" is part of the pattern.
There are many different stylistic devices that humans writers use. I believe that there's a subset of stylistic devices that all LLMs use. Do you believe this isn't the case?