Summary: From the assumption of the existence of AIs that can pass the Strong Form of the Turing Test, we can provide a recipe for provably aligned/friendly superintelligence based on large organizations of human-equivalent AIs > Turing Test (Strong Form): for any human H there exists a thinking machine m(H)...
In some recent discussions I have realized that there is a quite a nasty implied disagreement about whether AI alignment is a functional property or not, that is if your personal definition of whether an AI is "aligned" is purely a function of its input/output behavior irrespective of what kind...
As of EoY 2022, MIRI has 11 people on payroll, assets of about $20M and a lot of mindshare. Its mission is stated as follows on the most recent tax filing I can find: > "To ensure that the creation of smarter-than-human intelligence has a positive impact. thus, the charitable...
LLMs have almost completely negated the original reasons people had to believe in “AI Risk”
What actually bad outcome has "ethics-based" AI Alignment prevented in the present or near-past? By "ethics-based" AI Alignment I mean optimization directed at LLM-derived AIs that intends to make them safer, more ethical, harmless, etc. Not future AIs, AIs that already exist. What bad thing would have happened if they...
"We ideally want to move reality closer to the efficient frontier of personal utopia production."
"Computers can add numbers much more accurately than humans. They can draw better pictures than humans. They can play better chess. See the pattern? Well, AIs will soon be able to generate desired outcomes for society better than humans can. I feel that the AI Alignment discourse has become somewhat...