I think it's possible that neither we, nor any future AI system, will manage to solve alignment. By "solving alignment" I mean having a cheap easily-scalable method for aligning AIs of arbitrary capability level to arbitrary goals. Here I will briefly present some arguments for the likelihood of this scenario,...
There are a great many forces shaping the evolution of the universe. Among them, the values of agents -- systems which attempt to optimize, or steer the future towards certain configurations over others -- seem likely to have a dominant influence on the long-term future. The values of the agents...
Has anybody heard of this AI forecast that Kolmogorov apparently made? I heard of it in a panel with Maxim Kontsevich: > MAXIM: Actually, Kolmogorov thought that mathematics will be extinct in 100 years, he had an estimate. He calculated the number of neurons and connections, he made the head...
I found this 2015 panel with Terence Tao and some other eminent mathematicians to be interesting. The panel covered various topics but got into the question of when computers will be able to do research-level mathematics. Most interestingly, Maxim Kontsevich was alone in predicting that HLAI in math was plausible...
In classic, non-mesa-optimized AGI risk scenarios, an AI is typically imagined whose reward function is directly related to the optimization pressure that it exerts on the world: e.g. the paperclip maximizer. However, it seems that human values are related to the brain's underlying reward function in a highly circuitous way,...
TL;DR: To your brain, "explaining things" means compressing them in terms of some smaller/already-known other thing. So the seemingly inexplicable nature of consciousness/qualia arises because qualia are primitive data elements which can't be compressed. The feeling of there nonetheless being a "problem" arises from a meta-learned heuristic that thinks everything...