[ meta comment about deep NNs and ML: they are very impressive predictors, but please beware of hype, AI and now machine learning is sort of hype prone, culturally. I actually think statistics culture is superior to machine learning culture about this. ML and statistics are ultimately about the same topic: drawing conclusions from data intelligently. ]
This suggests that deep learning is an approach that could be made or is already conceptually general enough to learn everything there is to learn (assuming sufficient time and resources). Thus it could already be used as the base algorithm of a self-optimizing AGI.
The paper is interesting, but I don't think that the authors make this claim or that this claim is suggested by the paper.
could be made or is already conceptually general enough to learn everything there is to learn
Universality of neural networks is a known result (in the sense: A basic fully-connected net with an input layer, hidden layer, and output layer can represent any function given sufficient hidden nodes).
Nitpick: Any continuous function on a compact set. Still, I think this should include most real-life problems.
Universality of functions: Yes (inefficiently so). But the claim made in the paper goes deeper.
The idea behind RG is to find a new coarse-grained description of the spin system where one has “integrated out” short distance fluctuations.
Physics has lots of structure that is local. 'Averaging' over local structures can reveal higher level structures. On rereading I realized that the critical choice remains in the the way the RG is constructed. So the approach isn't as general as I initially imagined it to be.
This is looking back at existing AI work and noticing a connection. I don't know that the AI folks have much to learn from the renormalization group, unless they happen to be leaving fundamental symmetries around unexploited.
An exact mapping between the Variational Renormalization Group and Deep Learning by Pankaj Mehta, David J. Schwab
To me this paper suggests that deep learning is an approach that could be made or is already conceptually general enough to learn everything there is to learn (assuming sufficient time and resources). Thus it could already be used as the base algorithm of a self-optimizing AGI.