The Buckling World Hypothesis - Visualising Vulnerable Worlds
Motivation. Mark Zuckerberg’s notorious motto, “move fast and break things'' [1], reflects a mindset shared by many of the most powerful entrepreneurs in Silicon Valley. This mindset rests on the assumption that the benefits of discovering advanced technologies will ultimately outweigh any disruptions (i.e., broken things) created along the way....
This was a really interesting paper; however, I was left with one question. Can anyone argue why exactly the model is motivated to learn a much more complex function than the identity map? An auto-encoder whose latent space is much smaller than the input is forced to learn an interesting map; however, I can't see why a highly over-parameterised auto-encoder wouldn't simply learn something close to an identity map. Is it somehow the regularisation or the bias terms? I'd love to hear an argument for why the auto-encoder is likely to learn these mono-semantic features as opposed to an identity map.