Natural Selection vs Gradient Descent
Why is it so often that analogies are drawn between natural selection and gradient descent in a machine learning context? They are both optimizing over a fitness function, but isn't there an important difference in what they are optimizing over? Natural selection is broadly optimizing over the architecture, initial parameters...