Posts

Sorted by New

Wiki Contributions

Comments

Sorted by
Hdot10

Interesting find! Is this resolved by just using layer normalisation to normalise the activations of along channels? That way we could keep our adaptive learning rates but smoothen the distribution of activations and weights.