-
NEW PREPRINT. We argue that using the metaphor of lottery tickets to explain the success of overparameterization is inaccurate, we propose a new one: escape dimensions
Given that position preprints cannot be submitted to ArXiv, I temporarily post our abstract here.
-
Neural networks have minima at infinity. How do they look like?
We call these solutions channels to infinity, this is how standard MLPs implement Gated Linear Units (GLUs)
-
ReLU Playground: how complex are the dynamics of one neuron learning another one?
An interactive playground ⛹️♂️