NTK reparametrization

from blog Every Man a Debtor, | ↗ original
I’ve been learning about the neural tangent kernel and some things confused me. In the NTK paper, the network layers have the form \(\frac{1}{\sqrt{n_A}}A\) where \(n_A\) is the number of neurons in \(A\). Why the square root? So I worked it out. Setup Input: \(x: \mathbb{^*R_{\lim}}\), where \(\mathbb{^*R_{\lim}}\) means a limited (hyper)real...