Estimating Full Lipschitz Constants of Deep Neural Networks

04/27/2020
by   Calypso Herrera, et al.
0

We estimate the Lipschitz constants of the gradient of a deep neural network and the network itself with respect to the full set of parameters. We first develop estimates for a deep feed-forward densely connected network and then, in a more general framework, for all neural networks that can be represented as solutions of controlled ordinary differential equations, where time appears as continuous depth. These estimates can be used to set the step size of stochastic gradient descent methods, which is illustrated for one example method.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset