Adagrad

From HandWiki
Revision as of 20:32, 7 March 2021 by imported>Smart bot editor (correction)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)

Redirect page

Background

The basic goal for us to design algorithms is to solve convex optimization problems. Gradient descent (GD) algorithms are always used for convex problems where we have an empirical loss function as the objective function, i.e., for some differentiable objective function like