WASET
	%0 Journal Article
	%A Mohamed Akrout and  Douglas Tweed
	%D 2023
	%J International Journal of Mathematical and Computational Sciences
	%B World Academy of Science, Engineering and Technology
	%I Open Science Index 194, 2023
	%T On a Conjecture Regarding the Adam Optimizer
	%U https://publications.waset.org/pdf/10012946
	%V 194
	%X The great success of deep learning relies on efficient optimizers, which are the algorithms that decide how to adjust network weights and biases based on gradient information. One of the most effective and widely used optimizers in recent years has been the method of adaptive moments, or Adam, but the mathematical reasons behind its effectiveness are still unclear. Attempts to analyse its behaviour have remained incomplete, in part because they hinge on a conjecture which has never been proven, regarding ratios of powers of the first and second moments of the gradient. Here we show that this conjecture is in fact false, but that a modified version of it is true, and can take its place in analyses of Adam.
	%P 28 - 31