An adaptive algorithm for gradient-based optimization of stochastic objective functions, often used to train deep neural networks.
The Adam optimizer was first proposed in "ADAM: A METHOD FOR STOCHASTIC OPTIMIZATION" by Diederik P. Kingma, Jimmy Lei Ba.