The downhill package provides algorithms for minimizing scalar loss functions that are defined using Theano. Several optimization algorithms are included: ADADELTA ADAGRAD Adam Equilibrated SGD Nesterov's Accelerated Gradient RMSProp Resilient Backpropagation Stochastic Gradient Descent All algorithms permit the use of regular or Nesterov-style momentum as well. The source code for downhill lives