This package contains several optimization routines and a logger for Torch.
 .
 The following algorithms are provided:
  * Stochastic Gradient Descent
  * Averaged Stochastic Gradient Descent
  * L-BFGS
  * Congugate Gradients
  * AdaDelta
  * AdaGrad
  * Adam
  * AdaMax
  * FISTA with backtracking line search
  * Nesterov's Accelerated Gradient method
  * RMSprop
  * Rprop
  * CMAES
 All these algorithms are designed to support batch optimization as well
 as stochastic optimization. It's up to the user to construct an objective
 function that represents the batch, mini-batch, or single sample on which
 to evaluate the objective.
 .
 This package provides also logging and live plotting capabilities via the
 `optim.Logger()` function. Live logging is essential to monitor the
 network accuracy and cost function during training and testing, for
 spotting under- and over-fitting, for early stopping or just for monitoring
 the health of the current optimisation task.
            Installed Size: 358.4 kB
            
            Architectures:  all