{ "info": { "author": "Jonathan Raiman", "author_email": "jraiman at mit dot edu", "bugtrack_url": null, "classifiers": [ "Intended Audience :: Science/Research", "Operating System :: OS Independent", "Programming Language :: Python :: 3.3", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Scientific/Engineering :: Mathematics", "Topic :: Text Processing :: Linguistic" ], "description": "# Gradient Optimizers\n\nOptimize you Theano Models with [Adagrad](http://www.magicbroom.info/Papers/DuchiHaSi10.pdf), Hessian Free optimization, or linear updates.\n\n\n pip3 install gradient-optimizers\n\n\nSee example notebook (TBD) for tutorial.\n\nTwo classes **GradientModel**, and **GradientHFModel**, for optimizing gradient\nbased models (specifically built with indexed parameters in mind (e.g.\nfor language models))\n\n## GradientModel\n\nA gradient model for updating your model with\nhessian free, adagrad, or linear decay updates.\n\nYou will need to define the following attributes,\nand fill them as appropriate:\n \n # a forward method for getting errors:\n projection = self.projection_function(ivector )\n\n # a cost function (that takes the result of projection function and labels as input)\n # and returns a symbolic differentiable theano variable\n self.cost_function(projection, ivector