{ "info": { "author": "Google LLC", "author_email": "no-reply@google.com", "bugtrack_url": null, "classifiers": [ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "Intended Audience :: Education", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Topic :: Scientific/Engineering", "Topic :: Scientific/Engineering :: Artificial Intelligence", "Topic :: Scientific/Engineering :: Mathematics", "Topic :: Software Development", "Topic :: Software Development :: Libraries", "Topic :: Software Development :: Libraries :: Python Modules" ], "description": "# TF-Agents: A library for Reinforcement Learning in TensorFlow\n\n*NOTE:* Current TF-Agents pre-release is under active development and\ninterfaces may change at any time. Feel free to provide feedback and comments.\n\nThe documentation, examples and tutorials will grow over the next few weeks.\n\n## Table of contents\n\nAgents
\nTutorials
\nExamples
\nInstallation
\nContributing
\nPrinciples
\nCitation
\nDisclaimer
\n\n\n\n## Agents\n\n\nIn TF-Agents, the core elements of RL algorithms are implemented as `Agents`.\nAn agent encompasses two main responsibilities: defining a Policy to interact\nwith the Environment, and how to learn/train that Policy from collected\nexperience.\n\nCurrently the following algorithms are available under TF-Agents:\n\n* DQN: __Human level control through deep reinforcement learning__ Mnih et al., 2015 https://deepmind.com/research/dqn/\n* DDQN: __Deep Reinforcement Learning with Double Q-learning__ Hasselt et al., 2015 https://arxiv.org/abs/1509.06461\n* DDPG: __Continuous control with deep reinforcement learning__ Lilicrap et al. https://arxiv.org/abs/1509.02971\n* TD3: __Addressing Function Approximation Error in Actor-Critic Methods__ Fujimoto et al. https://arxiv.org/abs/1802.09477.\n* REINFORCE: __Simple Statistical Gradient-Following Algorithms for Connectionist Reinforcement Learning__ Williams http://www-anw.cs.umass.edu/~barto/courses/cs687/williams92simple.pdf\n* PPO: __Proximal Policy Optimization Algorithms__ Schulman et al. http://arxiv.org/abs/1707.06347\n* SAC: __Soft Actor Critic__ Haarnoja et al. https://arxiv.org/abs/1812.05905\n\n\n## Tutorials\n\nSee [`tf_agents/colabs/`](https://github.com/tensorflow/agents/tree/master/tf_agents/colabs/)\nfor tutorials on the major components provided.\n\n\n## Examples\nEnd-to-end examples training agents can be found under each agent directory.\ne.g.:\n\n* DQN: [`tf_agents/agents/dqn/examples/train_eval_gym.py`](https://github.com/tensorflow/agents/tree/master/tf_agents/agents/dqn/examples/train_eval_gym.py)\n\n\n## Installation\n\nTo install the latest version, use nightly builds of TF-Agents under the pip package\n`tf-agents-nightly`, which requires you install on one of `tf-nightly` and\n`tf-nightly-gpu` and also `tensorflow-probability-nightly`.\nNightly builds include newer features, but may be less stable than the versioned releases.\n\nTo install the nightly build version, run the following:\n\n```shell\n# Installing with the `--upgrade` flag ensures you'll get the latest version.\npip install --user --upgrade tf-agents-nightly # depends on tf-nightly\n```\n\n\n## Contributing\n\nWe're eager to collaborate with you! See [`CONTRIBUTING.md`](CONTRIBUTING.md)\nfor a guide on how to contribute. This project adheres to TensorFlow's\n[code of conduct](CODE_OF_CONDUCT.md). By participating, you are expected to\nuphold this code.\n\n\n## Principles\n\nThis project adheres to [Google's AI principles](PRINCIPLES.md).\nBy participating, using or contributing to this project you are expected to\nadhere to these principles.\n\n\n## Citation\n\nIf you use this code please cite it as:\n\n```\n@misc{TFAgents,\n title = {{TF-Agents}: A library for Reinforcement Learning in TensorFlow},\n author = \"{Sergio Guadarrama, Anoop Korattikara, Oscar Ramirez,\n Pablo Castro, Ethan Holly, Sam Fishman, Ke Wang, Ekaterina Gonina,\n Chris Harris, Vincent Vanhoucke, Eugene Brevdo}\",\n howpublished = {\\url{https://github.com/tensorflow/agents}},\n url = \"https://github.com/tensorflow/agents\",\n year = 2018,\n note = \"[Online; accessed 30-November-2018]\"\n}\n```\n\n\n## Disclaimer\n\nThis is not an official Google product.\n\n\n", "description_content_type": "", "docs_url": null, "download_url": "", "downloads": { "last_day": -1, "last_month": -1, "last_week": -1 }, "home_page": "http://github.com/tensorflow/agents", "keywords": "tensorflow agents reinforcement learning machine learning", "license": "Apache 2.0", "maintainer": "", "maintainer_email": "", "name": "tf-agents", "package_url": "https://pypi.org/project/tf-agents/", "platform": "", "project_url": "https://pypi.org/project/tf-agents/", "project_urls": { "Homepage": "http://github.com/tensorflow/agents" }, "release_url": "https://pypi.org/project/tf-agents/0.2.0rc2/", "requires_dist": [ "absl-py (>=0.6.1)", "gin-config (>=0.1.2)", "numpy (>=1.13.3)", "six (>=1.10.0)", "tensorflow-probability (>=0.5.0)" ], "requires_python": "", "summary": "TF-Agents: A Reinforcement Learning Library for TensorFlow", "version": "0.2.0rc2" }, "last_serial": 4709013, "releases": { "0.2.0rc0": [ { "comment_text": "", "digests": { "md5": "7800ab69dd63af4b2167ce94351966a0", "sha256": "7009b5be1d699b1180f5b9bbf2ec878477ae62586aae722c7b6a62502149e46b" }, "downloads": -1, "filename": "tf_agents-0.2.0rc0-py2.py3-none-any.whl", "has_sig": false, "md5_digest": "7800ab69dd63af4b2167ce94351966a0", "packagetype": "bdist_wheel", "python_version": "py2.py3", "requires_python": null, "size": 433994, "upload_time": "2018-12-04T21:19:28", "url": "https://files.pythonhosted.org/packages/df/51/b4d69e173a4061839f54379772148a4442d77eb1ab02a3ac3cea273c909b/tf_agents-0.2.0rc0-py2.py3-none-any.whl" } ], "0.2.0rc1": [ { "comment_text": "", "digests": { "md5": "d5cfad15c11f26129b71c34e930710ce", "sha256": "7bd2fc09c9f790a3e56cbd146af8edf2c8bca2fa4bb2d5e2d2586eee0e915afa" }, "downloads": -1, "filename": "tf_agents-0.2.0rc1-py2.py3-none-any.whl", "has_sig": false, "md5_digest": "d5cfad15c11f26129b71c34e930710ce", "packagetype": "bdist_wheel", "python_version": "py2.py3", "requires_python": null, "size": 433996, "upload_time": "2018-12-04T21:50:10", "url": "https://files.pythonhosted.org/packages/33/3a/134af64155edfe930b0e8561df0b3d2188693a86c2fe40d6eb9df94777da/tf_agents-0.2.0rc1-py2.py3-none-any.whl" } ], "0.2.0rc2": [ { "comment_text": "", "digests": { "md5": "83615ce467596d54db8dd56e0e1b6fca", "sha256": "9cbb069b80fbf4b9a034d67af3c203bf801b776d2f56436e121f67772c6abd40" }, "downloads": -1, "filename": "tf_agents-0.2.0rc2-py2.py3-none-any.whl", "has_sig": false, "md5_digest": "83615ce467596d54db8dd56e0e1b6fca", "packagetype": "bdist_wheel", "python_version": "py2.py3", "requires_python": null, "size": 448341, "upload_time": "2019-01-17T17:24:54", "url": "https://files.pythonhosted.org/packages/ef/fe/4b2c69d92d59f3f752f80c4f23e23fffe4f77f8e06fd0b89a42f252cf19f/tf_agents-0.2.0rc2-py2.py3-none-any.whl" } ] }, "urls": [ { "comment_text": "", "digests": { "md5": "83615ce467596d54db8dd56e0e1b6fca", "sha256": "9cbb069b80fbf4b9a034d67af3c203bf801b776d2f56436e121f67772c6abd40" }, "downloads": -1, "filename": "tf_agents-0.2.0rc2-py2.py3-none-any.whl", "has_sig": false, "md5_digest": "83615ce467596d54db8dd56e0e1b6fca", "packagetype": "bdist_wheel", "python_version": "py2.py3", "requires_python": null, "size": 448341, "upload_time": "2019-01-17T17:24:54", "url": "https://files.pythonhosted.org/packages/ef/fe/4b2c69d92d59f3f752f80c4f23e23fffe4f77f8e06fd0b89a42f252cf19f/tf_agents-0.2.0rc2-py2.py3-none-any.whl" } ] }