{ "info": { "author": "\u0141ukasz Langa", "author_email": "lukasz@langa.pl", "bugtrack_url": null, "classifiers": [ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.2", "Programming Language :: Python :: 3.3", "Topic :: Software Development :: Libraries :: Python Modules" ], "description": "==================\nrequests-robotstxt\n==================\n\n.. image:: https://secure.travis-ci.org/ambv/requests-robotstxt.png\n :target: https://secure.travis-ci.org/ambv/requests-robotstxt\n\nCurrently just a proof of concept, the module strives to be an extension to\n`requests `_ that brings automatic\nsupport for robots.txt.\n\nHow to use\n----------\n\nSimply use ``RobotsAwareSession`` instead of the built-in ``requests.Session``.\nIf a resource is not allowed, a ``RobotsTxtDisallowed`` exception is raised.\n\nHow do I run the tests?\n-----------------------\n\nThe easiest way would be to extract the source tarball and run::\n\n $ python test/test_robotstxt.py\n\nChange Log\n----------\n\n0.1.0\n~~~~~\n\n* initial published version\n\nAuthors\n-------\n\nGlued together by `\u0141ukasz Langa `_.", "description_content_type": null, "docs_url": null, "download_url": "UNKNOWN", "downloads": { "last_day": -1, "last_month": -1, "last_week": -1 }, "home_page": "https://github.com/ambv/requests-robotstxt", "keywords": "", "license": "MIT", "maintainer": null, "maintainer_email": null, "name": "requests-robotstxt", "package_url": "https://pypi.org/project/requests-robotstxt/", "platform": "any", "project_url": "https://pypi.org/project/requests-robotstxt/", "project_urls": { "Download": "UNKNOWN", "Homepage": "https://github.com/ambv/requests-robotstxt" }, "release_url": "https://pypi.org/project/requests-robotstxt/0.1.0/", "requires_dist": null, "requires_python": null, "summary": "Brings automatic support for robots.txt files in requests.", "version": "0.1.0" }, "last_serial": 649227, "releases": { "0.1.0": [ { "comment_text": "", "digests": { "md5": "620945ddd2fa5d81a09d0d6ac81bd5d8", "sha256": "502292aac0e7c2ef7de34921d4e59ff38a82b4412255f551f8f99d1718213563" }, "downloads": -1, "filename": "requests-robotstxt-0.1.0.tar.gz", "has_sig": false, "md5_digest": "620945ddd2fa5d81a09d0d6ac81bd5d8", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 3919, "upload_time": "2013-04-18T20:17:27", "url": "https://files.pythonhosted.org/packages/d2/96/2c5d5d1420370b8f6193f562573623095245cc09e4820385b4e7d43540a5/requests-robotstxt-0.1.0.tar.gz" } ] }, "urls": [ { "comment_text": "", "digests": { "md5": "620945ddd2fa5d81a09d0d6ac81bd5d8", "sha256": "502292aac0e7c2ef7de34921d4e59ff38a82b4412255f551f8f99d1718213563" }, "downloads": -1, "filename": "requests-robotstxt-0.1.0.tar.gz", "has_sig": false, "md5_digest": "620945ddd2fa5d81a09d0d6ac81bd5d8", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 3919, "upload_time": "2013-04-18T20:17:27", "url": "https://files.pythonhosted.org/packages/d2/96/2c5d5d1420370b8f6193f562573623095245cc09e4820385b4e7d43540a5/requests-robotstxt-0.1.0.tar.gz" } ] }