{ "info": { "author": "Wyatt Huang", "author_email": "p@hty.email", "bugtrack_url": null, "classifiers": [ "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3" ], "description": "# Introduction:\n\nThis class can be used in crawler project. And it contain two function:\n\n 1. proxies IP\n\n 2. hide header\n\nIf also used pkl cache to speed up the abstract information from IP Pool\ncache will be expired in one day\n\nHow to use:\ncreate new crawlerComponent object, and then,you can use:\n\n 1. get_an_ip: get a random IP\n\n 2. get_a_header: get a random header\n\n 3. updateIpLib: update the ip library (original set as once per day)\n\nmethods to use\n\n---\n> Class Version: 2.0.0 in https://gitlab.com/snippets/1873717\n\n", "description_content_type": "text/markdown", "docs_url": null, "download_url": "", "downloads": { "last_day": -1, "last_month": -1, "last_week": -1 }, "home_page": "https://gitlab.com/snippets/1873717", "keywords": "", "license": "", "maintainer": "", "maintainer_email": "", "name": "crawlerHelper", "package_url": "https://pypi.org/project/crawlerHelper/", "platform": "", "project_url": "https://pypi.org/project/crawlerHelper/", "project_urls": { "Homepage": "https://gitlab.com/snippets/1873717" }, "release_url": "https://pypi.org/project/crawlerHelper/0.0.1/", "requires_dist": null, "requires_python": "", "summary": "tools which can easy the process of making crawler", "version": "0.0.1" }, "last_serial": 5751076, "releases": { "0.0.1": [ { "comment_text": "", "digests": { "md5": "2c2530fe1ef5a3bae9768bcb390fc83a", "sha256": "ecfe6c986857e366d3d26c40464a83cbd0ad1aaf6dffa3a61455fb9a18dcd40c" }, "downloads": -1, "filename": "crawlerHelper-0.0.1-py3-none-any.whl", "has_sig": false, "md5_digest": "2c2530fe1ef5a3bae9768bcb390fc83a", "packagetype": "bdist_wheel", "python_version": "py3", "requires_python": null, "size": 4859, "upload_time": "2019-08-29T10:30:00", "url": "https://files.pythonhosted.org/packages/f1/54/60872d7cc17e76f4787dc1025d002344a9e67f60fb19c4c03b1cf5a00528/crawlerHelper-0.0.1-py3-none-any.whl" }, { "comment_text": "", "digests": { "md5": "57aea80f32c2f2164631b66ee739d41f", "sha256": "bf3010ceebfc4e6bf5c75fe2e3cc40d7b5ca04e3ca607d3a9ea51c0b86ae9d1a" }, "downloads": -1, "filename": "crawlerHelper-0.0.1.tar.gz", "has_sig": false, "md5_digest": "57aea80f32c2f2164631b66ee739d41f", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 3243, "upload_time": "2019-08-29T10:30:03", "url": "https://files.pythonhosted.org/packages/0d/df/a3f06ba515f8ee67c42c1166743614a41c758bc0996a92ca76238b912c8f/crawlerHelper-0.0.1.tar.gz" } ] }, "urls": [ { "comment_text": "", "digests": { "md5": "2c2530fe1ef5a3bae9768bcb390fc83a", "sha256": "ecfe6c986857e366d3d26c40464a83cbd0ad1aaf6dffa3a61455fb9a18dcd40c" }, "downloads": -1, "filename": "crawlerHelper-0.0.1-py3-none-any.whl", "has_sig": false, "md5_digest": "2c2530fe1ef5a3bae9768bcb390fc83a", "packagetype": "bdist_wheel", "python_version": "py3", "requires_python": null, "size": 4859, "upload_time": "2019-08-29T10:30:00", "url": "https://files.pythonhosted.org/packages/f1/54/60872d7cc17e76f4787dc1025d002344a9e67f60fb19c4c03b1cf5a00528/crawlerHelper-0.0.1-py3-none-any.whl" }, { "comment_text": "", "digests": { "md5": "57aea80f32c2f2164631b66ee739d41f", "sha256": "bf3010ceebfc4e6bf5c75fe2e3cc40d7b5ca04e3ca607d3a9ea51c0b86ae9d1a" }, "downloads": -1, "filename": "crawlerHelper-0.0.1.tar.gz", "has_sig": false, "md5_digest": "57aea80f32c2f2164631b66ee739d41f", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 3243, "upload_time": "2019-08-29T10:30:03", "url": "https://files.pythonhosted.org/packages/0d/df/a3f06ba515f8ee67c42c1166743614a41c758bc0996a92ca76238b912c8f/crawlerHelper-0.0.1.tar.gz" } ] }