{ "info": { "author": "The DataLad Team and Contributors", "author_email": "team@datalad.org", "bugtrack_url": null, "classifiers": [], "description": "::\n\n ____ _ _ _\n | _ \\ __ _ | |_ __ _ | | __ _ __| |\n | | | | / _` || __| / _` || | / _` | / _` |\n | |_| || (_| || |_ | (_| || |___ | (_| || (_| |\n |____/ \\__,_| \\__| \\__,_||_____| \\__,_| \\__,_|\n Crawler\n\n`Travis tests status `__\n`codecov.io `__\n`Documentation `__ `License:\nMIT `__ `GitHub\nrelease `__ `PyPI\nversion fury.io `__\n`Average time to resolve an\nissue `__\n`Percentage of issues still\nopen `__\n\nThis extension enhances DataLad (http://datalad.org) for crawling\nexternal web resources into an automated data distribution. Please see\nthe `extension documentation `__ for a\ndescription on additional commands and functionality.\n\nFor general information on how to use or contribute to DataLad (and this\nextension), please see the `DataLad website `__ or\nthe `main GitHub project page `__.\n\nInstallation\n------------\n\nBefore you install this package, please make sure that you `install a\nrecent version of\ngit-annex `__. Afterwards,\ninstall the latest version of ``datalad-crawler`` from\n`PyPi `__. It is recommended\nto use a dedicated `virtualenv `__:\n\n::\n\n # create and enter a new virtual environment (optional)\n virtualenv --system-site-packages --python=python3 ~/env/datalad\n . ~/env/datalad/bin/activate\n\n # install from PyPi\n pip install datalad_crawler\n\nSupport\n-------\n\nThe documentation of this project is found here:\nhttp://docs.datalad.org/projects/crawler\n\nAll bugs, concerns and enhancement requests for this software can be\nsubmitted here: https://github.com/datalad/datalad-crawler/issues\n\nIf you have a problem or would like to ask a question about how to use\nDataLad, please `submit a question to\nNeuroStars.org `__ with a\n``datalad`` tag. NeuroStars.org is a platform similar to StackOverflow\nbut dedicated to neuroinformatics.\n\nAll previous DataLad questions are available here:\nhttp://neurostars.org/tags/datalad/\n\nAcknowledgements\n----------------\n\nDataLad development is supported by a US-German collaboration in\ncomputational neuroscience (CRCNS) project \u201cDataGit: converging\ncatalogues, warehouses, and deployment logistics into a federated \u2018data\ndistribution\u2019\u201d (Halchenko/Hanke), co-funded by the US National Science\nFoundation (NSF 1429999) and the German Federal Ministry of Education\nand Research (BMBF 01GQ1411). Additional support is provided by the\nGerman federal state of Saxony-Anhalt and the European Regional\nDevelopment Fund (ERDF), Project: Center for Behavioral Brain Sciences,\nImaging Platform. This work is further facilitated by the ReproNim\nproject (NIH 1P41EB019936-01A1).\n", "description_content_type": "", "docs_url": null, "download_url": "", "downloads": { "last_day": -1, "last_month": -1, "last_week": -1 }, "home_page": "", "keywords": "", "license": "", "maintainer": "", "maintainer_email": "", "name": "datalad_crawler", "package_url": "https://pypi.org/project/datalad_crawler/", "platform": "", "project_url": "https://pypi.org/project/datalad_crawler/", "project_urls": null, "release_url": "https://pypi.org/project/datalad_crawler/0.4.1/", "requires_dist": null, "requires_python": "", "summary": "DataLad extension package for crawling external web resources into an automated data distribution", "version": "0.4.1" }, "last_serial": 5426255, "releases": { "0.1": [ { "comment_text": "", "digests": { "md5": "fc825d00c7c8c410ba788ddccb0de2e9", "sha256": "4691c1633034f97d44ba0c77d1b84caac3e37d82c8868c1d8dd9d1d96f72ac98" }, "downloads": -1, "filename": "datalad_crawler-0.1-py2.py3-none-any.whl", "has_sig": false, "md5_digest": "fc825d00c7c8c410ba788ddccb0de2e9", "packagetype": "bdist_wheel", "python_version": "py2.py3", "requires_python": null, "size": 127417, "upload_time": "2018-05-11T06:47:09", "url": "https://files.pythonhosted.org/packages/18/52/d05c0d7a8ad8f8c862db47b5de53462556d1d68cd31aad6db4d79534f721/datalad_crawler-0.1-py2.py3-none-any.whl" }, { "comment_text": "", "digests": { "md5": "03a5dfd76f6710760f47c452c4b55b79", "sha256": "5add832a7cab5078eae5ddcd4c2c8f6e68755a6e9f587f8c44d76f69a8d9e0a9" }, "downloads": -1, "filename": "datalad_crawler-0.1.tar.gz", "has_sig": false, "md5_digest": "03a5dfd76f6710760f47c452c4b55b79", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 120201, "upload_time": "2018-05-11T06:47:11", "url": "https://files.pythonhosted.org/packages/bc/d4/079faea8fa08c3af33e89f33741ee8a2ed05682af90624ae5f6712603d39/datalad_crawler-0.1.tar.gz" } ], "0.1rc1": [ { "comment_text": "", "digests": { "md5": "17e2e81f113017fc09ad44095d5d3507", "sha256": "d6231d81b7c3b218ce0a674123c72652dbc7e6864d7a613750a9c0c7d8a175c9" }, "downloads": -1, "filename": "datalad_crawler-0.1rc1-py2.py3-none-any.whl", "has_sig": false, "md5_digest": "17e2e81f113017fc09ad44095d5d3507", "packagetype": "bdist_wheel", "python_version": "py2.py3", "requires_python": null, "size": 127475, "upload_time": "2018-05-09T15:33:55", "url": "https://files.pythonhosted.org/packages/aa/1b/54cc49da2dedd577ccb72538830feaf23e62509a6d46be546054372a15a5/datalad_crawler-0.1rc1-py2.py3-none-any.whl" }, { "comment_text": "", "digests": { "md5": "a44302943c9eb871c21d327995097246", "sha256": "bdb243e79d6b4b48f93d77eb26da8556520a5f28599a406e48cb4fbdedd8121d" }, "downloads": -1, "filename": "datalad_crawler-0.1rc1.tar.gz", "has_sig": false, "md5_digest": "a44302943c9eb871c21d327995097246", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 120211, "upload_time": "2018-05-09T15:33:57", "url": "https://files.pythonhosted.org/packages/37/5b/ba0e6ced797f1a4c5caa751751bce9ba8f18867acd719199d854736a52e6/datalad_crawler-0.1rc1.tar.gz" } ], "0.2": [ { "comment_text": "", "digests": { "md5": "cff51cdee6b6e7bd1c47a8312e2327ae", "sha256": "33837688f4674e839c5db53eb9d5ba508cd3935935e99e177f62f41de4cd8312" }, "downloads": -1, "filename": "datalad_crawler-0.2.tar.gz", "has_sig": false, "md5_digest": "cff51cdee6b6e7bd1c47a8312e2327ae", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 119485, "upload_time": "2018-05-17T14:36:09", "url": "https://files.pythonhosted.org/packages/b5/67/2035981f22be85d34444c3468cbe58b672533174cdc76b29d5497a716ed1/datalad_crawler-0.2.tar.gz" } ], "0.3": [ { "comment_text": "", "digests": { "md5": "5d65027f77e7cf3027e9981b921146cf", "sha256": "e1e82742eacf15131e18cfad0a5f672238a9c0086ab296d0225bc1da20478828" }, "downloads": -1, "filename": "datalad_crawler-0.3.tar.gz", "has_sig": false, "md5_digest": "5d65027f77e7cf3027e9981b921146cf", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 120500, "upload_time": "2019-02-06T20:33:11", "url": "https://files.pythonhosted.org/packages/03/0f/de78d373ee64fb69be6855b26efdc762dcb037b88421ebb57c1978bfdec6/datalad_crawler-0.3.tar.gz" } ], "0.4": [ { "comment_text": "", "digests": { "md5": "7305cc47bbb5c415b169031622f01a6e", "sha256": "adcc5549791ccbb95bfbb5d1eaddf553e1fb7d2bb2c6802ec3e1703002a297dc" }, "downloads": -1, "filename": "datalad_crawler-0.4.tar.gz", "has_sig": false, "md5_digest": "7305cc47bbb5c415b169031622f01a6e", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 121168, "upload_time": "2019-03-14T18:52:11", "url": "https://files.pythonhosted.org/packages/fe/8d/8008c3ea01889a3f0e955394b10aaf1c3d8dc56b67d98d762f7ad8450512/datalad_crawler-0.4.tar.gz" } ], "0.4.1": [ { "comment_text": "", "digests": { "md5": "e64cc71b9e0fc4b29101f172fec131a1", "sha256": "22c6d7889e02d3bb415d9b4574f688c20dca51bed547b44a8e0ecefee8ffd2b4" }, "downloads": -1, "filename": "datalad_crawler-0.4.1.tar.gz", "has_sig": false, "md5_digest": "e64cc71b9e0fc4b29101f172fec131a1", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 121890, "upload_time": "2019-06-20T14:52:18", "url": "https://files.pythonhosted.org/packages/1c/f4/e3e839639ff5da6f683de8642740819f07356927bc382609cbc9575ad61f/datalad_crawler-0.4.1.tar.gz" } ] }, "urls": [ { "comment_text": "", "digests": { "md5": "e64cc71b9e0fc4b29101f172fec131a1", "sha256": "22c6d7889e02d3bb415d9b4574f688c20dca51bed547b44a8e0ecefee8ffd2b4" }, "downloads": -1, "filename": "datalad_crawler-0.4.1.tar.gz", "has_sig": false, "md5_digest": "e64cc71b9e0fc4b29101f172fec131a1", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 121890, "upload_time": "2019-06-20T14:52:18", "url": "https://files.pythonhosted.org/packages/1c/f4/e3e839639ff5da6f683de8642740819f07356927bc382609cbc9575ad61f/datalad_crawler-0.4.1.tar.gz" } ] }