{ "info": { "author": "Daniel Perez", "author_email": "tuvistavie@gmail.com", "bugtrack_url": null, "classifiers": [ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3" ], "description": "# bigcode-tokenizer\n\nTool to tokenize source code.\n\n\n", "description_content_type": null, "docs_url": null, "download_url": "https://github.com/tuvistavie/bigcode-tools/archive/master.zip", "downloads": { "last_day": -1, "last_month": -1, "last_week": -1 }, "home_page": "https://github.com/tuvistavie/bigcode-tools/tree/master/bigcode-tokenizer", "keywords": "", "license": "", "maintainer": "", "maintainer_email": "", "name": "bigcode-tokenizer", "package_url": "https://pypi.org/project/bigcode-tokenizer/", "platform": "", "project_url": "https://pypi.org/project/bigcode-tokenizer/", "project_urls": { "Download": "https://github.com/tuvistavie/bigcode-tools/archive/master.zip", "Homepage": "https://github.com/tuvistavie/bigcode-tools/tree/master/bigcode-tokenizer" }, "release_url": "https://pypi.org/project/bigcode-tokenizer/0.1.0/", "requires_dist": [ "pygments", "nose; extra == 'test'" ], "requires_python": "", "summary": "Tool to tokenize source code", "version": "0.1.0" }, "last_serial": 3584621, "releases": { "0.1.0": [ { "comment_text": "", "digests": { "md5": "c04bbddf718608456e67102946149f7c", "sha256": "591654a617fb3f88fdedd8c1d953d48c68529f024cbb66945898ef5e9a3b86bf" }, "downloads": -1, "filename": "bigcode_tokenizer-0.1.0-py3-none-any.whl", "has_sig": true, "md5_digest": "c04bbddf718608456e67102946149f7c", "packagetype": "bdist_wheel", "python_version": "py3", "requires_python": null, "size": 7459, "upload_time": "2018-02-15T14:55:51", "url": "https://files.pythonhosted.org/packages/16/c3/1fbd0dc211803208ec7a5f61a9dcf2bae476f3dd1a6969209095ac2b3f08/bigcode_tokenizer-0.1.0-py3-none-any.whl" }, { "comment_text": "", "digests": { "md5": "e1fb91c6b03c1ae26951d61f3bebe428", "sha256": "0ace2a55c667c72e0245b159309c38b0b2acd7fbe53ccb98e9fe165e072ba3cf" }, "downloads": -1, "filename": "bigcode-tokenizer-0.1.0.tar.gz", "has_sig": true, "md5_digest": "e1fb91c6b03c1ae26951d61f3bebe428", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 4368, "upload_time": "2018-02-15T14:55:53", "url": "https://files.pythonhosted.org/packages/13/43/c9cdc41794471f3d2ee9ef4f266ffea2f8e0cd0271f0100c5a1742c758a9/bigcode-tokenizer-0.1.0.tar.gz" } ] }, "urls": [ { "comment_text": "", "digests": { "md5": "c04bbddf718608456e67102946149f7c", "sha256": "591654a617fb3f88fdedd8c1d953d48c68529f024cbb66945898ef5e9a3b86bf" }, "downloads": -1, "filename": "bigcode_tokenizer-0.1.0-py3-none-any.whl", "has_sig": true, "md5_digest": "c04bbddf718608456e67102946149f7c", "packagetype": "bdist_wheel", "python_version": "py3", "requires_python": null, "size": 7459, "upload_time": "2018-02-15T14:55:51", "url": "https://files.pythonhosted.org/packages/16/c3/1fbd0dc211803208ec7a5f61a9dcf2bae476f3dd1a6969209095ac2b3f08/bigcode_tokenizer-0.1.0-py3-none-any.whl" }, { "comment_text": "", "digests": { "md5": "e1fb91c6b03c1ae26951d61f3bebe428", "sha256": "0ace2a55c667c72e0245b159309c38b0b2acd7fbe53ccb98e9fe165e072ba3cf" }, "downloads": -1, "filename": "bigcode-tokenizer-0.1.0.tar.gz", "has_sig": true, "md5_digest": "e1fb91c6b03c1ae26951d61f3bebe428", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 4368, "upload_time": "2018-02-15T14:55:53", "url": "https://files.pythonhosted.org/packages/13/43/c9cdc41794471f3d2ee9ef4f266ffea2f8e0cd0271f0100c5a1742c758a9/bigcode-tokenizer-0.1.0.tar.gz" } ] }