{ "info": { "author": "Patrick Lehmann", "author_email": "Paebbels@gmail.com", "bugtrack_url": null, "classifiers": [ "Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Operating System :: OS Independent", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3 :: Only", "Topic :: Utilities" ], "description": "![PyPI - License](https://img.shields.io/pypi/l/pyTokenizer)\n![GitHub tag (latest by date)](https://img.shields.io/github/v/tag/Paebbels/pyTokenizer) \n![GitHub release (latest by date)](https://img.shields.io/github/v/release/Paebbels/pyTokenizer)\n[![Documentation Status](https://readthedocs.org/projects/pytokenizer/badge/?version=latest)](https://pyTokenizer.readthedocs.io/en/latest/?badge=latest) \n[![PyPI](https://img.shields.io/pypi/v/pyTokenizer)](https://pypi.org/project/pyTokenizer/)\n![PyPI - Python Version](https://img.shields.io/pypi/pyversions/pyTokenizer)\n![PyPI - Wheel](https://img.shields.io/pypi/wheel/pyTokenizer)\n![PyPI - Status](https://img.shields.io/pypi/status/pyTokenizer)\n\n# pyTokenizer\n\nA streaming tokenizer.\n\n\n## Contributors:\n\n* [Patrick Lehmann](https://github.com/Paebbels) (Maintainer)\n\n\n## License\n\nThis library is licensed under [Apache License 2.0](LICENSE.md)\n\n-------------------------\n\nSPDX-License-Identifier: Apache-2.0\n\n\n", "description_content_type": "text/markdown", "docs_url": null, "download_url": "", "downloads": { "last_day": -1, "last_month": -1, "last_week": -1 }, "home_page": "https://github.com/Paebbels/pyTokenizer", "keywords": "Python3 Token Tokenizer Streaming", "license": "", "maintainer": "", "maintainer_email": "", "name": "pyTokenizer", "package_url": "https://pypi.org/project/pyTokenizer/", "platform": "", "project_url": "https://pypi.org/project/pyTokenizer/", "project_urls": { "Documentation": "https://pyTokenizer.readthedocs.io/en/latest/", "Homepage": "https://github.com/Paebbels/pyTokenizer", "Issue Tracker": "https://github.com/Paebbels/pyTokenizer/issues", "Source Code": "https://github.com/Paebbels/pyTokenizer" }, "release_url": "https://pypi.org/project/pyTokenizer/1.1.1/", "requires_dist": null, "requires_python": ">=3.5", "summary": "A streaming tokenizer.", "version": "1.1.1" }, "last_serial": 5966290, "releases": { "0.1.3": [ { "comment_text": "", "digests": { "md5": "fe86374d109f9104da264c50e6c3806b", "sha256": "bb4509feabca85867d6f7eafa74216b6a9b6c693dc8c69ede34d0a9df687d1e5" }, "downloads": -1, "filename": "pyTokenizer-0.1.3-py3-none-any.whl", "has_sig": false, "md5_digest": "fe86374d109f9104da264c50e6c3806b", "packagetype": "bdist_wheel", "python_version": "py3", "requires_python": ">=3.5", "size": 8175, "upload_time": "2019-10-13T04:18:32", "url": "https://files.pythonhosted.org/packages/d9/e1/9af20afe462a6fb06d3b1c843b9e683d08e544c42613cf3d6f4d9345f120/pyTokenizer-0.1.3-py3-none-any.whl" }, { "comment_text": "", "digests": { "md5": "a93612945512f9dbdab28b7a86e1e795", "sha256": "d5cd368a4cc9513a870558a0937a66250adbb86406b68933979829160b5da4ee" }, "downloads": -1, "filename": "pyTokenizer-0.1.3.tar.gz", "has_sig": false, "md5_digest": "a93612945512f9dbdab28b7a86e1e795", "packagetype": "sdist", "python_version": "source", "requires_python": ">=3.5", "size": 4086, "upload_time": "2019-10-13T04:18:35", "url": "https://files.pythonhosted.org/packages/d9/4a/16b81038e1fd332a330ffb73a78af369000651742c2e8b4772293af7ed83/pyTokenizer-0.1.3.tar.gz" } ], "1.1.0": [ { "comment_text": "", "digests": { "md5": "625b9e0c952cc7c66bbcc6c5157d173b", "sha256": "a8ba0bb856a525dbbbfccfb363422db178bf23d14a171d3d6bc86c9f95946154" }, "downloads": -1, "filename": "pyTokenizer-1.1.0-py3-none-any.whl", "has_sig": false, "md5_digest": "625b9e0c952cc7c66bbcc6c5157d173b", "packagetype": "bdist_wheel", "python_version": "py3", "requires_python": ">=3.5", "size": 8175, "upload_time": "2019-10-13T04:19:55", "url": "https://files.pythonhosted.org/packages/1a/7f/06ff34a034938510f6c8a295bdf65947ada51ef2298652d3e36941967483/pyTokenizer-1.1.0-py3-none-any.whl" }, { "comment_text": "", "digests": { "md5": "586c63e8b52deff8e109f2369974cf13", "sha256": "27bfed3aa9e4bc501225d64103bf0582c3b5e3c84ea02b0306cd18b7ae893e5e" }, "downloads": -1, "filename": "pyTokenizer-1.1.0.tar.gz", "has_sig": false, "md5_digest": "586c63e8b52deff8e109f2369974cf13", "packagetype": "sdist", "python_version": "source", "requires_python": ">=3.5", "size": 4096, "upload_time": "2019-10-13T04:19:58", "url": "https://files.pythonhosted.org/packages/af/ea/0855b8df88f6c7d0e7a81133a6ac969f893c21b5f175e150ceb8fd8987c6/pyTokenizer-1.1.0.tar.gz" } ], "1.1.1": [ { "comment_text": "", "digests": { "md5": "c6caa5f2f13fa570e85f36312b9b89b9", "sha256": "16b1831af36e7c2bf1d1f29bf9455afe0f4dda01dcf489d7347fadf2c233e31c" }, "downloads": -1, "filename": "pyTokenizer-1.1.1-py3-none-any.whl", "has_sig": false, "md5_digest": "c6caa5f2f13fa570e85f36312b9b89b9", "packagetype": "bdist_wheel", "python_version": "py3", "requires_python": ">=3.5", "size": 8138, "upload_time": "2019-10-13T04:37:38", "url": "https://files.pythonhosted.org/packages/a8/2e/1bbef671260f327e464e22ee3ddfdc9de1465947bb7e771c71c9e8695dd1/pyTokenizer-1.1.1-py3-none-any.whl" }, { "comment_text": "", "digests": { "md5": "f01f014f0ee64efec7d98fc86fcc79da", "sha256": "8b82f83ed01f2ab81eebcee1ee8d40c0c7c67e8270d026de05b19596da1b26d6" }, "downloads": -1, "filename": "pyTokenizer-1.1.1.tar.gz", "has_sig": false, "md5_digest": "f01f014f0ee64efec7d98fc86fcc79da", "packagetype": "sdist", "python_version": "source", "requires_python": ">=3.5", "size": 4064, "upload_time": "2019-10-13T04:37:39", "url": "https://files.pythonhosted.org/packages/82/ed/fc340f7c35703a7733d43a8e6e782526d353a22176763b5b63bd8be40c6f/pyTokenizer-1.1.1.tar.gz" } ] }, "urls": [ { "comment_text": "", "digests": { "md5": "c6caa5f2f13fa570e85f36312b9b89b9", "sha256": "16b1831af36e7c2bf1d1f29bf9455afe0f4dda01dcf489d7347fadf2c233e31c" }, "downloads": -1, "filename": "pyTokenizer-1.1.1-py3-none-any.whl", "has_sig": false, "md5_digest": "c6caa5f2f13fa570e85f36312b9b89b9", "packagetype": "bdist_wheel", "python_version": "py3", "requires_python": ">=3.5", "size": 8138, "upload_time": "2019-10-13T04:37:38", "url": "https://files.pythonhosted.org/packages/a8/2e/1bbef671260f327e464e22ee3ddfdc9de1465947bb7e771c71c9e8695dd1/pyTokenizer-1.1.1-py3-none-any.whl" }, { "comment_text": "", "digests": { "md5": "f01f014f0ee64efec7d98fc86fcc79da", "sha256": "8b82f83ed01f2ab81eebcee1ee8d40c0c7c67e8270d026de05b19596da1b26d6" }, "downloads": -1, "filename": "pyTokenizer-1.1.1.tar.gz", "has_sig": false, "md5_digest": "f01f014f0ee64efec7d98fc86fcc79da", "packagetype": "sdist", "python_version": "source", "requires_python": ">=3.5", "size": 4064, "upload_time": "2019-10-13T04:37:39", "url": "https://files.pythonhosted.org/packages/82/ed/fc340f7c35703a7733d43a8e6e782526d353a22176763b5b63bd8be40c6f/pyTokenizer-1.1.1.tar.gz" } ] }