{ "info": { "author": "Katsuya Horiuchi", "author_email": "katsuya.horiuchi.biz@gmail.com", "bugtrack_url": null, "classifiers": [ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "Intended Audience :: Science/Research", "License :: OSI Approved :: Apache Software License", "Natural Language :: English", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: Implementation :: CPython" ], "description": "", "description_content_type": "", "docs_url": null, "download_url": "", "downloads": { "last_day": -1, "last_month": -1, "last_week": -1 }, "home_page": "https://github.com/katsuya-horiuchi/multiword_tokenization", "keywords": "NLP", "license": "Apache Software License", "maintainer": "", "maintainer_email": "", "name": "multiword-tokenization", "package_url": "https://pypi.org/project/multiword-tokenization/", "platform": "", "project_url": "https://pypi.org/project/multiword-tokenization/", "project_urls": { "Homepage": "https://github.com/katsuya-horiuchi/multiword_tokenization" }, "release_url": "https://pypi.org/project/multiword-tokenization/0.2.0a1/", "requires_dist": null, "requires_python": "", "summary": "Multi-word tokenization", "version": "0.2.0a1" }, "last_serial": 3710467, "releases": { "0.1.0a1": [ { "comment_text": "", "digests": { "md5": "16dd84350a9fceab357c28adbd4b10c5", "sha256": "57f984ae0e0ad74c132e95e8e5e022222231f20473debe03245c1eb273d7e337" }, "downloads": -1, "filename": "multiword_tokenization-0.1.0a1-py3-none-any.whl", "has_sig": false, "md5_digest": "16dd84350a9fceab357c28adbd4b10c5", "packagetype": "bdist_wheel", "python_version": "py3", "requires_python": null, "size": 4884, "upload_time": "2017-12-30T16:21:08", "url": "https://files.pythonhosted.org/packages/ea/a6/c24ebd0763468652296de431b7e12dcad2da249490fe1c1fe49b7213977d/multiword_tokenization-0.1.0a1-py3-none-any.whl" }, { "comment_text": "", "digests": { "md5": "d0dd95541a2502c18828f3d4c8cc683a", "sha256": "c35479e7bde2553af1852dd4debda61fe4267fc62add0154ce03761d9a55545d" }, "downloads": -1, "filename": "multiword_tokenization-0.1.0a1.tar.gz", "has_sig": false, "md5_digest": "d0dd95541a2502c18828f3d4c8cc683a", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 3824, "upload_time": "2017-12-30T16:21:10", "url": "https://files.pythonhosted.org/packages/80/81/f60d265ace34930737bd1f7938ddcd6a2b7c420ddc1d4b9d37b55367341a/multiword_tokenization-0.1.0a1.tar.gz" } ], "0.1.0a2": [ { "comment_text": "", "digests": { "md5": "04d23477740839af15b99aed83510152", "sha256": "4f51404967c1a06f565830423dec53a0a245cfebcb556490ca7c2f0da0889ffe" }, "downloads": -1, "filename": "multiword_tokenization-0.1.0a2-py3-none-any.whl", "has_sig": false, "md5_digest": "04d23477740839af15b99aed83510152", "packagetype": "bdist_wheel", "python_version": "py3", "requires_python": null, "size": 4877, "upload_time": "2017-12-30T16:26:08", "url": "https://files.pythonhosted.org/packages/1f/92/94d4d86d725e7914ce026f6fb08178e69e660ee2127f22a0ca94f75ffe71/multiword_tokenization-0.1.0a2-py3-none-any.whl" }, { "comment_text": "", "digests": { "md5": "5b3c94c43741941d563a19a95fada8a2", "sha256": "be288b891a7d0715cc51271c319564a93ebcb9e457ae76e15635ebcea77b52d9" }, "downloads": -1, "filename": "multiword_tokenization-0.1.0a2.tar.gz", "has_sig": false, "md5_digest": "5b3c94c43741941d563a19a95fada8a2", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 3823, "upload_time": "2017-12-30T16:26:10", "url": "https://files.pythonhosted.org/packages/6d/92/4c4bf51ee8a74dd2c6bd0e5c4cd1228a1b3f4a77aaf6bfed0147204e10d4/multiword_tokenization-0.1.0a2.tar.gz" } ], "0.2.0a1": [ { "comment_text": "", "digests": { "md5": "73dbb8771cb7958b4c6385157b9a6847", "sha256": "922958dff14557789ecd9c1f8c99ac3f96c3d1d24377b27add9e4474dfbc5818" }, "downloads": -1, "filename": "multiword_tokenization-0.2.0a1.tar.gz", "has_sig": false, "md5_digest": "73dbb8771cb7958b4c6385157b9a6847", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 4122, "upload_time": "2018-03-27T14:46:35", "url": "https://files.pythonhosted.org/packages/dd/7e/8375e47a4ee3f2659dbe69a382e4c9301ec95b0b8aa6ded8798dea658294/multiword_tokenization-0.2.0a1.tar.gz" } ] }, "urls": [ { "comment_text": "", "digests": { "md5": "73dbb8771cb7958b4c6385157b9a6847", "sha256": "922958dff14557789ecd9c1f8c99ac3f96c3d1d24377b27add9e4474dfbc5818" }, "downloads": -1, "filename": "multiword_tokenization-0.2.0a1.tar.gz", "has_sig": false, "md5_digest": "73dbb8771cb7958b4c6385157b9a6847", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 4122, "upload_time": "2018-03-27T14:46:35", "url": "https://files.pythonhosted.org/packages/dd/7e/8375e47a4ee3f2659dbe69a382e4c9301ec95b0b8aa6ded8798dea658294/multiword_tokenization-0.2.0a1.tar.gz" } ] }