{ "info": { "author": "Xiao Ma", "author_email": "Marshalma0923@gmail.com", "bugtrack_url": null, "classifiers": [ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Topic :: Software Development :: Build Tools" ], "description": "", "description_content_type": "", "docs_url": null, "download_url": "https://github.com/ALaughingHorse/tokenizer_xm/archive/v_02.tar.gz", "downloads": { "last_day": -1, "last_month": -1, "last_week": -1 }, "home_page": "https://github.com/ALaughingHorse/tokenizer_xm", "keywords": "text preprocessing,tokenize,NLP", "license": "MIT", "maintainer": "", "maintainer_email": "", "name": "tokenizer-xm", "package_url": "https://pypi.org/project/tokenizer-xm/", "platform": "", "project_url": "https://pypi.org/project/tokenizer-xm/", "project_urls": { "Download": "https://github.com/ALaughingHorse/tokenizer_xm/archive/v_02.tar.gz", "Homepage": "https://github.com/ALaughingHorse/tokenizer_xm" }, "release_url": "https://pypi.org/project/tokenizer-xm/0.2/", "requires_dist": null, "requires_python": "", "summary": "Tokenizing with options to exclude contractions, lemmatize and stem.", "version": "0.2" }, "last_serial": 5863971, "releases": { "0.1": [ { "comment_text": "", "digests": { "md5": "a0edfa44bf4e72e7c86162d61cf8727d", "sha256": "ede096ef449268c8bf955f48e495430ca1e4616f5f58a6d7761a75cc19a157aa" }, "downloads": -1, "filename": "tokenizer_xm-0.1.tar.gz", "has_sig": false, "md5_digest": "a0edfa44bf4e72e7c86162d61cf8727d", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 2994, "upload_time": "2019-09-20T19:44:48", "url": "https://files.pythonhosted.org/packages/55/2c/4d19ddee2423395aeb70dca5413716e4028665b354c5b489df460fb7da53/tokenizer_xm-0.1.tar.gz" } ], "0.2": [ { "comment_text": "", "digests": { "md5": "7d6e59a9959d14539d308d4c53c1099e", "sha256": "e6cb18dceaa547eef85fe7d472b82536e3a024d77737b629ed223953c91ec454" }, "downloads": -1, "filename": "tokenizer_xm-0.2.tar.gz", "has_sig": false, "md5_digest": "7d6e59a9959d14539d308d4c53c1099e", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 3013, "upload_time": "2019-09-20T20:12:08", "url": "https://files.pythonhosted.org/packages/b7/c8/8b2ce0816e126c9b3661ba921ff3616d10e080ad007b83b775a98d383601/tokenizer_xm-0.2.tar.gz" } ] }, "urls": [ { "comment_text": "", "digests": { "md5": "7d6e59a9959d14539d308d4c53c1099e", "sha256": "e6cb18dceaa547eef85fe7d472b82536e3a024d77737b629ed223953c91ec454" }, "downloads": -1, "filename": "tokenizer_xm-0.2.tar.gz", "has_sig": false, "md5_digest": "7d6e59a9959d14539d308d4c53c1099e", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 3013, "upload_time": "2019-09-20T20:12:08", "url": "https://files.pythonhosted.org/packages/b7/c8/8b2ce0816e126c9b3661ba921ff3616d10e080ad007b83b775a98d383601/tokenizer_xm-0.2.tar.gz" } ] }