{ "info": { "author": "Iv\u00e1n Arias Rodr\u00edguez", "author_email": "ivan.arias.rodriguez@gmail.com", "bugtrack_url": null, "classifiers": [ "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 3" ], "description": "# IAR Tokenizer\n\nThe IAR (Iv\u00c3\u00a1n Arias Rodr\u00c3\u00adguez) Tokenizer is a tokenizer developed mainly for Spanish. It is able to divide a text in paragraphs, those in \nsentences, and each sentence in a list of tokens.\n\nMore information to be added in the future...\n\n", "description_content_type": "text/markdown", "docs_url": null, "download_url": "", "downloads": { "last_day": -1, "last_month": -1, "last_week": -1 }, "home_page": "https://github.com/ivansiiito/iar_tokenizer", "keywords": "", "license": "", "maintainer": "", "maintainer_email": "", "name": "iar-tokenizer", "package_url": "https://pypi.org/project/iar-tokenizer/", "platform": "", "project_url": "https://pypi.org/project/iar-tokenizer/", "project_urls": { "Homepage": "https://github.com/ivansiiito/iar_tokenizer" }, "release_url": "https://pypi.org/project/iar-tokenizer/1.0.10/", "requires_dist": [ "nltk" ], "requires_python": "", "summary": "A tokenizer focused on Spanish language.", "version": "1.0.10" }, "last_serial": 5372628, "releases": { "1.0.0": [ { "comment_text": "", "digests": { "md5": "e068e32b3baea20b7f89824a2d36479c", "sha256": "4c394ee333d4dd0c26bf2a6573a1975fe7bf8416a0c01715a19b1f8dbc378353" }, "downloads": -1, "filename": "iar_tokenizer-1.0.0-py3-none-any.whl", "has_sig": false, "md5_digest": "e068e32b3baea20b7f89824a2d36479c", "packagetype": "bdist_wheel", "python_version": "py3", "requires_python": null, "size": 11740, "upload_time": "2019-05-17T16:14:59", "url": "https://files.pythonhosted.org/packages/37/c8/4b8e37d27cc2041a3920e1ca3526eec1d21ecbd9668d9017d7a5d889b5c1/iar_tokenizer-1.0.0-py3-none-any.whl" } ], "1.0.1": [ { "comment_text": "", "digests": { "md5": "572b68445c8e6785b512aa2242de05f7", "sha256": "4d9809b5b13c18e854d7bdbee6efed389f8894f65a4e36bd6e6e20e84175091b" }, "downloads": -1, "filename": "iar_tokenizer-1.0.1-py3-none-any.whl", "has_sig": false, "md5_digest": "572b68445c8e6785b512aa2242de05f7", "packagetype": "bdist_wheel", "python_version": "py3", "requires_python": null, "size": 11801, "upload_time": "2019-05-18T09:25:54", "url": "https://files.pythonhosted.org/packages/84/94/1033436402bf7ca5526b0b7dcaa66efe3933893b13326d2137935be86ede/iar_tokenizer-1.0.1-py3-none-any.whl" }, { "comment_text": "", "digests": { "md5": "c4578d58f9a56297305e73c5681d084c", "sha256": "a5bd6ad6facab7ad064c192cd0da3c8a7d1492f7f0eaf5b493482a4419257df5" }, "downloads": -1, "filename": "iar_tokenizer-1.0.1.tar.gz", "has_sig": false, "md5_digest": "c4578d58f9a56297305e73c5681d084c", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 10887, "upload_time": "2019-05-18T09:25:55", "url": "https://files.pythonhosted.org/packages/0e/18/cd66e70ef1327a75ed86c4dc78f2c7806be1e031c21edf13b714f651b598/iar_tokenizer-1.0.1.tar.gz" } ], "1.0.10": [ { "comment_text": "", "digests": { "md5": "20fa152d03875c1488c7a8404707ee9e", "sha256": "bc28adc449f9afcfabf6e439e87064e737129c19bec3dc6dbc4989d894afae6d" }, "downloads": -1, "filename": "iar_tokenizer-1.0.10-py3-none-any.whl", "has_sig": false, "md5_digest": "20fa152d03875c1488c7a8404707ee9e", "packagetype": "bdist_wheel", "python_version": "py3", "requires_python": null, "size": 21591, "upload_time": "2019-06-07T18:16:10", "url": "https://files.pythonhosted.org/packages/4c/f5/680c56f689d67d6471469cca6d4be7a10673966204dd47ccf3901b136e48/iar_tokenizer-1.0.10-py3-none-any.whl" }, { "comment_text": "", "digests": { "md5": "86950cdf470a1f98f5ca645632cbe9c0", "sha256": "5c54125b73d759edb75ccbc2319d36b3a2fe40ed4186e2014fc2cb76a405afec" }, "downloads": -1, "filename": "iar_tokenizer-1.0.10.tar.gz", "has_sig": false, "md5_digest": "86950cdf470a1f98f5ca645632cbe9c0", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 11165, "upload_time": "2019-06-07T18:16:11", "url": "https://files.pythonhosted.org/packages/aa/f8/c452ffe617110ec6618d54155b3a0ca1898f06d40f04e6845f000d5f22f7/iar_tokenizer-1.0.10.tar.gz" } ], "1.0.2": [ { "comment_text": "", "digests": { "md5": "8ff5ff320a642070d7a9f49ac56b51c2", "sha256": "8ad6b9cbcf9df237576d9ac843c37ca4b4ac2fb858ce9c3a300d5637494bd872" }, "downloads": -1, "filename": "iar_tokenizer-1.0.2-py3-none-any.whl", "has_sig": false, "md5_digest": "8ff5ff320a642070d7a9f49ac56b51c2", "packagetype": "bdist_wheel", "python_version": "py3", "requires_python": null, "size": 11801, "upload_time": "2019-05-18T09:40:05", "url": "https://files.pythonhosted.org/packages/9d/c1/46acfe469720da92500fded65ee5ea6f1aef847e7db49b603181638931cb/iar_tokenizer-1.0.2-py3-none-any.whl" }, { "comment_text": "", "digests": { "md5": "7652668ba9fdcd1ff95606bedb57c33a", "sha256": "ad4c8a9cec2fd327de3e55a764c7eaae0934f1b9670ad651bd1f23e1921f88e9" }, "downloads": -1, "filename": "iar_tokenizer-1.0.2.tar.gz", "has_sig": false, "md5_digest": "7652668ba9fdcd1ff95606bedb57c33a", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 10874, "upload_time": "2019-05-18T09:40:07", "url": "https://files.pythonhosted.org/packages/c2/87/b4b6e1abc46e72f61aedfda70d5dd204b17ac27e92fb52d815f6fb50c6b9/iar_tokenizer-1.0.2.tar.gz" } ], "1.0.4": [ { "comment_text": "", "digests": { "md5": "0df4ba5987da8ae3a03d3108a3354ae1", "sha256": "91b768612c2789a3a1eae6591e5dfe607a1862bd67202c8a89062d0b6229e310" }, "downloads": -1, "filename": "iar_tokenizer-1.0.4-py3-none-any.whl", "has_sig": false, "md5_digest": "0df4ba5987da8ae3a03d3108a3354ae1", "packagetype": "bdist_wheel", "python_version": "py3", "requires_python": null, "size": 11820, "upload_time": "2019-05-22T00:46:44", "url": "https://files.pythonhosted.org/packages/dc/3f/2afce4d6202dd25a4c7b53c508afe8ae9cb15c22d5586753ed822b3d2ce1/iar_tokenizer-1.0.4-py3-none-any.whl" }, { "comment_text": "", "digests": { "md5": "c414ce614b82d764abdc6ae8d5ce444b", "sha256": "cf1bb5170286470faddb9485670a83743a8ac9e4b7578f2c7b38cf4edb633ddc" }, "downloads": -1, "filename": "iar_tokenizer-1.0.4.tar.gz", "has_sig": false, "md5_digest": "c414ce614b82d764abdc6ae8d5ce444b", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 10936, "upload_time": "2019-05-22T00:46:46", "url": "https://files.pythonhosted.org/packages/91/ce/2447428960f18d12c14eb4db0660eada4dcca2c7fee6631cd749d55aaae1/iar_tokenizer-1.0.4.tar.gz" } ], "1.0.5": [ { "comment_text": "", "digests": { "md5": "c882cb85c554d04642f75e8b81bde1a3", "sha256": "9448cf68ceaaa477f4d9d2d4548b14c334aafb3b1fd472e055f6599a8c02a5ce" }, "downloads": -1, "filename": "iar_tokenizer-1.0.5-py3-none-any.whl", "has_sig": false, "md5_digest": "c882cb85c554d04642f75e8b81bde1a3", "packagetype": "bdist_wheel", "python_version": "py3", "requires_python": null, "size": 12102, "upload_time": "2019-06-07T16:17:43", "url": "https://files.pythonhosted.org/packages/de/19/414047d25dff6e155ca2fef9fb6bffce9c2de6899cc1136e03c6084e277b/iar_tokenizer-1.0.5-py3-none-any.whl" }, { "comment_text": "", "digests": { "md5": "c13c093bd3cf04ca0704aeedd37ed023", "sha256": "bc1d119450e625857fcbaf4e34c0d51ee211c21e1046fa2b09b8ad487ed9191d" }, "downloads": -1, "filename": "iar_tokenizer-1.0.5.tar.gz", "has_sig": false, "md5_digest": "c13c093bd3cf04ca0704aeedd37ed023", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 11169, "upload_time": "2019-06-07T16:17:53", "url": "https://files.pythonhosted.org/packages/20/0a/f1aa6474b2ab26e30828c72e3486a2ad135fad62823d21bba32ff6eb7494/iar_tokenizer-1.0.5.tar.gz" } ], "1.0.6": [ { "comment_text": "", "digests": { "md5": "eacbb8fa2758f4cd61e6f507afe82e3f", "sha256": "66985deb46570d23e53e8a3936113c7158ddc948e99745ea8791711c1fc4e725" }, "downloads": -1, "filename": "iar_tokenizer-1.0.6-py3-none-any.whl", "has_sig": false, "md5_digest": "eacbb8fa2758f4cd61e6f507afe82e3f", "packagetype": "bdist_wheel", "python_version": "py3", "requires_python": null, "size": 12097, "upload_time": "2019-06-07T16:22:42", "url": "https://files.pythonhosted.org/packages/5a/56/4802326209df979a705046eeefa3025f6664cfa371a1fcecba0518d7a894/iar_tokenizer-1.0.6-py3-none-any.whl" }, { "comment_text": "", "digests": { "md5": "30d5da165512485f537d38d9299d0f77", "sha256": "681bcad324615c3ad46b25da46adbb79f95e2a09597d3dff8ba637dd3c7ac996" }, "downloads": -1, "filename": "iar_tokenizer-1.0.6.tar.gz", "has_sig": false, "md5_digest": "30d5da165512485f537d38d9299d0f77", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 11165, "upload_time": "2019-06-07T16:22:43", "url": "https://files.pythonhosted.org/packages/12/92/23f9c0683e2b69cf290ea86ba720b9084e69e2052f69c5b2467675fe7366/iar_tokenizer-1.0.6.tar.gz" } ], "1.0.7": [ { "comment_text": "", "digests": { "md5": "64be1664e42affcea2b09aed47f82453", "sha256": "8d9831e6e248397f89c51c0c1d9f00327565ca9acbc8776f452ec00782c37cd9" }, "downloads": -1, "filename": "iar_tokenizer-1.0.7-py3-none-any.whl", "has_sig": false, "md5_digest": "64be1664e42affcea2b09aed47f82453", "packagetype": "bdist_wheel", "python_version": "py3", "requires_python": null, "size": 12019, "upload_time": "2019-06-07T17:18:04", "url": "https://files.pythonhosted.org/packages/22/d2/d61942beabd7b8138b83fb2eccbcf9596e76e983e7557df5de3740538f24/iar_tokenizer-1.0.7-py3-none-any.whl" }, { "comment_text": "", "digests": { "md5": "6b850bb709a0ab028afada28d1b57545", "sha256": "7dd4bef0765fdf3c99c2753d52ae22064b7eae71b59e8ffc4660e1aac4c6009b" }, "downloads": -1, "filename": "iar_tokenizer-1.0.7.tar.gz", "has_sig": false, "md5_digest": "6b850bb709a0ab028afada28d1b57545", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 11117, "upload_time": "2019-06-07T17:18:05", "url": "https://files.pythonhosted.org/packages/6e/a0/bb6a26e54a98fc1ced4d223ab945ea2926bcb031a6f6a4e62d8596924354/iar_tokenizer-1.0.7.tar.gz" } ], "1.0.9": [ { "comment_text": "", "digests": { "md5": "27de4ddf1ace96bd5743de0784cef9b2", "sha256": "011f5c6cc01a229faeb67f3fad9e86849bf63a404e3c5ef8cc72872f341ebd98" }, "downloads": -1, "filename": "iar_tokenizer-1.0.9-py3-none-any.whl", "has_sig": false, "md5_digest": "27de4ddf1ace96bd5743de0784cef9b2", "packagetype": "bdist_wheel", "python_version": "py3", "requires_python": null, "size": 21577, "upload_time": "2019-06-07T17:45:29", "url": "https://files.pythonhosted.org/packages/93/01/84bb3ae34bd8593b94d860989c46b38ac56db5c30f205c6b3bac86d9cadc/iar_tokenizer-1.0.9-py3-none-any.whl" }, { "comment_text": "", "digests": { "md5": "a79b8447af61b5ccf1e98e340d0d792e", "sha256": "218e9e5f92f28b6351f8f55e4aa886965a3f1031941524834ec3e2587fbfe919" }, "downloads": -1, "filename": "iar_tokenizer-1.0.9.tar.gz", "has_sig": false, "md5_digest": "a79b8447af61b5ccf1e98e340d0d792e", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 11123, "upload_time": "2019-06-07T17:45:30", "url": "https://files.pythonhosted.org/packages/30/84/8abe8318231b7d5f847240006693d1116e9cde71d567118f17b2cdb83ad9/iar_tokenizer-1.0.9.tar.gz" } ] }, "urls": [ { "comment_text": "", "digests": { "md5": "20fa152d03875c1488c7a8404707ee9e", "sha256": "bc28adc449f9afcfabf6e439e87064e737129c19bec3dc6dbc4989d894afae6d" }, "downloads": -1, "filename": "iar_tokenizer-1.0.10-py3-none-any.whl", "has_sig": false, "md5_digest": "20fa152d03875c1488c7a8404707ee9e", "packagetype": "bdist_wheel", "python_version": "py3", "requires_python": null, "size": 21591, "upload_time": "2019-06-07T18:16:10", "url": "https://files.pythonhosted.org/packages/4c/f5/680c56f689d67d6471469cca6d4be7a10673966204dd47ccf3901b136e48/iar_tokenizer-1.0.10-py3-none-any.whl" }, { "comment_text": "", "digests": { "md5": "86950cdf470a1f98f5ca645632cbe9c0", "sha256": "5c54125b73d759edb75ccbc2319d36b3a2fe40ed4186e2014fc2cb76a405afec" }, "downloads": -1, "filename": "iar_tokenizer-1.0.10.tar.gz", "has_sig": false, "md5_digest": "86950cdf470a1f98f5ca645632cbe9c0", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 11165, "upload_time": "2019-06-07T18:16:11", "url": "https://files.pythonhosted.org/packages/aa/f8/c452ffe617110ec6618d54155b3a0ca1898f06d40f04e6845f000d5f22f7/iar_tokenizer-1.0.10.tar.gz" } ] }