{ "info": { "author": "Li Fang", "author_email": "golifang1234@gmail.com", "bugtrack_url": null, "classifiers": [ "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 2", "Programming Language :: Python :: 3" ], "description": "", "description_content_type": "", "docs_url": null, "download_url": "", "downloads": { "last_day": -1, "last_month": -1, "last_week": -1 }, "home_page": "", "keywords": "", "license": "", "maintainer": "", "maintainer_email": "", "name": "bert-tokenizer", "package_url": "https://pypi.org/project/bert-tokenizer/", "platform": "", "project_url": "https://pypi.org/project/bert-tokenizer/", "project_urls": null, "release_url": "https://pypi.org/project/bert-tokenizer/0.1.5/", "requires_dist": null, "requires_python": "", "summary": "A Tokenizer for Bert model", "version": "0.1.5" }, "last_serial": 4498769, "releases": { "0.1.0": [ { "comment_text": "", "digests": { "md5": "d226bb8f7869f4040e77f5c1215cd74f", "sha256": "d5e6616cef6e4ec025a2d31dc0fafda3c3deaf9d6a890e956e09bb4ab2da5de7" }, "downloads": -1, "filename": "bert_tokenizer-0.1.0-py3-none-any.whl", "has_sig": false, "md5_digest": "d226bb8f7869f4040e77f5c1215cd74f", "packagetype": "bdist_wheel", "python_version": "py3", "requires_python": null, "size": 2301, "upload_time": "2018-11-17T03:18:56", "url": "https://files.pythonhosted.org/packages/ef/f1/3ac003fbad66efc5d4cb9fd652b3f5e3a352b231bfeacc6093cfb3c1df50/bert_tokenizer-0.1.0-py3-none-any.whl" }, { "comment_text": "", "digests": { "md5": "28978e235fe8226b043bab923c5df383", "sha256": "f35704703d26710ecb1b75a38f2149972c530945761c4c998d2837730c39c4a1" }, "downloads": -1, "filename": "bert_tokenizer-0.1.0.tar.gz", "has_sig": false, "md5_digest": "28978e235fe8226b043bab923c5df383", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 1604, "upload_time": "2018-11-17T03:18:59", "url": "https://files.pythonhosted.org/packages/f3/e1/bb4a3f9e5a4409b76b1cb9d70e5e80ec295576ebafa6939f0f7686e7f799/bert_tokenizer-0.1.0.tar.gz" } ], "0.1.1": [ { "comment_text": "", "digests": { "md5": "e0114c2fca1d010e511e20763b3c3d39", "sha256": "dfc8aad2d6adfeb2c2af987a196eb038ea65d7d08104969181db8035babcf6c7" }, "downloads": -1, "filename": "bert_tokenizer-0.1.1-py3-none-any.whl", "has_sig": false, "md5_digest": "e0114c2fca1d010e511e20763b3c3d39", "packagetype": "bdist_wheel", "python_version": "py3", "requires_python": null, "size": 2280, "upload_time": "2018-11-17T03:37:14", "url": "https://files.pythonhosted.org/packages/a6/78/8c6eb72ec2c9223e21d76786bb57274129ba25eb98d8262738c0393e49e1/bert_tokenizer-0.1.1-py3-none-any.whl" }, { "comment_text": "", "digests": { "md5": "e703c3957d55e30d66b4174957502d6e", "sha256": "e2eea769663a05353089087de58e6d36ae09b77e6942c771f8f45a865a32ab2f" }, "downloads": -1, "filename": "bert_tokenizer-0.1.1.tar.gz", "has_sig": false, "md5_digest": "e703c3957d55e30d66b4174957502d6e", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 1570, "upload_time": "2018-11-17T03:37:16", "url": "https://files.pythonhosted.org/packages/03/f0/d7bb1ce02dd893e2d04dfbf89cb900a3800a0f3d8597ec2b3ef71e162edc/bert_tokenizer-0.1.1.tar.gz" } ], "0.1.2": [ { "comment_text": "", "digests": { "md5": "26f2ffc88fdab9d2652bd73f393ff9e8", "sha256": "432e63c9e8c20138eff25eb51fc17d247e9d9cc0e14c6bb4c2015c677567fe91" }, "downloads": -1, "filename": "bert_tokenizer-0.1.2-py3-none-any.whl", "has_sig": false, "md5_digest": "26f2ffc88fdab9d2652bd73f393ff9e8", "packagetype": "bdist_wheel", "python_version": "py3", "requires_python": null, "size": 1041738, "upload_time": "2018-11-17T03:41:50", "url": "https://files.pythonhosted.org/packages/85/f2/814627b8dd89aa2c0e9bd548951c960149cd6d6f69d885d1744a75616fc3/bert_tokenizer-0.1.2-py3-none-any.whl" }, { "comment_text": "", "digests": { "md5": "70d4b1a00d6b2f8cd1f80d987cc87ce8", "sha256": "df240306d6a7a04d4df8e9474c070a51a0e0d76901572af65ebecae68606a0f6" }, "downloads": -1, "filename": "bert_tokenizer-0.1.2.tar.gz", "has_sig": false, "md5_digest": "70d4b1a00d6b2f8cd1f80d987cc87ce8", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 1031867, "upload_time": "2018-11-17T03:41:55", "url": "https://files.pythonhosted.org/packages/0c/b6/cd29383d465daa6b59bb57b49f24c26ed21dc159f4fb7961409f9d323406/bert_tokenizer-0.1.2.tar.gz" } ], "0.1.3": [ { "comment_text": "", "digests": { "md5": "5f0bbdb71c2843a78c82a83efdf630ae", "sha256": "870ab38377f8e5f5754e08b403793f16fbefda6992c628c14e866cca8c1d46ad" }, "downloads": -1, "filename": "bert_tokenizer-0.1.3-py3-none-any.whl", "has_sig": false, "md5_digest": "5f0bbdb71c2843a78c82a83efdf630ae", "packagetype": "bdist_wheel", "python_version": "py3", "requires_python": null, "size": 1041922, "upload_time": "2018-11-17T13:02:22", "url": "https://files.pythonhosted.org/packages/57/be/3766df968d719ed61b10da0a376af5fca82b51626037559cd6bba73ab34b/bert_tokenizer-0.1.3-py3-none-any.whl" }, { "comment_text": "", "digests": { "md5": "1aaa6d47bb2b4d170cc399ec79b72a2e", "sha256": "f495efd489e775b5c1707c8d7b83062f205a07a42692a5f7dfc6f1db8dad9e9b" }, "downloads": -1, "filename": "bert_tokenizer-0.1.3.tar.gz", "has_sig": false, "md5_digest": "1aaa6d47bb2b4d170cc399ec79b72a2e", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 1032105, "upload_time": "2018-11-17T13:02:27", "url": "https://files.pythonhosted.org/packages/48/0d/c4a13070c3d1d8a33fbba7b03afe2c24f5a3122faad7ddf8fceb2216ee92/bert_tokenizer-0.1.3.tar.gz" } ], "0.1.4": [ { "comment_text": "", "digests": { "md5": "1b927d2e1da2d89fd65ee0e18a1ee512", "sha256": "d5e452d90ed0548aa2aade1e152a23dce231b03dc953b9777413ec5480d69cf4" }, "downloads": -1, "filename": "bert_tokenizer-0.1.4-py3-none-any.whl", "has_sig": false, "md5_digest": "1b927d2e1da2d89fd65ee0e18a1ee512", "packagetype": "bdist_wheel", "python_version": "py3", "requires_python": null, "size": 1182377, "upload_time": "2018-11-17T15:06:13", "url": "https://files.pythonhosted.org/packages/31/48/12306ecc1248b72201c7b4d46d80bd1bab0a32f269973f05f75004a7442b/bert_tokenizer-0.1.4-py3-none-any.whl" }, { "comment_text": "", "digests": { "md5": "e9fe7ee5aefbd20719128593eaf2c550", "sha256": "47756ac91040dc944dac9de6e44eef5d533fd87dc03ec9ec1acd4fd104c5e275" }, "downloads": -1, "filename": "bert_tokenizer-0.1.4.tar.gz", "has_sig": false, "md5_digest": "e9fe7ee5aefbd20719128593eaf2c550", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 1172293, "upload_time": "2018-11-17T15:06:25", "url": "https://files.pythonhosted.org/packages/af/36/3fce2933273197afa3b301d145c98df833776337e8d8f23343161e2aa6b7/bert_tokenizer-0.1.4.tar.gz" } ], "0.1.5": [ { "comment_text": "", "digests": { "md5": "5898e2e1f15583656af0dceb47b6e2c5", "sha256": "0e7a36d937c2ae40b62dd3ff55a30f650c2f9d7a48629efed46df43181b505d9" }, "downloads": -1, "filename": "bert_tokenizer-0.1.5-py3-none-any.whl", "has_sig": false, "md5_digest": "5898e2e1f15583656af0dceb47b6e2c5", "packagetype": "bdist_wheel", "python_version": "py3", "requires_python": null, "size": 1182403, "upload_time": "2018-11-18T02:42:01", "url": "https://files.pythonhosted.org/packages/93/a3/eb0f8c1df4b8ee38c1487c32c0d21e4735f30487dea3fb2ea00b6fe61f09/bert_tokenizer-0.1.5-py3-none-any.whl" }, { "comment_text": "", "digests": { "md5": "4c5670db98e97b324fa84b902c7229db", "sha256": "c75f3449f0c7743593f2513b63e289b20c9f1763063a391f3efc1c29d8564462" }, "downloads": -1, "filename": "bert_tokenizer-0.1.5.tar.gz", "has_sig": false, "md5_digest": "4c5670db98e97b324fa84b902c7229db", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 1172326, "upload_time": "2018-11-18T02:42:06", "url": "https://files.pythonhosted.org/packages/de/a8/43a19cde294d9ecee2b0cafb7de5931cd062e155cf42be6fb10d4b417a93/bert_tokenizer-0.1.5.tar.gz" } ] }, "urls": [ { "comment_text": "", "digests": { "md5": "5898e2e1f15583656af0dceb47b6e2c5", "sha256": "0e7a36d937c2ae40b62dd3ff55a30f650c2f9d7a48629efed46df43181b505d9" }, "downloads": -1, "filename": "bert_tokenizer-0.1.5-py3-none-any.whl", "has_sig": false, "md5_digest": "5898e2e1f15583656af0dceb47b6e2c5", "packagetype": "bdist_wheel", "python_version": "py3", "requires_python": null, "size": 1182403, "upload_time": "2018-11-18T02:42:01", "url": "https://files.pythonhosted.org/packages/93/a3/eb0f8c1df4b8ee38c1487c32c0d21e4735f30487dea3fb2ea00b6fe61f09/bert_tokenizer-0.1.5-py3-none-any.whl" }, { "comment_text": "", "digests": { "md5": "4c5670db98e97b324fa84b902c7229db", "sha256": "c75f3449f0c7743593f2513b63e289b20c9f1763063a391f3efc1c29d8564462" }, "downloads": -1, "filename": "bert_tokenizer-0.1.5.tar.gz", "has_sig": false, "md5_digest": "4c5670db98e97b324fa84b902c7229db", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 1172326, "upload_time": "2018-11-18T02:42:06", "url": "https://files.pythonhosted.org/packages/de/a8/43a19cde294d9ecee2b0cafb7de5931cd062e155cf42be6fb10d4b417a93/bert_tokenizer-0.1.5.tar.gz" } ] }