{ "info": { "author": "Florian Friesdorf", "author_email": "flo@chaoflow.net", "bugtrack_url": null, "classifiers": [ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: GNU Library or Lesser General Public License (LGPL)", "Natural Language :: English", "Operating System :: OS Independent", "Programming Language :: Python", "Topic :: Software Development :: Testing" ], "description": "Introduction\n============\n\nThe purpose of chaoflow.testing.crawler is to find tests in your package and\nrun them along with tests in files you explicitly define.\n\nTests are found in the following files:\n\n- all .py files that are part of (subpackages of) your package, i.e. files that\n you can import with ``import``\n- all .txt files that have a corresponding .py file\n- all files you excplicitly specify\n\nAll tests found can be run individually, e.g. with zc.recipe.testrunner.\n\n\nUsing the test crawler\n======================\n\nIn order to use the test crawler for your package, you only need to copy one\nfile and declare the dependency on chaoflow.testing.crawler in your setup.py.\n\nDrop this file into your package root and adapt the files list to explicitly\nspecify test files that are not found otherwise, tests.py_::\n\n # chaoflow.testing.crawler.tests.py\n #\n # You can simply copy this file to your package and adjust it to your needs\n\n from chaoflow.testing.crawler import create_test_suite\n\n # File to test, relative to the package root\n # all .py files are found\n # all .txt files with corresponding .py file are found\n files = [\n 'README.txt'\n ]\n\n # We assume that this modules is in the root of your package\n pkgname = __name__[:-6]\n\n test_suite = create_test_suite(pkgname, files)\n\n.. _tests.py: http://github.com/chaoflow/chaoflow.testing.crawler/blob/master/src/chaoflow/testing/crawler/tests.py\n\n\n\nDeclare the dependency in setup.py::\n\n setup(...\n extras_require={\n 'test': [\n 'interlude',\n 'chaoflow.testing.ipython',\n 'chaoflow.testing.crawler',\n ],\n },\n )\n\nIf interlude is available, ``interlude.interact`` will be available as\n``interact`` in your test environment.\n\nIf chaoflow.testing.ipython is available, ``ipshell`` will be available as\n``ipshell`` in your test environment.\n\nexample buildout.cfg using chaoflow.testing.crawler and zc.recipe.testrunner::\n\n [buildout]\n develop = .\n parts = test py\n\n [test]\n recipe = zc.recipe.testrunner\n eggs = chaoflow.testing.crawler [test]\n\n [py]\n recipe = zc.recipe.egg\n interpreter = py\n eggs = ${test:eggs}\n\n\nAfter buildout, you can run your tests using ``./bin/test``.\nRun ``./bin/test --list-tests`` to get a list of all registered tests and see\n``./bin/test --help`` for further information.\n\n\nLicense\n=======\n\nchaoflow.testing.crawler is licensed under LGPLv3. Please let me know if this\npresents a problem for you.\n\nChangelog\n=========", "description_content_type": null, "docs_url": null, "download_url": "UNKNOWN", "downloads": { "last_day": -1, "last_month": -1, "last_week": -1 }, "home_page": "http://github.com/chaoflow/chaoflow.testing.crawler", "keywords": "", "license": "LGPL", "maintainer": null, "maintainer_email": null, "name": "chaoflow.testing.crawler", "package_url": "https://pypi.org/project/chaoflow.testing.crawler/", "platform": "UNKNOWN", "project_url": "https://pypi.org/project/chaoflow.testing.crawler/", "project_urls": { "Download": "UNKNOWN", "Homepage": "http://github.com/chaoflow/chaoflow.testing.crawler" }, "release_url": "https://pypi.org/project/chaoflow.testing.crawler/0.5/", "requires_dist": null, "requires_python": null, "summary": "Recursively finds tests in your package", "version": "0.5" }, "last_serial": 738924, "releases": { "0.3": [ { "comment_text": "", "digests": { "md5": "5fe99beae0aaf2541c38a01744e4614a", "sha256": "11cd520cb5e22906366c6a4b0cd7a4146583d08669707f9a3c6ce4a83b724732" }, "downloads": -1, "filename": "chaoflow.testing.crawler-0.3.zip", "has_sig": true, "md5_digest": "5fe99beae0aaf2541c38a01744e4614a", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 36766, "upload_time": "2009-11-01T18:11:45", "url": "https://files.pythonhosted.org/packages/57/19/8a19ce5a72e59c9c71a1b7334c8da5525921c6924f53ba78f591b3aad708/chaoflow.testing.crawler-0.3.zip" } ], "0.4": [ { "comment_text": "", "digests": { "md5": "aa2f86f07d9f1c9481869310a4466e8f", "sha256": "8e7aa313b7329b1ab9bb4bb51fabf321c907d28a26189d08191492b1df10f3a7" }, "downloads": -1, "filename": "chaoflow.testing.crawler-0.4.zip", "has_sig": true, "md5_digest": "aa2f86f07d9f1c9481869310a4466e8f", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 37239, "upload_time": "2009-11-26T15:23:30", "url": "https://files.pythonhosted.org/packages/17/dc/90f9d33d6c539dc1adafee2e20fcb4a1ac59e1f6eccc71dd9cc8c96921a7/chaoflow.testing.crawler-0.4.zip" } ], "0.5": [ { "comment_text": "", "digests": { "md5": "5e2327c71b8807b76561dc71511d3c30", "sha256": "9880738483b8f8da86a1015215bcdf806679c70f5eb30b2cdc1be9a03b9004e0" }, "downloads": -1, "filename": "chaoflow.testing.crawler-0.5.zip", "has_sig": true, "md5_digest": "5e2327c71b8807b76561dc71511d3c30", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 37249, "upload_time": "2009-11-26T17:26:14", "url": "https://files.pythonhosted.org/packages/46/68/81813c411eb4e40359b9dfd01b0e334f72a259911a370d229d2132f76c62/chaoflow.testing.crawler-0.5.zip" } ] }, "urls": [ { "comment_text": "", "digests": { "md5": "5e2327c71b8807b76561dc71511d3c30", "sha256": "9880738483b8f8da86a1015215bcdf806679c70f5eb30b2cdc1be9a03b9004e0" }, "downloads": -1, "filename": "chaoflow.testing.crawler-0.5.zip", "has_sig": true, "md5_digest": "5e2327c71b8807b76561dc71511d3c30", "packagetype": "sdist", "python_version": "source", "requires_python": null, "size": 37249, "upload_time": "2009-11-26T17:26:14", "url": "https://files.pythonhosted.org/packages/46/68/81813c411eb4e40359b9dfd01b0e334f72a259911a370d229d2132f76c62/chaoflow.testing.crawler-0.5.zip" } ] }