{ "info": { "author": "Spark Installation authors", "author_email": "authors@example.com", "bugtrack_url": null, "classifiers": [ "Development Status :: 4 - Beta", "Intended Audience :: Developers", "License :: OSI Approved :: Apache Software License", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.3", "Programming Language :: Python :: 3.4", "Programming Language :: Python :: 3.5", "Topic :: Software Development" ], "description": "# Introduction \nThis Python script is intended to provide a smooth, cross-platform installation experience\nfor Spark, including WinUtils on Windows.\n\n# Getting Started\nPython 2.7 or 3.5 is required to execute this script. If installing on\nPython 3.6, ensure you choose Spark version 2.1.1 or higher (see [SPARK-19109](https://issues.apache.org/jira/browse/SPARK-19019)).\n\nRunning the script with no parameters will grab the latest Spark/Hadoop combination\nversion available.\n\nCommand line options -sv and -hv (or --sparkversion and --hadoopversion) allow the user\nto specify exactly which version pairing to use. Invalid pairings will present the list\nof valid options to the user.\n\n\n", "description_content_type": null, "docs_url": null, "download_url": "", "downloads": { "last_day": -1, "last_month": -1, "last_week": -1 }, "home_page": "https://github.com/rstudio/spark-install", "keywords": "Apache Spark PySpark Hadoop WinUtils", "license": "Apache", "maintainer": "", "maintainer_email": "", "name": "rstudio-spark-install", "package_url": "https://pypi.org/project/rstudio-spark-install/", "platform": "", "project_url": "https://pypi.org/project/rstudio-spark-install/", "project_urls": { "Homepage": "https://github.com/rstudio/spark-install" }, "release_url": "https://pypi.org/project/rstudio-spark-install/0.8.0/", "requires_dist": null, "requires_python": "", "summary": "Utility to setup various versions of Apache Spark on multiple platforms.", "version": "0.8.0" }, "last_serial": 3004727, "releases": { "0.8.0": [ { "comment_text": "", "digests": { "md5": "2379b9fdca00a7d293c6e6a37806cf2a", "sha256": "e203f4a0338b0893ee432a95f5ff102a0649598bd192b2f036935ef92f6338b6" }, "downloads": -1, "filename": "rstudio_spark_install-0.8.0-py2.py3-none-any.whl", "has_sig": false, "md5_digest": "2379b9fdca00a7d293c6e6a37806cf2a", "packagetype": "bdist_wheel", "python_version": "py2.py3", "requires_python": null, "size": 8124, "upload_time": "2017-07-06T20:56:05", "url": "https://files.pythonhosted.org/packages/ce/8a/631d73f07e155e0b033e11140cc660be5b8a3b4fce7991610f259f4c7b57/rstudio_spark_install-0.8.0-py2.py3-none-any.whl" } ] }, "urls": [ { "comment_text": "", "digests": { "md5": "2379b9fdca00a7d293c6e6a37806cf2a", "sha256": "e203f4a0338b0893ee432a95f5ff102a0649598bd192b2f036935ef92f6338b6" }, "downloads": -1, "filename": "rstudio_spark_install-0.8.0-py2.py3-none-any.whl", "has_sig": false, "md5_digest": "2379b9fdca00a7d293c6e6a37806cf2a", "packagetype": "bdist_wheel", "python_version": "py2.py3", "requires_python": null, "size": 8124, "upload_time": "2017-07-06T20:56:05", "url": "https://files.pythonhosted.org/packages/ce/8a/631d73f07e155e0b033e11140cc660be5b8a3b4fce7991610f259f4c7b57/rstudio_spark_install-0.8.0-py2.py3-none-any.whl" } ] }