1

I have a production cluster with no internet connection and I would like to run spark scripts. Also, I only have installed python 2.6 and I cannot install 2.7 Considering these limitations where can I find the python package for pyspark 1.6.0?

Best regards, João

Nipun Thennakoon
  • 3,586
  • 1
  • 18
  • 26
João
  • 177
  • 4
  • 15
  • You can download spark 1.6.0 from [here](https://spark.apache.org/downloads.html) and manually install it. (navigate to python/pyspark and install it) – mkaran Jan 16 '18 at 17:12
  • Hi thanks for your reply. How can I install pyspark python library with the pyspark binary? – João Jan 16 '18 at 17:46
  • oh i see. I found the answer here: https://stackoverflow.com/questions/25205264/how-do-i-install-pyspark-for-use-in-standalone-scripts thanks! – João Jan 16 '18 at 18:18
  • Glad to help, good luck! :) – mkaran Jan 16 '18 at 19:16
  • pyspark is part of the Spark distribution, you don't need to download it separately – desertnaut Jan 17 '18 at 10:08
  • 1
    @mkaran, you should answer the question so the OP may accept it and make it clear how the issue was solved. The way it is, it seems this question has no answer, which is untrue. It will also better document the question, keep SO cleaner, and grant you some SO points. Wouldn't that be nice for everyone? – Igor Donin Jul 09 '18 at 14:17

1 Answers1

0

Adding here from the comments as Igor Donin suggested:

You can download spark 1.6.0 or other older versions from here and manually install it: navigate to python/pyspark and install it.

As you've commented, the way to install it is described here

mkaran
  • 2,528
  • 20
  • 23