6

I am struggling to add spark to my python path:

(myenv)me@me /home/me$ set SPARK_HOME="/home/me/spark-1.2.1-bin-hadoop2.4"
(myenv)me@me /home/me$ set PYTHONPATH=$PYTHONPATH:$SPARK_HOME:$SPARK_HOME/python:$SPARK_HOME/python/build:$SPARK_HOME/bin

(myenv)me@me /home/me$ python -c 'import sys; print(sys.path)'
['', '/home/me/.virtualenvs/default/lib/python2.7', '/home/me/.virtualenvs/default/lib/python2.7/plat-x86_64-linux-gnu', '/home/me/.virtualenvs/default/lib/python2.7/lib-tk', '/home/me/.virtualenvs/default/lib/python2.7/lib-old', '/home/me/.virtualenvs/default/lib/python2.7/lib-dynload', '/usr/lib/python2.7', '/usr/lib/python2.7/plat-x86_64-linux-gnu', '/usr/lib/python2.7/lib-tk', '/home/me/.virtualenvs/default/local/lib/python2.7/site-packages', '/home/me/.virtualenvs/default/lib/python2.7/site-packages']

(myenv)me@me /home/me$ python -c 'import pyspark'
Traceback (most recent call last):
  File "<string>", line 1, in <module>
ImportError: No module named pyspark
thecheech
  • 2,041
  • 3
  • 18
  • 25
  • Possible duplicate of [How do I install pyspark for use in standalone scripts?](http://stackoverflow.com/questions/25205264/how-do-i-install-pyspark-for-use-in-standalone-scripts) – Kamil Sindi Dec 31 '16 at 15:18

2 Answers2

7

I've got the same problem, but this helped.

Just add the following command in your .bashrc

export SPARK_HOME=/path/to/your/spark-1.4.1-bin-hadoop2.6
export PYTHONPATH=$SPARK_HOME/python:$SPARK_HOME/python/build:$PYTHONPATH
export PYTHONPATH=$SPARK_HOME/python/lib/py4j-0.8.2.1-src.zip:$PYTHONPATH
Community
  • 1
  • 1
ppk28
  • 367
  • 4
  • 6
0

I think you mixed up PYTHONPATH and sys.path. But are you sure you need to modify PYTHONPATH if you have pyspark installed properly?

EDIT:

I haven't used pyspark, but would this help? importing pyspark in python shell

Community
  • 1
  • 1
laike9m
  • 18,344
  • 20
  • 107
  • 140