-2

I have Ubuntu 18.04.I am working with anaconda distribution for using jupyter notebooks and python.I want to install spark and pyspark to work with HDFS. I want to know the proper installation procedure for the same. Thank you

OneCricketeer
  • 179,855
  • 19
  • 132
  • 245
Alay Majmudar
  • 60
  • 1
  • 9
  • I tried using conda installation and manual installation but i am getting confused between setting up the environmnet variables in .bashrc file – Alay Majmudar Sep 09 '18 at 11:58

1 Answers1

2
conda install -c conda-forge pyspark 

This allows you to install PySpark into your anaconda environment using the conda-forge channel. In order for it to work with Spark, just run your code on the Spark cluster. For more information, look here which has some references with using anaconda specifically with PySpark and Spark.