I have installed pyspark in a miniconda environment on Ubuntu through conda install pyspark
. So far everything works fine: I can run jobs through spark-submit
and I can inspect running jobs at localhost:4040
. But I can't locate start-history-server.sh
, which I need to look at jobs that have completed.
It is supposed to be in {spark}/sbin
, where {spark}
is the installation directory of spark. I'm not sure where that is supposed to be when spark is installed through conda, but I have searched through the entire miniconda directory and I can't seem to locate start-history-server.sh
. For what it's worth, this is for both python 3.7 and 2.7 environments.
My question is: is start-history-server.sh
included in a conda installation of pyspark?
If yes, where? If no, what's the recommended alternative way of evaluating spark jobs after the fact?