I'm submitting a spark job from a shell script that has a bunch of env vars and parameters to pass to spark. Strangely, the driver host is not one of these parameters (there are driver cores and memory however). So if I have 3 machines in the cluster, a driver will be chosen randomly. I don't want this behaviour since 1) the jar I'm submitting is only on one of the machines and 2) the driver machine should often be smaller than the other machines which is not the case if it's random choice.
So far, I found no way to specify this param on the command line to spark-submit
. I've tried --conf SPARK_DRIVER_HOST="172.30.1.123
, --conf spark.driver.host="172.30.1.123
and many other things but nothing has any effect. I'm using spark 2.1.0. Thanks.