I'm seeing a strange problem. I have a spark cluster in standalone mode. I submit spark jobs from a remote node as follows from the terminal
$> spark-submit --master spark://10.1.40.18:7077 --class com.test.Ping spark-jobs.jar
when the app is running , when I press ctrl-C on the console terminal, then the process is killed and so is the app in the spark master UI. When I go to spark master ui, i see that this app is in state Killed under Completed applications, which is what I expected to see.
Now, I created a shell script as follows to do the same
#!/bin/bash
spark-submit --master spark://10.1.40.18:7077 --class com.test.Ping spark-jobs.jar &
echo $! > my.pid
When I execute the shell script from terminal, as follows
$> bash myscript.sh
The application is submitted correctly to spark master and I can see it as one of the running apps in the spark master UI. But when I kill the process in my terminal as follows
$> ps kill $(cat my.pid)
I see that the process is killed on my machine but the spark application is still running in spark master! It doesn't get killed.
I noticed one more thing that, when I launch the spark job via shell script and kill the application from spark master UI by clicking on "kill" next to the running application, it gets killed in spark ui but I still see the process running in my machine.
In both cases, I would expect the remote spark app to be killed and my local process to be killed.
Why is this happening? and how can I kill a spark app from the terminal launced via shell script w.o going to the spark master UI?
I want to launch the spark app via script and log the pid so i can monitor it remotely
thanks for the help