I understand Below are daemons required for spark cluster
- Master
- Workers (Slave)
- Driver(Launched when Application is submmited)
- Executors(Launched when Application is submmited)
I have some very basic questions on Spark when its being set up on yarn cluster
- Are there any master daemon or worker daemons started separately for spark ? I understand Resource manager and nodemanager of yarn cluster itself will act as master and workers for spark jobs. From this article http://blog.cloudera.com/blog/2014/05/apache-spark-resource-management-and-yarn-app-models/, it looks like there are are no master/slave daemons separately for spark on yarn.
- If answer to above question is NO. When we are trying to set up spark on existing yarn, do we need to start any Persistent daemon at all before submitting the spark application ?
- Any of the start-stop scripts inside spark-1.5.0-bin-hadoop2.4\sbin directory will be useful in this case at all ?
- Spark WEB UI is not available once driver is finished its execution . Am I correct ?