I want to use some of the classifiers provided by MLLib (random forests, etc) but I want to use them without connecting to a Spark cluster.
If I need to somehow run some Spark stuff in-process so that I have a Spark context to use, that's fine. But I haven't been able to find any information or an example for such a use case.
So my two questions are:
- Is there a way to use the MLLib classifiers without a Spark context at all?
- Otherwise, can I use them by starting a Spark context in-process, without needing any kind of actual Spark installation?