I have a R script running in local: each record/row is feed into a function called func to perform some calculation. so the flow is as follows.
new <- lapply(old, func)
Ideally, using sparkR, I would expect each worker has the function func and perform calculation on a subset of "old". In this case, func is very simple and can be calculated locally (no need a distributed version of this func).
Any one know how to achieve this in using SparkR? Basically the question is if there is any support in SparkR functioning like doparallel but on multiple workers.