I've got sequence of values (from a reduce by key). I know that in theory the keys are an ordered sequence of things, and I should be able to reduce over them.
I want to run a window over these sequences. I could store the window in the accumulator for the reduce function, but I think that the way Spark parallelizes the workflow (requiring functions to be commutative and associative) means that the windows may get chopped.
Is there a way to do this?