2

I have three Spark Streaming jobs that use ConsumerStrategies.Assign[]() to seek the latest offset that was committed into a database.

Each one of these jobs is reading from one of three partitions in a topic (for example: partition 0,1 and 2). If one of the Spark Streaming jobs fails, is it possible to rebalance that partition to one of the other two jobs that are currently running.

I know you can do that in normal Kafka using ConsumerRebalanceListener, onPartitionsProvoked() and onPartitionsAssigned(). But how would you do that in Spark-Streaming-Kafka?

pheeleeppoo
  • 1,491
  • 6
  • 25
  • 29

0 Answers0