I'm running a hadoop job with mapred.reduce.tasks = 100
(just experimenting). The number of maps spawned are 537 as that depends on the input splits. Problem is the number of reducers "Running" in parallel won't go beyond 4. Even after the maps are 100% complete. Is there a way to increase the number of reducers running as the CPU usage is sub optimal and the Reduce is very slow.
I have also set mapred.tasktracker.reduce.tasks.maximum = 100
. But this doesn't seem to affect the numbers of reducers running in parallel.