Home > Software > BIGDATA > HADOOP
Interview Questions   Tutorials   Discussions   Programs   Videos   Discussion   

HADOOP - How many Reducers should be configured?

asked SRVMTrainings November 11, 2012 02:10 AM  

How many Reducers should be configured?


1 Answers

answered By   0  
The right number of reduces seems to be 0.95 or 1.75 multiplied by (<no. of nodes> * mapreduce.tasktracker.reduce.tasks.maximum).
With 0.95 all of the reduces can launch immediately and start transfering map outputs as the maps finish. With 1.75 the faster nodes will finish their first round of reduces and launch a second wave of reduces doing a much better job of load balancing.
Increasing the number of reduces increases the framework overhead, but increases load balancing and lowers the cost of failures.
The scaling factors above are slightly less than whole numbers to reserve a few reduce slots in the framework for speculative-tasks and failed tasks.
   add comment

Your answer

Join with account you already have



 Write A Tutorials
Online-Classroom Classes

  1 person following this question

  5 people following this tag

  Question tags

hadoop × 7

Asked 5 years and 2 months ago ago
Number of Views -470
Number of Answers -1
Last updated
5 years and 29 days ago ago

Ready to start your tutorial with us? That's great! Send us an email and we will get back to you as soon as possible!