I was trying performance testing of a Map reduce program in a cluster of 10 Data nodes. During the process I strated with 5 Reducers , then 10 and so on.
I was thinking that increasing the number of reducers would also make the job finis faster, but what I observed that after 30, even you increase the number till 70 the job runtime didnt make any difference, though there are more then 400 unique keys and increasing the number of reducers beyond 30 should make a difference Isn't it?
Can anyone help me to understand if there could be any particular reason for this?
Thanks in advance