Description
When the output sequence files from partitioning job are large(bigger than two hdfs file block size), the second round of the job (using these sequence file as input) will start up more tasks than client want. Some times, this uncertainty make the job exceed the cluster slot capacity.
In the real project, I implemented an new Inputformat which marked as un-splitable to solve the problem. Is there any better way?