Details
-
Bug
-
Status: Closed
-
Major
-
Resolution: Fixed
-
0.20.2, 0.20.205.0
-
None
-
None
Description
If there's a reduce task that needs more disk space than is available on any mapred.local.dir in the cluster, that task will stay pending forever. For example, we produced this in a QA cluster by accidentally running terasort with one reducer - since no mapred.local.dir had 1T free, the job remained in pending state for several days. The reason for the "stuck" task wasn't clear from a user perspective until we looked at the JT logs.
Probably better to just fail the job if a reduce task goes through all TTs and finds that there isn't enough space.