Details
-
Bug
-
Status: Resolved
-
Major
-
Resolution: Duplicate
-
None
-
None
-
None
-
None
-
~30 node cluster, various size/number of disks, CPUs, memory
Description
This appears to have been introduced with the "check for enough free space" before startup.
It's debatable how best to fix this bug. I will submit a patch which ignores directories for which the DF utility fails. This is letting me continue operation on my cluster (where the number of drives varies, so there are entries in mapred.local.dir for drives that aren't on all cluster nodes), but a cleaner solution is probably better. I'd lean towards "check for existence", and ignore the dir if it doesn't - but don't depend on DF to fail, since DF could fail for other reasons without meaning you're out of disk space. I argue that a TaskTracker should start up if all directories that can be written to in the list have enough space. Otherwise, a failed drive per cluster machine means no work ever gets done.
Attachments
Attachments
Issue Links
- is superceded by
-
MAPREDUCE-2413 TaskTracker should handle disk failures at both startup and runtime
- Closed