Uploaded image for project: 'Hadoop Common'
  1. Hadoop Common
  2. HADOOP-3333

job failing because of reassigning same tasktracker to failing tasks

    XMLWordPrintableJSON

Details

    • Bug
    • Status: Closed
    • Blocker
    • Resolution: Fixed
    • 0.16.3
    • 0.18.0
    • None
    • None
    • Reviewed

    Description

      We have a long running a job in a 2nd atttempt. Previous job was failing and current jobs risks to fail as well, because reduce tasks failing on marginal TaskTrackers are assigned repeatedly to the same TaskTrackers (probably because it is the only available slot), eventually running out of attempts.
      Reduce tasks should be assigned to the same TaskTrackers at most twice, or TaskTrackers need to get some better smarts to find failing hardware.
      BTW, mapred.reduce.max.attempts=12, which is high, but does not help in this case.

      Attachments

        1. hadoop-3333-v3.patch
          17 kB
          Jothi Padmanabhan
        2. hadoop-3333-v2.patch
          16 kB
          Jothi Padmanabhan
        3. hadoop-3333-v1.patch
          16 kB
          Jothi Padmanabhan
        4. hadoop-3333.patch
          16 kB
          Jothi Padmanabhan
        5. HADOOP-3333_2_20080506.patch
          12 kB
          Arun Murthy
        6. HADOOP-3333_1_20080505.patch
          11 kB
          Arun Murthy
        7. HADOOP-3333_0_20080503.patch
          9 kB
          Arun Murthy

        Issue Links

          Activity

            People

              jothipn Jothi Padmanabhan
              ckunz Christian Kunz
              Votes:
              0 Vote for this issue
              Watchers:
              0 Start watching this issue

              Dates

                Created:
                Updated:
                Resolved: