Uploaded image for project: 'Nutch'
  1. Nutch
  2. NUTCH-1978

solrindex will fail when indexing corrupted segments

    XMLWordPrintableJSON

Details

    • Bug
    • Status: Closed
    • Minor
    • Resolution: Duplicate
    • 1.10
    • 1.10
    • indexer
    • None

    Description

      The same issue from NUTCH-1771 but seems like this bug will appear in most of the versions since they all don't have the code to handle the corrupted segments.

      Form NUTCH-1771, people pointed out that it will be very hard to handle this in the hadoop layer, and the program should skip the corrupted segments instead of end the program. By corrupted segments I mean that the segment may be just generated and doesn't have the content.

      So my initial idea is to check if the segment folder is valid before putting the segment into the hadoop job. If the segment is not valid, we can simply just skip that segment. We can check if the segment folder contains exactly 6 sub directories as there should be. The other approach will be to check all the six sub directories and see if they are exactly the six dir that should appear.

      Attachments

        Issue Links

          Activity

            People

              Unassigned Unassigned
              chongli Chong Li
              Votes:
              0 Vote for this issue
              Watchers:
              3 Start watching this issue

              Dates

                Created:
                Updated:
                Resolved: