Uploaded image for project: 'Nutch'
  1. Nutch
  2. NUTCH-1184

Fetcher to parse and follow Nth degree outlinks



    • Type: New Feature
    • Status: Closed
    • Priority: Major
    • Resolution: Fixed
    • Affects Version/s: None
    • Fix Version/s: 1.5
    • Component/s: fetcher
    • Labels:
    • Patch Info:
      Patch Available


      Fetcher improvements to parse and follow outlinks up to a specified depth. The number of outlinks to follow can be decreased by depth using a divisor. This patch introduces three new configuration directives:

        <description>(EXPERT)When fetcher.parse is true and this value is greater than 0 the fetcher will extract outlinks
        and follow until the desired depth is reached. A value of 1 means all generated pages are fetched and their first degree
        outlinks are fetched and parsed too. Be careful, this feature is in itself agnostic of the state of the CrawlDB and does not
        know about already fetched pages. A setting larger than 2 will most likely fetch home pages twice in the same fetch cycle.
        It is highly recommended to set db.ignore.external.links to true to restrict the outlink follower to URL's within the same
        domain. When disabled (false) the feature is likely to follow duplicates even when depth=1.
        A value of -1 of 0 disables this feature.
        <description>(EXPERT)The number of outlinks to follow when fetcher.follow.outlinks.depth is enabled. Be careful, this can multiply
        the total number of pages to fetch. This works with fetcher.follow.outlinks.depth.divisor, by default settings the followed outlinks
        at depth 1 is 8, not 4.
        <description>(EXPERT)The divisor of fetcher.follow.outlinks.num.links per fetcher.follow.outlinks.depth. This decreases the number
        of outlinks to follow by increasing depth. The formula used is: outlinks = floor(divisor / depth * num.links). This prevents
        exponential growth of the fetch list.

      Please, do not use this unless you know what you're doing. This feature does not consider the state of the CrawlDB nor does it consider generator settings such as limiting the number of pages per (domain|host|ip) queue. It is not polite to use this feature with high settings as it can fetch many pages from the same domain including duplicates.

      Also, this feature will not work if fetcher.parse is disabled. With parsing enabled you might want to consider not to store downloaded content.


        1. NUTCH-1184-1.5-1.patch
          9 kB
          Markus Jelsma
        2. NUTCH-1184-1.5-2.patch
          8 kB
          Markus Jelsma
        3. NUTCH-1184-1.5-3.patch
          13 kB
          Markus Jelsma
        4. NUTCH-1184-1.5-4.patch
          13 kB
          Markus Jelsma
        5. NUTCH-1184-1.5-5.patch
          23 kB
          Markus Jelsma
        6. NUTCH-1184-1.5-5-ParseData.patch
          0.5 kB
          Markus Jelsma
        7. NUTCH-1185-1.5-6.patch
          31 kB
          Markus Jelsma
        8. NUTCH-1185-1.5-7.patch
          31 kB
          Markus Jelsma
        9. NUTCH-1185-1.5-8.patch
          32 kB
          Markus Jelsma
        10. NUTCH-1185-1.5-9.patch
          33 kB
          Markus Jelsma
        11. NUTCH-1184-1.5-9-ParseOutputFormat.patch
          6 kB
          Markus Jelsma

          Issue Links



              • Assignee:
                markus17 Markus Jelsma
                markus17 Markus Jelsma
              • Votes:
                0 Vote for this issue
                0 Start watching this issue


                • Created: