Details
-
Bug
-
Status: Closed
-
Minor
-
Resolution: Won't Fix
-
0.7
-
None
-
None
Description
Here's a simple example that the current RobotRulesParser gets wrong:
User-agent: *
Disallow: /
Allow: /rss
The problem is that the isAllowed function takes the first rule that matches and incorrectly decides that URLs starting with "/rss" are Disallowed. The correct algorithm is to take the longest rule that matches. I will attach a patch that fixes this.