Package | Description |
---|---|
crawlercommons.robots |
The robots package contains all of the robots.txt rule inference, parsing and utilities contained within Crawler Commons.
|
Class and Description |
---|
BaseRobotRules
Result from parsing a single robots.txt file - which means we get a set of
rules, and a crawl-delay.
|
BaseRobotsParser |
SimpleRobotRules.RobotRule
Single rule that maps from a path prefix to an allow flag.
|
SimpleRobotRules.RobotRulesMode |
Copyright © 2009–2016 Crawler-Commons. All rights reserved.