Uses of Class
crawlercommons.robots.BaseRobotRules
-
Packages that use BaseRobotRules Package Description crawlercommons.robots The robots package contains all of the robots.txt rule inference, parsing and utilities contained within Crawler Commons. -
-
Uses of BaseRobotRules in crawlercommons.robots
Subclasses of BaseRobotRules in crawlercommons.robots Modifier and Type Class Description class
SimpleRobotRules
Result from parsing a single robots.txt file - which means we get a set of rules, and an optional crawl-delay, and an optional sitemap URL.Methods in crawlercommons.robots that return BaseRobotRules Modifier and Type Method Description abstract BaseRobotRules
BaseRobotsParser. failedFetch(int httpStatusCode)
The fetch of robots.txt failed, so return rules appropriate give the HTTP status code.abstract BaseRobotRules
BaseRobotsParser. parseContent(String url, byte[] content, String contentType, String robotNames)
Parse the robots.txt file in content, and return rules appropriate for processing paths by userAgent.
-