Class BaseRobotRules

    • Constructor Detail

      • BaseRobotRules

        public BaseRobotRules()
    • Method Detail

      • isAllowed

        public abstract boolean isAllowed​(String url)
      • isAllowAll

        public abstract boolean isAllowAll()
      • isAllowNone

        public abstract boolean isAllowNone()
      • getCrawlDelay

        public long getCrawlDelay()
        Get Crawl-delay (in milliseconds)
        Returns:
        Crawl-delay defined in the robots.txt for the given agent name, or UNSET_CRAWL_DELAY if not defined.
      • setCrawlDelay

        public void setCrawlDelay​(long crawlDelay)
        Parameters:
        crawlDelay - Crawl-Delay in milliseconds
      • isDeferVisits

        public boolean isDeferVisits()
        Returns:
        whether to defer visits to the server
      • setDeferVisits

        public void setDeferVisits​(boolean deferVisits)
        Indicate to defer visits to the server, e.g. to wait until the robots.txt becomes available.
      • addSitemap

        public void addSitemap​(String sitemap)
        Add sitemap URL to rules if not a duplicate
      • getSitemaps

        public List<String> getSitemaps()
        Get URLs of sitemap links found in robots.txt
      • hashCode

        public int hashCode()
        Overrides:
        hashCode in class Object
      • toString

        public String toString()
        Returns a string with the crawl delay as well as a list of sitemaps if they exist (and aren't more than 10).
        Overrides:
        toString in class Object