Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 1 of 1 for processRobotsTxt (0.08 sec)

  1. fess-crawler/src/main/java/org/codelibs/fess/crawler/client/http/HcHttpClient.java

         * This method fetches and parses the robots.txt file to extract disallow/allow rules
         * and sitemap information.
         *
         * @param url The URL to process robots.txt for
         */
        protected void processRobotsTxt(final String url) {
            if (StringUtil.isBlank(url)) {
                throw new CrawlerSystemException("HTTP URL is null or empty. Cannot process robots.txt without a valid URL.");
            }
    
    Registered: Sat Dec 20 11:21:39 UTC 2025
    - Last Modified: Sun Nov 23 12:19:14 UTC 2025
    - 53.7K bytes
    - Viewed (0)
Back to top