- Sort Score
- Num 10 results
- Language All
Results 1 - 2 of 2 for processRobotsTxt (0.06 seconds)
-
fess-crawler/src/main/java/org/codelibs/fess/crawler/client/http/Hc4HttpClient.java
* This method fetches and parses the robots.txt file to extract disallow/allow rules * and sitemap information. * * @param url The URL to process robots.txt for */ protected void processRobotsTxt(final String url) { if (StringUtil.isBlank(url)) { throw new CrawlerSystemException("HTTP URL is null or empty. Cannot process robots.txt without a valid URL."); }Created: Sun Apr 12 03:50:13 GMT 2026 - Last Modified: Fri Jan 09 23:46:52 GMT 2026 - 54.4K bytes - Click Count (0) -
fess-crawler/src/main/java/org/codelibs/fess/crawler/client/http/Hc5HttpClient.java
* This method fetches and parses the robots.txt file to extract disallow/allow rules * and sitemap information. * * @param url The URL to process robots.txt for */ protected void processRobotsTxt(final String url) { if (StringUtil.isBlank(url)) { throw new CrawlerSystemException("HTTP URL is null or empty. Cannot process robots.txt without a valid URL."); }Created: Sun Apr 12 03:50:13 GMT 2026 - Last Modified: Sat Jan 31 12:23:29 GMT 2026 - 62.2K bytes - Click Count (0)