Search Options

Results per page
Sort
Preferred Languages
Advance

Results 11 - 20 of 26 for potser (0.06 sec)

  1. fess-crawler/src/main/java/org/codelibs/fess/crawler/entity/RequestData.java

                return false;
            if (getClass() != obj.getClass())
                return false;
            RequestData other = (RequestData) obj;
            return method == other.method && Objects.equals(url, other.url)
                    && Float.floatToIntBits(weight) == Float.floatToIntBits(other.weight);
        }
    
        /**
         * Returns a string representation of this object.
         * @return A string representation.
    Registered: Sun Sep 21 03:50:09 UTC 2025
    - Last Modified: Sun Jul 06 02:13:03 UTC 2025
    - 4K bytes
    - Viewed (0)
  2. fess-crawler/src/test/java/org/codelibs/fess/crawler/filter/UrlFilterTest.java

            assertFalse(urlFilter.match("https://other.com/page"));
    
            // Clear the filter
            urlFilter.clear();
    
            // After clear, all URLs should match (no filters applied)
            assertTrue(urlFilter.match("https://example.com/page"));
            assertTrue(urlFilter.match("https://other.com/page"));
            assertTrue(urlFilter.match("https://any.com/image.jpg"));
        }
    
    Registered: Sun Sep 21 03:50:09 UTC 2025
    - Last Modified: Wed Sep 03 14:42:53 UTC 2025
    - 19K bytes
    - Viewed (0)
  3. .github/ISSUE_TEMPLATE/bug_report.md

    **Screenshots**
    If applicable, add screenshots to help explain your problem.
    
    **Environment (please complete the following information):**
     - OS: [e.g. Windows 10]
     - Version [e.g. 3.5.0]
    
    **Additional context**
    Registered: Sun Sep 21 03:50:09 UTC 2025
    - Last Modified: Mon Feb 10 22:18:26 UTC 2020
    - 748 bytes
    - Viewed (0)
  4. fess-crawler/src/main/java/org/codelibs/fess/crawler/util/IgnoreCloseInputStream.java

    import java.io.InputStream;
    
    /**
     * A wrapper for an {@link InputStream} that ignores the {@link #close()} method call.
     * This can be useful when you want to prevent the underlying input stream from being closed.
     *
     * <p>All other methods delegate to the wrapped input stream.</p>
     *
     */
    public class IgnoreCloseInputStream extends InputStream {
    
        /** The wrapped input stream. */
        private transient InputStream inputStream;
    
        /**
    Registered: Sun Sep 21 03:50:09 UTC 2025
    - Last Modified: Sun Jul 06 02:13:03 UTC 2025
    - 5.3K bytes
    - Viewed (0)
  5. fess-crawler/src/main/java/org/codelibs/fess/crawler/exception/SitemapsException.java

    /**
     * Exception thrown during sitemaps processing in the crawler.
     * This exception extends {@link org.codelibs.fess.crawler.exception.CrawlerSystemException}
     * and provides constructors for wrapping other exceptions or creating a new exception with a message.
     */
    public class SitemapsException extends CrawlerSystemException {
        private static final long serialVersionUID = 1L;
    
        /**
    Registered: Sun Sep 21 03:50:09 UTC 2025
    - Last Modified: Sun Jul 06 02:13:03 UTC 2025
    - 1.7K bytes
    - Viewed (0)
  6. fess-crawler/src/main/java/org/codelibs/fess/crawler/entity/SitemapUrl.java

         */
        private String changefreq;
    
        /**
         * The priority of this URL relative to other URLs on your site. Valid
         * values range from 0.0 to 1.0. This value does not affect how your pages
         * are compared to pages on other sites—it only lets the search engines know
         * which pages you deem most important for the crawlers.
         *
         * The default priority of a page is 0.5.
    Registered: Sun Sep 21 03:50:09 UTC 2025
    - Last Modified: Sun Jul 06 02:13:03 UTC 2025
    - 6.5K bytes
    - Viewed (0)
  7. fess-crawler/src/main/java/org/codelibs/fess/crawler/exception/CrawlingAccessException.java

     * It extends CrawlerSystemException and provides functionality to set and check the log level for the exception.
     *
     * <p>
     * This exception can be thrown when there are problems accessing URLs, files, or any other resources needed for crawling.
     * It includes constructors to handle messages, causes, or both.
     * </p>
     *
     * <p>
    Registered: Sun Sep 21 03:50:09 UTC 2025
    - Last Modified: Sun Jul 06 02:13:03 UTC 2025
    - 3.8K bytes
    - Viewed (0)
  8. fess-crawler/src/test/java/org/codelibs/fess/crawler/rule/RuleTest.java

            responseData2.setUrl("https://api.example.com/v1/users");
            assertTrue(rule.match(responseData2));
    
            ResponseData responseData3 = new ResponseData();
            responseData3.setUrl("http://other.com/page");
            assertFalse(rule.match(responseData3));
        }
    
        /**
         * Test ConfigurableRule with MIME type condition
         */
        public void test_configurableRule_mimeTypeCondition() {
    Registered: Sun Sep 21 03:50:09 UTC 2025
    - Last Modified: Wed Sep 03 14:42:53 UTC 2025
    - 22.7K bytes
    - Viewed (0)
  9. fess-crawler/src/test/java/org/codelibs/fess/crawler/extractor/impl/ApiExtractorTest.java

            final ExtractData text = extractor.getText(new ByteArrayInputStream(testStr.getBytes()), params);
            assertEquals(content, text.getContent());
        }
    
        // TODO other tests
    
        static class TestApiExtractorServer {
            private Server server;
    
            public TestApiExtractorServer(final int port) {
                server = new Server(port);
    
    Registered: Sun Sep 21 03:50:09 UTC 2025
    - Last Modified: Sat Mar 15 06:52:00 UTC 2025
    - 5.4K bytes
    - Viewed (0)
  10. fess-crawler/src/main/java/org/codelibs/fess/crawler/entity/RobotsTxt.java

    import java.util.ArrayList;
    import java.util.LinkedHashMap;
    import java.util.List;
    import java.util.Map;
    import java.util.regex.Pattern;
    
    import org.codelibs.core.lang.StringUtil;
    
    /**
     * Represents a robots.txt file parser and handler.
     * This class manages the rules defined in a robots.txt file, including user agent directives,
     * allowed/disallowed paths, crawl delays, and sitemap URLs.
     *
    Registered: Sun Sep 21 03:50:09 UTC 2025
    - Last Modified: Sun Jul 06 02:13:03 UTC 2025
    - 10K bytes
    - Viewed (0)
Back to top