Search Options

Results per page
Sort
Preferred Languages
Advance

Results 11 - 20 of 104 for url (0.03 sec)

  1. fess-crawler/src/main/java/org/codelibs/fess/net/protocol/storage/Handler.java

            /**
             * Constructs a new StorageURLConnection for the specified URL.
             * This constructor parses the URL to extract bucket and object names.
             *
             * @param url The storage URL to connect to
             */
            protected StorageURLConnection(final URL url) {
                super(url);
                final String[] values = url.toExternalForm().split("/", 2);
                if (values.length == 2) {
    Registered: Sun Sep 21 03:50:09 UTC 2025
    - Last Modified: Sun Jul 06 02:13:03 UTC 2025
    - 10.5K bytes
    - Viewed (0)
  2. fess-crawler/src/main/java/org/codelibs/fess/crawler/entity/AccessResult.java

        /**
         * Returns the URL of the accessed resource.
         *
         * @return the URL
         */
        String getUrl();
    
        /**
         * Sets the URL of the accessed resource.
         *
         * @param url the URL
         */
        void setUrl(String url);
    
        /**
         * Returns the parent URL of the accessed resource.
         *
         * @return the parent URL
         */
        String getParentUrl();
    
        /**
    Registered: Sun Sep 21 03:50:09 UTC 2025
    - Last Modified: Sun Jul 06 02:13:03 UTC 2025
    - 5.2K bytes
    - Viewed (10)
  3. pom.xml

    	<url>https://github.com/codelibs/fess-crawler</url>
    	<inceptionYear>2009</inceptionYear>
    	<licenses>
    		<license>
    			<name>The Apache Software License, Version 2.0</name>
    			<url>https://www.apache.org/licenses/LICENSE-2.0.txt</url>
    			<distribution>repo</distribution>
    		</license>
    	</licenses>
    	<organization>
    		<name>CodeLibs</name>
    		<url>https://fess.codelibs.org/</url>
    	</organization>
    Registered: Sun Sep 21 03:50:09 UTC 2025
    - Last Modified: Sat Sep 06 04:15:37 UTC 2025
    - 2.8K bytes
    - Viewed (0)
  4. fess-crawler/src/main/java/org/codelibs/fess/crawler/entity/RequestData.java

            }
        }
    
        /**
         * Gets the URL for this request.
         * @return the URL
         */
        public String getUrl() {
            return url;
        }
    
        /**
         * Sets the URL for this request.
         * @param url the URL
         */
        public void setUrl(final String url) {
            this.url = url;
        }
    
        /**
         * Gets the weight/priority of this request.
    Registered: Sun Sep 21 03:50:09 UTC 2025
    - Last Modified: Sun Jul 06 02:13:03 UTC 2025
    - 4K bytes
    - Viewed (0)
  5. fess-crawler/src/main/java/org/codelibs/fess/crawler/interval/impl/DefaultIntervalController.java

        /** Delay in milliseconds after processing a URL */
        protected long delayMillisAfterProcessing = 0L;
    
        /** Delay in milliseconds when no URL is in the queue */
        protected long delayMillisAtNoUrlInQueue = 500L;
    
        /** Delay in milliseconds before processing a URL */
        protected long delayMillisBeforeProcessing = 0L;
    
    Registered: Sun Sep 21 03:50:09 UTC 2025
    - Last Modified: Sun Jul 06 02:13:03 UTC 2025
    - 3.4K bytes
    - Viewed (0)
  6. fess-crawler/src/main/java/org/codelibs/fess/crawler/service/impl/UrlFilterServiceImpl.java

    /**
     * Implementation of the {@link UrlFilterService} interface.
     * This class provides methods for managing URL filtering rules,
     * including adding include and exclude URL patterns, deleting patterns,
     * and retrieving lists of compiled URL patterns. It utilizes a
     * {@link MemoryDataHelper} to store and manage the URL patterns in memory.
     *
     */
    public class UrlFilterServiceImpl implements UrlFilterService {
    
        /**
    Registered: Sun Sep 21 03:50:09 UTC 2025
    - Last Modified: Sun Jul 06 02:13:03 UTC 2025
    - 4.2K bytes
    - Viewed (0)
  7. fess-crawler/src/test/java/org/codelibs/fess/net/protocol/storage/HandlerTest.java

            super.setUp();
            new StandardCrawlerContainer();
        }
    
        public void test_protocol() throws Exception {
            URL url = new URL("storage:/home/foo");
            assertEquals("storage", url.getProtocol());
        }
    Registered: Sun Sep 21 03:50:09 UTC 2025
    - Last Modified: Sat Mar 15 06:52:00 UTC 2025
    - 1.1K bytes
    - Viewed (0)
  8. fess-crawler/src/test/java/org/codelibs/fess/crawler/client/CrawlerClientFactoryTest.java

        public void test_getClient_null() {
            String url;
            CrawlerClient client;
    
            url = null;
            client = clientFactory.getClient(url);
            assertNull(client);
    
            url = "";
            client = clientFactory.getClient(url);
            assertNull(client);
    
            url = " ";
            client = clientFactory.getClient(url);
            assertNull(client);
        }
    Registered: Sun Sep 21 03:50:09 UTC 2025
    - Last Modified: Sat Mar 15 06:52:00 UTC 2025
    - 5.3K bytes
    - Viewed (0)
  9. fess-crawler/src/main/java/org/codelibs/fess/crawler/service/UrlFilterService.java

    /**
     * Service interface for managing URL filters.
     * Provides methods to add and remove include/exclude URL filters,
     * as well as retrieve the patterns of these filters.
     */
    public interface UrlFilterService {
    
        /**
         * Adds a URL to the include filter list for the specified session.
         *
         * @param sessionId the ID of the session for which the URL filter is being added
    Registered: Sun Sep 21 03:50:09 UTC 2025
    - Last Modified: Sat Mar 15 06:52:00 UTC 2025
    - 3.1K bytes
    - Viewed (0)
  10. fess-crawler/src/main/java/org/codelibs/fess/crawler/client/http/HcHttpClient.java

        /**
         * Processes robots.txt for the given URL.
         * This method fetches and parses the robots.txt file to extract disallow/allow rules
         * and sitemap information.
         *
         * @param url The URL to process robots.txt for
         */
        protected void processRobotsTxt(final String url) {
            if (StringUtil.isBlank(url)) {
                throw new CrawlerSystemException("url is null or empty.");
            }
    
    Registered: Sun Sep 21 03:50:09 UTC 2025
    - Last Modified: Thu Aug 07 02:55:08 UTC 2025
    - 52.2K bytes
    - Viewed (0)
Back to top