Search Options

Results per page
Sort
Preferred Languages
Advance

Results 41 - 44 of 44 for considers (0.58 sec)

  1. LICENSE

        root function must still compute square roots.)
    
    These requirements apply to the modified work as a whole.  If
    identifiable sections of that work are not derived from the Library,
    and can be reasonably considered independent and separate works in
    themselves, then this License, and its terms, do not apply to those
    sections when you distribute them as separate works.  But when you
    Registered: Sat Dec 20 13:44:44 UTC 2025
    - Last Modified: Mon Jan 18 20:25:38 UTC 2016
    - 25.8K bytes
    - Viewed (0)
  2. src/main/resources/fess_config.properties

    crawler.ignore.robots.txt=false
    # Whether to ignore robots meta tags during crawling.
    crawler.ignore.robots.tags=false
    # Whether to ignore content exceptions during crawling.
    crawler.ignore.content.exception=true
    # HTTP status codes considered as failure URLs.
    crawler.failure.url.status.codes=404
    # Interval (seconds) for system monitor during crawling.
    crawler.system.monitor.interval=60
    # Whether to ignore idle threads in hot thread monitoring.
    Registered: Sat Dec 20 09:19:18 UTC 2025
    - Last Modified: Thu Dec 11 09:47:03 UTC 2025
    - 54.8K bytes
    - Viewed (0)
  3. src/main/java/jcifs/smb1/smb1/SmbFile.java

        }
    
        /**
         * Tests to see if the file this <code>SmbFile</code> represents
         * exists and is not marked read-only. By default, resources are
         * considered to be read-only and therefore for <code>smb1://</code>,
         * <code>smb1://workgroup/</code>, and <code>smb1://server/</code> resources
         * will be read-only.
         *
    Registered: Sat Dec 20 13:44:44 UTC 2025
    - Last Modified: Sat Aug 30 05:58:03 UTC 2025
    - 112.2K bytes
    - Viewed (0)
  4. src/main/java/org/codelibs/fess/mylasta/direction/FessConfig.java

         */
        boolean isCrawlerIgnoreContentException();
    
        /**
         * Get the value for the key 'crawler.failure.url.status.codes'. <br>
         * The value is, e.g. 404 <br>
         * comment: HTTP status codes considered as failure URLs.
         * @return The value of found property. (NotNull: if not found, exception but basically no way)
         */
        String getCrawlerFailureUrlStatusCodes();
    
        /**
    Registered: Sat Dec 20 09:19:18 UTC 2025
    - Last Modified: Sat Dec 13 02:21:17 UTC 2025
    - 525.7K bytes
    - Viewed (2)
Back to top