- Sort Score
- Result 10 results
- Languages All
Results 1 - 5 of 5 for standards (0.08 sec)
-
README.md
logger.debug.org.codelibs.fess.suggest=DEBUG ``` ## Contributing 1. Fork the repository 2. Create a feature branch: `git checkout -b feature/amazing-feature` 3. Follow coding standards: `mvn formatter:format` 4. Add tests for new functionality 5. Ensure all tests pass: `mvn test` 6. Check license headers: `mvn license:format` 7. Commit changes: `git commit -m 'Add amazing feature'`
Registered: Fri Sep 19 09:08:11 UTC 2025 - Last Modified: Sun Aug 31 03:31:14 UTC 2025 - 12.1K bytes - Viewed (1) -
fess-crawler/src/main/java/org/codelibs/fess/crawler/extractor/impl/CommandExtractor.java
*/ public void setMaxOutputLine(final int maxOutputLine) { this.maxOutputLine = maxOutputLine; } /** * Sets whether to redirect standard output. * @param standardOutput true to redirect standard output, false otherwise. */ public void setStandardOutput(final boolean standardOutput) { this.standardOutput = standardOutput; }Registered: Sun Sep 21 03:50:09 UTC 2025 - Last Modified: Sun Jul 06 02:13:03 UTC 2025 - 16K bytes - Viewed (0) -
src/test/java/org/codelibs/fess/suggest/converter/AnalyzerConverterTest.java
Registered: Fri Sep 19 09:08:11 UTC 2025 - Last Modified: Mon Sep 01 13:33:03 UTC 2025 - 12.5K bytes - Viewed (0) -
src/main/java/org/codelibs/fess/suggest/util/SuggestUtil.java
import java.util.Collections; import java.util.HashMap; import java.util.List; import java.util.Map; import java.util.stream.Stream; import org.apache.lucene.queryparser.flexible.standard.StandardQueryParser; import org.apache.lucene.queryparser.flexible.standard.config.StandardQueryConfigHandler; import org.apache.lucene.search.BooleanClause; import org.apache.lucene.search.BooleanQuery; import org.apache.lucene.search.Query;
Registered: Fri Sep 19 09:08:11 UTC 2025 - Last Modified: Mon Sep 01 13:33:03 UTC 2025 - 17.4K bytes - Viewed (1) -
fess-crawler/src/main/java/org/codelibs/fess/crawler/entity/RobotsTxt.java
* This class manages the rules defined in a robots.txt file, including user agent directives, * allowed/disallowed paths, crawl delays, and sitemap URLs. * * <p>The robots.txt protocol is implemented according to the standard specification, * supporting pattern matching for user agents, path-based access control, and crawl delay settings.</p> * * <p>Key features:</p> * <ul>
Registered: Sun Sep 21 03:50:09 UTC 2025 - Last Modified: Sun Jul 06 02:13:03 UTC 2025 - 10K bytes - Viewed (0)