Search Options

Results per page
Sort
Preferred Languages
Advance

Results 21 - 30 of 38 for Tokenizer (0.14 sec)

  1. src/main/resources/fess_indices/fess_config.file_config.json

          "number_of_shards": 1,
          "number_of_replicas": 0,
          "auto_expand_replicas": "0-1"
        },
        "analysis": {
          "analyzer": {
            "standard_analyzer": {
              "type": "custom",
              "tokenizer": "standard",
              "filter": [
                "cjk_width",
                "asciifolding",
                "lowercase",
                "stop",
                "stemmer"
              ]
            }
          }
        }
      }
    Registered: Thu Oct 31 13:40:30 UTC 2024
    - Last Modified: Thu Dec 02 13:14:56 UTC 2021
    - 484 bytes
    - Viewed (0)
  2. src/main/resources/fess_indices/fess_config.web_config.json

          "number_of_shards": 1,
          "number_of_replicas": 0,
          "auto_expand_replicas": "0-1"
        },
        "analysis": {
          "analyzer": {
            "standard_analyzer": {
              "type": "custom",
              "tokenizer": "standard",
              "filter": [
                "cjk_width",
                "asciifolding",
                "lowercase",
                "stop",
                "stemmer"
              ]
            }
          }
        }
      }
    Registered: Thu Oct 31 13:40:30 UTC 2024
    - Last Modified: Thu Dec 02 13:14:56 UTC 2021
    - 484 bytes
    - Viewed (0)
  3. compat/maven-artifact/src/main/java/org/apache/maven/artifact/versioning/DefaultArtifactVersion.java

                    }
                }
                if (idx < tok.length) {
                    qualifier = tok[idx++];
                    fallback = isDigits(qualifier);
                }
    
                // string tokenizer won't detect these and ignores them
                if (part1.contains("..") || part1.startsWith(".") || part1.endsWith(".")) {
                    fallback = true;
                }
    
                if (fallback) {
    Registered: Sun Nov 03 03:35:11 UTC 2024
    - Last Modified: Fri Oct 25 12:31:46 UTC 2024
    - 6K bytes
    - Viewed (0)
  4. src/test/java/org/codelibs/opensearch/extension/analysis/NGramSynonymTokenizer.java

     */
    
    import java.io.IOException;
    import java.util.ArrayList;
    import java.util.Comparator;
    import java.util.List;
    import java.util.PriorityQueue;
    
    import org.apache.lucene.analysis.Tokenizer;
    import org.apache.lucene.analysis.synonym.SynonymMap;
    import org.apache.lucene.analysis.tokenattributes.CharTermAttribute;
    import org.apache.lucene.analysis.tokenattributes.OffsetAttribute;
    Registered: Fri Nov 08 09:08:12 UTC 2024
    - Last Modified: Thu Feb 22 01:36:54 UTC 2024
    - 17K bytes
    - Viewed (0)
  5. src/main/java/org/codelibs/fess/suggest/converter/KatakanaConverter.java

            final Reader rd = new StringReader(inputStr);
            try (TokenStream stream = createTokenStream(rd)) {
                if (stream == null) {
                    throw new IOException("Invalid tokenizer.");
                }
                stream.reset();
    
                int offset = 0;
                while (stream.incrementToken()) {
                    final CharTermAttribute att = stream.getAttribute(CharTermAttribute.class);
    Registered: Fri Nov 08 09:08:12 UTC 2024
    - Last Modified: Thu Feb 22 01:36:54 UTC 2024
    - 4.7K bytes
    - Viewed (0)
  6. src/cmd/asm/internal/lex/input.go

    		if i > 0 {
    			name, value = name[:i], name[i+1:]
    		}
    		tokens := Tokenize(name)
    		if len(tokens) != 1 || tokens[0].ScanToken != scanner.Ident {
    			fmt.Fprintf(os.Stderr, "asm: parsing -D: %q is not a valid identifier name\n", tokens[0])
    			flags.Usage()
    		}
    		macros[name] = &Macro{
    			name:   name,
    			args:   nil,
    			tokens: Tokenize(value),
    		}
    	}
    	return macros
    }
    
    var panicOnError bool // For testing.
    Registered: Tue Nov 05 11:13:11 UTC 2024
    - Last Modified: Fri Sep 06 13:17:27 UTC 2024
    - 12.5K bytes
    - Viewed (0)
  7. src/cmd/asm/internal/asm/pseudo_test.go

    package asm
    
    import (
    	"strings"
    	"testing"
    
    	"cmd/asm/internal/lex"
    )
    
    func tokenize(s string) [][]lex.Token {
    	res := [][]lex.Token{}
    	if len(s) == 0 {
    		return res
    	}
    	for _, o := range strings.Split(s, ",") {
    		res = append(res, lex.Tokenize(o))
    	}
    	return res
    }
    
    func TestErroneous(t *testing.T) {
    
    	type errtest struct {
    		pseudo   string
    		operands string
    Registered: Tue Nov 05 11:13:11 UTC 2024
    - Last Modified: Tue Aug 29 07:48:38 UTC 2023
    - 3.1K bytes
    - Viewed (0)
  8. src/cmd/asm/internal/asm/expr_test.go

    }
    
    func TestExpr(t *testing.T) {
    	p := NewParser(nil, nil, nil) // Expression evaluation uses none of these fields of the parser.
    	for i, test := range exprTests {
    		p.start(lex.Tokenize(test.input))
    		result := int64(p.expr())
    		if result != test.output {
    			t.Errorf("%d: %q evaluated to %d; expected %d", i, test.input, result, test.output)
    		}
    		tok := p.next()
    Registered: Tue Nov 05 11:13:11 UTC 2024
    - Last Modified: Tue Aug 29 07:48:38 UTC 2023
    - 3.2K bytes
    - Viewed (0)
  9. src/cmd/asm/internal/lex/lex.go

    type Macro struct {
    	name   string   // The #define name.
    	args   []string // Formal arguments.
    	tokens []Token  // Body of macro.
    }
    
    // Tokenize turns a string into a list of Tokens; used to parse the -D flag and in tests.
    func Tokenize(str string) []Token {
    	t := NewTokenizer("command line", strings.NewReader(str), nil)
    	var tokens []Token
    	for {
    		tok := t.Next()
    		if tok == scanner.EOF {
    			break
    Registered: Tue Nov 05 11:13:11 UTC 2024
    - Last Modified: Tue Aug 29 18:31:05 UTC 2023
    - 4.1K bytes
    - Viewed (0)
  10. build-logic/documentation/src/main/groovy/gradlebuild/docs/dsl/docbook/model/ClassDoc.groovy

            this.className = className
            id = className
            simpleName = className.tokenize('.').last()
            this.classMetaData = classMetaData
            this.extensionMetaData = extensionMetaData
    
            classSection = targetDocument.createElement('chapter')
    
            classContent.childNodes.each { Node n ->
    Registered: Wed Nov 06 11:36:14 UTC 2024
    - Last Modified: Wed Dec 09 08:14:05 UTC 2020
    - 6.2K bytes
    - Viewed (0)
Back to top