Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 20 for Tokenizer (0.2 sec)

  1. src/cmd/asm/internal/lex/tokenizer.go

    	return i > 0 && unicode.IsDigit(ch)
    }
    
    func (t *Tokenizer) Text() string {
    	switch t.tok {
    	case LSH:
    		return "<<"
    	case RSH:
    		return ">>"
    	case ARR:
    		return "->"
    	case ROT:
    		return "@>"
    	}
    	return t.s.TokenText()
    }
    
    func (t *Tokenizer) File() string {
    	return t.base.Filename()
    }
    
    func (t *Tokenizer) Base() *src.PosBase {
    	return t.base
    }
    
    Registered: Tue Nov 05 11:13:11 UTC 2024
    - Last Modified: Thu Aug 04 20:35:21 UTC 2022
    - 3K bytes
    - Viewed (0)
  2. src/main/java/org/codelibs/core/text/Tokenizer.java

        static {
            setup(defaultCtype);
        }
    
        /**
         * {@link Tokenizer}を作成します。
         *
         * @param str
         *            文字列。{@literal null}であってはいけません
         */
        public Tokenizer(final String str) {
            this(str, defaultCtype);
        }
    
        /**
         * {@link Tokenizer}を作成します。
         *
         * @param str
         *            文字列。{@literal null}であってはいけません
         * @param ctype
    Registered: Fri Nov 01 20:58:10 UTC 2024
    - Last Modified: Thu Mar 07 01:59:08 UTC 2024
    - 8.6K bytes
    - Viewed (0)
  3. src/test/java/org/codelibs/opensearch/extension/analysis/ReloadableKuromojiTokenizerFactory.java

                    tokenizer.setReader(inputPending);
                }
                tokenizer.reset();
            }
    
            @Override
            public boolean incrementToken() throws IOException {
                return tokenizer.incrementToken();
            }
    
            @Override
            public void end() throws IOException {
                tokenizer.end();
            }
    
            @Override
    Registered: Fri Nov 08 09:08:12 UTC 2024
    - Last Modified: Thu Feb 22 01:36:54 UTC 2024
    - 9.5K bytes
    - Viewed (0)
  4. src/test/java/org/codelibs/core/text/TokenizerTest.java

        public void testEOF() throws Exception {
            final Tokenizer tokenizer = new Tokenizer("");
            assertThat(tokenizer.nextToken(), is(Tokenizer.TT_EOF));
            assertThat(tokenizer.nextToken(), is(Tokenizer.TT_EOF));
        }
    
        /**
         * @throws Exception
         */
        @Test
        public void testWhitespace() throws Exception {
            final Tokenizer tokenizer = new Tokenizer("\t       \n");
    Registered: Fri Nov 01 20:58:10 UTC 2024
    - Last Modified: Thu Mar 07 01:59:08 UTC 2024
    - 2K bytes
    - Viewed (0)
  5. src/main/java/jcifs/http/Handler.java

                }
                if ( handler == null ) {
                    String path = System.getProperty(HANDLER_PKGS_PROPERTY);
                    StringTokenizer tokenizer = new StringTokenizer(path, "|");
                    while ( tokenizer.hasMoreTokens() ) {
                        String provider = tokenizer.nextToken().trim();
                        if ( provider.equals("jcifs") )
                            continue;
    Registered: Sun Nov 03 00:10:13 UTC 2024
    - Last Modified: Sun Jul 01 13:12:10 UTC 2018
    - 6.9K bytes
    - Viewed (0)
  6. src/main/java/jcifs/smb1/http/Handler.java

                }
                if (handler == null) {
                    String path = System.getProperty(HANDLER_PKGS_PROPERTY);
                    StringTokenizer tokenizer = new StringTokenizer(path, "|");
                    while (tokenizer.hasMoreTokens()) {
                        String provider = tokenizer.nextToken().trim();
                        if (provider.equals("jcifs.smb1")) continue;
    Registered: Sun Nov 03 00:10:13 UTC 2024
    - Last Modified: Fri Mar 22 20:39:42 UTC 2019
    - 6.1K bytes
    - Viewed (0)
  7. src/test/java/org/codelibs/opensearch/extension/kuromoji/index/analysis/KuromojiTokenizerFactory.java

    import java.io.IOException;
    import java.io.StringReader;
    import java.util.HashSet;
    import java.util.List;
    import java.util.Set;
    
    import org.apache.lucene.analysis.Tokenizer;
    import org.apache.lucene.analysis.ja.JapaneseTokenizer;
    import org.apache.lucene.analysis.ja.JapaneseTokenizer.Mode;
    import org.apache.lucene.analysis.ja.dict.UserDictionary;
    import org.apache.lucene.analysis.util.CSVUtil;
    Registered: Fri Nov 08 09:08:12 UTC 2024
    - Last Modified: Thu Feb 22 01:36:54 UTC 2024
    - 4.7K bytes
    - Viewed (0)
  8. src/test/java/org/codelibs/opensearch/extension/analysis/SynonymLoader.java

                @Override
                protected TokenStreamComponents createComponents(final String fieldName) {
                    final Tokenizer tokenizer = new KeywordTokenizer();
                    final TokenStream stream = ignoreCase ? new LowerCaseFilter(tokenizer) : tokenizer;
                    return new TokenStreamComponents(tokenizer, stream);
                }
            };
        }
    Registered: Fri Nov 08 09:08:12 UTC 2024
    - Last Modified: Thu Feb 22 01:36:54 UTC 2024
    - 6.7K bytes
    - Viewed (0)
  9. src/test/java/org/codelibs/opensearch/extension/analysis/NGramSynonymTokenizerFactory.java

     * either express or implied. See the License for the specific language
     * governing permissions and limitations under the License.
     */
    package org.codelibs.opensearch.extension.analysis;
    
    import org.apache.lucene.analysis.Tokenizer;
    import org.opensearch.common.settings.Settings;
    import org.opensearch.env.Environment;
    import org.opensearch.index.IndexSettings;
    import org.opensearch.index.analysis.AbstractTokenizerFactory;
    
    /**
    Registered: Fri Nov 08 09:08:12 UTC 2024
    - Last Modified: Thu Feb 22 01:36:54 UTC 2024
    - 2.4K bytes
    - Viewed (0)
  10. src/cmd/asm/internal/asm/line_test.go

    	})
    }
    
    func testBadInstParser(t *testing.T, goarch string, tests []badInstTest) {
    	for i, test := range tests {
    		arch, ctxt := setArch(goarch)
    		tokenizer := lex.NewTokenizer("", strings.NewReader(test.input+"\n"), nil)
    		parser := NewParser(ctxt, arch, tokenizer)
    
    		err := tryParse(t, func() {
    			parser.Parse()
    		})
    
    		switch {
    		case err == nil:
    Registered: Tue Nov 05 11:13:11 UTC 2024
    - Last Modified: Tue Aug 29 07:48:38 UTC 2023
    - 1.9K bytes
    - Viewed (0)
Back to top