Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 8 of 8 for Tokenize (0.05 sec)

  1. src/cmd/asm/internal/lex/input.go

    		if i > 0 {
    			name, value = name[:i], name[i+1:]
    		}
    		tokens := Tokenize(name)
    		if len(tokens) != 1 || tokens[0].ScanToken != scanner.Ident {
    			fmt.Fprintf(os.Stderr, "asm: parsing -D: %q is not a valid identifier name\n", tokens[0])
    			flags.Usage()
    		}
    		macros[name] = &Macro{
    			name:   name,
    			args:   nil,
    			tokens: Tokenize(value),
    		}
    	}
    	return macros
    }
    
    var panicOnError bool // For testing.
    Registered: Tue Sep 09 11:13:09 UTC 2025
    - Last Modified: Fri Sep 06 13:17:27 UTC 2024
    - 12.5K bytes
    - Viewed (0)
  2. src/cmd/asm/internal/asm/operand_test.go

    		}
    	}()
    
    	parse()
    
    	return nil
    }
    
    func testBadOperandParser(t *testing.T, parser *Parser, tests []badOperandTest) {
    	for _, test := range tests {
    		err := tryParse(t, func() {
    			parser.start(lex.Tokenize(test.input))
    			addr := obj.Addr{}
    			parser.operand(&addr)
    		})
    
    		switch {
    		case err == nil:
    			t.Errorf("fail at %s: got no errors; expected %s\n", test.input, test.error)
    Registered: Tue Sep 09 11:13:09 UTC 2025
    - Last Modified: Tue Aug 29 18:31:05 UTC 2023
    - 23.9K bytes
    - Viewed (0)
  3. src/bufio/scan.go

    	scanCalled   bool      // Scan has been called; buffer is in use.
    	done         bool      // Scan has finished.
    }
    
    // SplitFunc is the signature of the split function used to tokenize the
    // input. The arguments are an initial substring of the remaining unprocessed
    // data and a flag, atEOF, that reports whether the [Reader] has no more data
    Registered: Tue Sep 09 11:13:09 UTC 2025
    - Last Modified: Wed May 21 18:05:26 UTC 2025
    - 14.2K bytes
    - Viewed (0)
  4. src/main/resources/fess_indices/fess.json

    Registered: Thu Sep 04 12:52:25 UTC 2025
    - Last Modified: Fri Jun 06 14:17:42 UTC 2025
    - 39.8K bytes
    - Viewed (0)
  5. src/test/java/jcifs/internal/smb2/session/Smb2SessionSetupRequestTest.java

            // Verify token length
            assertEquals(tokenSize, SMBUtil.readInt2(buffer, bodyOffset + 14));
    
            if (tokenSize > 0) {
                // Verify token content
                int tokenOffset = bodyOffset + 24;
                while ((tokenOffset % 8) != 0) {
                    tokenOffset++; // Account for pad8
                }
                byte[] actualToken = new byte[tokenSize];
    Registered: Sun Sep 07 00:10:21 UTC 2025
    - Last Modified: Thu Aug 14 05:31:44 UTC 2025
    - 21.2K bytes
    - Viewed (0)
  6. android/guava/src/com/google/common/net/MediaType.java

        checkNotNull(input);
        Tokenizer tokenizer = new Tokenizer(input);
        try {
          String type = tokenizer.consumeToken(TOKEN_MATCHER);
          consumeSeparator(tokenizer, '/');
          String subtype = tokenizer.consumeToken(TOKEN_MATCHER);
          ImmutableListMultimap.Builder<String, String> parameters = ImmutableListMultimap.builder();
          while (tokenizer.hasMore()) {
            consumeSeparator(tokenizer, ';');
    Registered: Fri Sep 05 12:43:10 UTC 2025
    - Last Modified: Tue May 13 17:27:14 UTC 2025
    - 48K bytes
    - Viewed (0)
  7. guava/src/com/google/common/net/MediaType.java

        checkNotNull(input);
        Tokenizer tokenizer = new Tokenizer(input);
        try {
          String type = tokenizer.consumeToken(TOKEN_MATCHER);
          consumeSeparator(tokenizer, '/');
          String subtype = tokenizer.consumeToken(TOKEN_MATCHER);
          ImmutableListMultimap.Builder<String, String> parameters = ImmutableListMultimap.builder();
          while (tokenizer.hasMore()) {
            consumeSeparator(tokenizer, ';');
    Registered: Fri Sep 05 12:43:10 UTC 2025
    - Last Modified: Tue May 13 17:27:14 UTC 2025
    - 48K bytes
    - Viewed (0)
  8. README.md

    String escaped = JsonUtil.escape("Hello \"World\" with special chars");
    String unescaped = JsonUtil.unescape(escaped);
    
    // Text tokenization
    Tokenizer tokenizer = new Tokenizer("field1,field2,field3", ",");
    while (tokenizer.hasMoreTokens()) {
        String token = tokenizer.nextToken();
        // Process each token
    }
    
    // Decimal formatting
    DecimalFormat format = DecimalFormatUtil.getDecimalFormat("###,###.00");
    ```
    
    Registered: Fri Sep 05 20:58:11 UTC 2025
    - Last Modified: Sun Aug 31 02:56:02 UTC 2025
    - 12.7K bytes
    - Viewed (0)
Back to top