Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 4 of 4 for NewTokenizer (0.25 sec)

  1. src/cmd/asm/internal/lex/lex_test.go

    			"A",
    			"B(THIS)",
    		),
    		"x.\n",
    	},
    	*/
    }
    
    func TestLex(t *testing.T) {
    	for _, test := range lexTests {
    		input := NewInput(test.name)
    		input.Push(NewTokenizer(test.name, strings.NewReader(test.input), nil))
    		result := drain(input)
    		if result != test.output {
    			t.Errorf("%s: got %q expected %q", test.name, result, test.output)
    		}
    	}
    }
    
    Registered: Tue Nov 05 11:13:11 UTC 2024
    - Last Modified: Tue Aug 29 07:48:38 UTC 2023
    - 5.8K bytes
    - Viewed (0)
  2. src/cmd/asm/internal/lex/lex.go

    func NewLexer(name string) TokenReader {
    	input := NewInput(name)
    	fd, err := os.Open(name)
    	if err != nil {
    		log.Fatalf("%s\n", err)
    	}
    	input.Push(NewTokenizer(name, fd, fd))
    	return input
    }
    
    // The other files in this directory each contain an implementation of TokenReader.
    
    // A TokenReader is like a reader, but returns lex tokens of type Token. It also can tell you what
    Registered: Tue Nov 05 11:13:11 UTC 2024
    - Last Modified: Tue Aug 29 18:31:05 UTC 2023
    - 4.1K bytes
    - Viewed (0)
  3. src/cmd/asm/internal/lex/tokenizer.go

    // turning text from readers into tokens.
    type Tokenizer struct {
    	tok  ScanToken
    	s    *scanner.Scanner
    	base *src.PosBase
    	line int
    	file *os.File // If non-nil, file descriptor to close.
    }
    
    func NewTokenizer(name string, r io.Reader, file *os.File) *Tokenizer {
    	var s scanner.Scanner
    	s.Init(r)
    	// Newline is like a semicolon; other space characters are fine.
    	s.Whitespace = 1<<'\t' | 1<<'\r' | 1<<' '
    Registered: Tue Nov 05 11:13:11 UTC 2024
    - Last Modified: Thu Aug 04 20:35:21 UTC 2022
    - 3K bytes
    - Viewed (0)
  4. src/cmd/asm/internal/asm/line_test.go

    		{"BSWAPW R11", `unrecognized instruction`},
    	})
    }
    
    func testBadInstParser(t *testing.T, goarch string, tests []badInstTest) {
    	for i, test := range tests {
    		arch, ctxt := setArch(goarch)
    		tokenizer := lex.NewTokenizer("", strings.NewReader(test.input+"\n"), nil)
    		parser := NewParser(ctxt, arch, tokenizer)
    
    		err := tryParse(t, func() {
    			parser.Parse()
    		})
    
    		switch {
    		case err == nil:
    Registered: Tue Nov 05 11:13:11 UTC 2024
    - Last Modified: Tue Aug 29 07:48:38 UTC 2023
    - 1.9K bytes
    - Viewed (0)
Back to top