- Sort Score
- Result 10 results
- Languages All
Results 1 - 3 of 3 for Bush (0.12 sec)
-
src/cmd/asm/internal/lex/lex_test.go
"#define A #define B (x) x", "A", "B(THIS)", ), "x.\n", }, */ } func TestLex(t *testing.T) { for _, test := range lexTests { input := NewInput(test.name) input.Push(NewTokenizer(test.name, strings.NewReader(test.input), nil)) result := drain(input) if result != test.output { t.Errorf("%s: got %q expected %q", test.name, result, test.output) } } }
Go - Registered: Tue Apr 30 11:13:12 GMT 2024 - Last Modified: Tue Aug 29 07:48:38 GMT 2023 - 5.8K bytes - Viewed (0) -
src/cmd/asm/internal/lex/lex.go
// NewLexer returns a lexer for the named file and the given link context. func NewLexer(name string) TokenReader { input := NewInput(name) fd, err := os.Open(name) if err != nil { log.Fatalf("%s\n", err) } input.Push(NewTokenizer(name, fd, fd)) return input } // The other files in this directory each contain an implementation of TokenReader.
Go - Registered: Tue Apr 30 11:13:12 GMT 2024 - Last Modified: Tue Aug 29 18:31:05 GMT 2023 - 4.1K bytes - Viewed (0) -
src/cmd/asm/internal/lex/stack.go
) // A Stack is a stack of TokenReaders. As the top TokenReader hits EOF, // it resumes reading the next one down. type Stack struct { tr []TokenReader } // Push adds tr to the top (end) of the input stack. (Popping happens automatically.) func (s *Stack) Push(tr TokenReader) { s.tr = append(s.tr, tr) } func (s *Stack) Next() ScanToken { tos := s.tr[len(s.tr)-1] tok := tos.Next()
Go - Registered: Tue Apr 30 11:13:12 GMT 2024 - Last Modified: Mon Jan 09 22:33:23 GMT 2017 - 1.2K bytes - Viewed (0)