- Sort Score
- Result 10 results
- Languages All
Results 1 - 7 of 7 for newline (0.17 sec)
-
src/cmd/cgo/godefs.go
// semicolon at each newline, so we can replace newline with semicolon. // However, we can't do that in cases where the lexer would not insert // a semicolon. We only have to worry about cases that can occur in an // expression passed through gofmt, which means composite literals and // (due to the printer possibly inserting newlines because of position // information) operators.
Go - Registered: Tue Apr 30 11:13:12 GMT 2024 - Last Modified: Thu Sep 08 14:33:35 GMT 2022 - 4.5K bytes - Viewed (0) -
src/cmd/asm/internal/lex/tokenizer.go
} func NewTokenizer(name string, r io.Reader, file *os.File) *Tokenizer { var s scanner.Scanner s.Init(r) // Newline is like a semicolon; other space characters are fine. s.Whitespace = 1<<'\t' | 1<<'\r' | 1<<' ' // Don't skip comments: we need to count newlines. s.Mode = scanner.ScanChars | scanner.ScanFloats | scanner.ScanIdents | scanner.ScanInts | scanner.ScanStrings |
Go - Registered: Tue Apr 30 11:13:12 GMT 2024 - Last Modified: Thu Aug 04 20:35:21 GMT 2022 - 3K bytes - Viewed (0) -
src/archive/tar/strconv.go
return "", "", s, ErrHeader } n -= int64(len(nStr) + 1) // convert from index in s to index in rest if n <= 0 { return "", "", s, ErrHeader } // Extract everything between the space and the final newline. rec, nl, rem := rest[:n-1], rest[n-1:n], rest[n:] if nl != "\n" { return "", "", s, ErrHeader } // The first equals separates the key from the value. k, v, ok = strings.Cut(rec, "=") if !ok {
Go - Registered: Tue Apr 30 11:13:12 GMT 2024 - Last Modified: Tue Aug 01 14:28:42 GMT 2023 - 9K bytes - Viewed (0) -
src/cmd/asm/internal/lex/lex.go
// A TokenReader is like a reader, but returns lex tokens of type Token. It also can tell you what // the text of the most recently returned token is, and where it was found. // The underlying scanner elides all spaces except newline, so the input looks like a stream of // Tokens; original spacing is lost but we don't need it. type TokenReader interface { // Next returns the next token. Next() ScanToken
Go - Registered: Tue Apr 30 11:13:12 GMT 2024 - Last Modified: Tue Aug 29 18:31:05 GMT 2023 - 4.1K bytes - Viewed (0) -
src/cmd/asm/internal/lex/lex_test.go
}, { "#define A() A()\nA()", "recursive macro invocation", }, { "#define A a\n#define A a\n", "redefinition of macro", }, { "#define A a", "no newline after macro definition", }, } func TestBadLex(t *testing.T) { for _, test := range badLexTests { input := NewInput(test.error) input.Push(NewTokenizer(test.error, strings.NewReader(test.input), nil))
Go - Registered: Tue Apr 30 11:13:12 GMT 2024 - Last Modified: Tue Aug 29 07:48:38 GMT 2023 - 5.8K bytes - Viewed (0) -
misc/cgo/gmp/fib.go
// Keep the fibbers in dedicated operating system // threads, so that this program tests coordination // between pthreads and not just goroutines. runtime.LockOSThread() i := big.NewInt(n) if n == 0 { c <- i } for { j := <-c out <- j.String() i.Add(i, j) c <- i } } func main() { c := make(chan *big.Int) out := make(chan string) go fibber(c, out, 0)
Go - Registered: Tue Apr 30 11:13:12 GMT 2024 - Last Modified: Mon Apr 10 22:32:35 GMT 2023 - 919 bytes - Viewed (0) -
src/buildall.bash
linux_targets() { selectedtargets | grep 'linux' | sort } non_linux_targets() { selectedtargets | grep -v 'linux' | sort } # Note words in $targets are separated by both newlines and spaces. targets="$(linux_targets) $(non_linux_targets)" failed=false for target in $targets do echo "" echo "### Building $target" export GOOS=$(echo $target | sed 's/-.*//')
Shell Script - Registered: Tue Apr 30 11:13:12 GMT 2024 - Last Modified: Tue Apr 23 17:45:23 GMT 2024 - 2.1K bytes - Viewed (0)