- Sort Score
- Result 10 results
- Languages All
Results 111 - 120 of 269 for Atack (0.04 sec)
-
src/crypto/internal/nistec/p224.go
func (p *P224Point) ScalarMult(q *P224Point, scalar []byte) (*P224Point, error) { // Compute a p224Table for the base point q. The explicit NewP224Point // calls get inlined, letting the allocations live on the stack. var table = p224Table{NewP224Point(), NewP224Point(), NewP224Point(), NewP224Point(), NewP224Point(), NewP224Point(), NewP224Point(), NewP224Point(), NewP224Point(), NewP224Point(), NewP224Point(),
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Aug 12 00:04:29 UTC 2022 - 15.9K bytes - Viewed (0) -
src/crypto/internal/nistec/p521.go
func (p *P521Point) ScalarMult(q *P521Point, scalar []byte) (*P521Point, error) { // Compute a p521Table for the base point q. The explicit NewP521Point // calls get inlined, letting the allocations live on the stack. var table = p521Table{NewP521Point(), NewP521Point(), NewP521Point(), NewP521Point(), NewP521Point(), NewP521Point(), NewP521Point(), NewP521Point(), NewP521Point(), NewP521Point(), NewP521Point(),
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Aug 12 00:04:29 UTC 2022 - 17K bytes - Viewed (0) -
src/runtime/mbarrier.go
// the stack to the heap, but this requires first having a pointer // hidden on the stack. Immediately after a stack is scanned, it only // points to shaded objects, so it's not hiding anything, and the // shade(*slot) prevents it from hiding any other pointers on its // stack. // // For a detailed description of this barrier and proof of
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 15.7K bytes - Viewed (0) -
src/log/slog/logger_test.go
for pb.Next() { l.LogAttrs(ctx, LevelInfo, "msg", Int("a", 1), String("b", "two"), Bool("c", true)) } }) }) } // callerPC returns the program counter at the given stack depth. func callerPC(depth int) uintptr { var pcs [1]uintptr runtime.Callers(depth, pcs[:]) return pcs[0] } func wantAllocs(t *testing.T, want int, f func()) { if race.Enabled {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Nov 10 21:25:30 UTC 2023 - 19.5K bytes - Viewed (0) -
src/fmt/scan_test.go
testScanInts(t, func(r *RecursiveInt, b *bytes.Buffer) (err error) { _, err = Fscan(b, r) return }) } // 800 is small enough to not overflow the stack when using gccgo on a // platform that does not support split stack. const intCount = 800 func testScanInts(t *testing.T, scan func(*RecursiveInt, *bytes.Buffer) error) { r := new(RecursiveInt) ints := makeInts(intCount)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue May 23 20:25:13 UTC 2023 - 39.3K bytes - Viewed (0) -
src/cmd/vendor/golang.org/x/sys/unix/zsysnum_netbsd_arm64.go
SYS_GETSID = 286 // { pid_t|sys||getsid(pid_t pid); } SYS___CLONE = 287 // { pid_t|sys||__clone(int flags, void *stack); } SYS_FKTRACE = 288 // { int|sys||fktrace(int fd, int ops, int facs, pid_t pid); } SYS_PREADV = 289 // { ssize_t|sys||preadv(int fd, const struct iovec *iovp, int iovcnt, int PAD, off_t offset); }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Oct 19 23:33:33 UTC 2023 - 25.7K bytes - Viewed (0) -
src/crypto/internal/nistec/generate.go
func (p *{{.P}}Point) ScalarMult(q *{{.P}}Point, scalar []byte) (*{{.P}}Point, error) { // Compute a {{.p}}Table for the base point q. The explicit New{{.P}}Point // calls get inlined, letting the allocations live on the stack. var table = {{.p}}Table{New{{.P}}Point(), New{{.P}}Point(), New{{.P}}Point(), New{{.P}}Point(), New{{.P}}Point(), New{{.P}}Point(), New{{.P}}Point(), New{{.P}}Point(), New{{.P}}Point(), New{{.P}}Point(), New{{.P}}Point(),
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 04 17:29:44 UTC 2024 - 19.7K bytes - Viewed (0) -
src/cmd/go/internal/cache/cache.go
entry := fmt.Sprintf("v1 %x %x %20d %20d\n", id, out, size, time.Now().UnixNano()) if verify && allowVerify { old, err := c.get(id) if err == nil && (old.OutputID != out || old.Size != size) { // panic to show stack trace, so we can see what code is generating this cache entry. msg := fmt.Sprintf("go: internal cache error: cache verify failed: id=%x changed:<<<\n%s\n>>>\nold: %x %d\nnew: %x %d", id, reverseHash(id), out, size, old.OutputID, old.Size)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Sat Mar 09 14:19:39 UTC 2024 - 20.3K bytes - Viewed (0) -
src/encoding/pem/pem_test.go
} if string(rest) != test.input { t.Errorf("unexpected rest: %q; want = %q", rest, test.input) } } } func TestCVE202224675(t *testing.T) { // Prior to CVE-2022-24675, this input would cause a stack overflow. input := []byte(strings.Repeat("-----BEGIN \n", 10000000)) result, rest := Decode(input) if result != nil || !reflect.DeepEqual(rest, input) { t.Errorf("Encode of %#v decoded as %#v", input, rest) } }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Sep 29 22:56:00 UTC 2022 - 23.5K bytes - Viewed (0) -
src/runtime/crash_cgo_test.go
t.Skip("skipping windows specific test") } testenv.SkipFlaky(t, 22575) o := runTestProg(t, "testprogcgo", "StackMemory") stackUsage, err := strconv.Atoi(o) if err != nil { t.Fatalf("Failed to read stack usage: %v", err) } if expected, got := 100<<10, stackUsage; got > expected { t.Fatalf("expected < %d bytes of memory per thread, got %d", expected, got) } } func TestSigStackSwapping(t *testing.T) {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Apr 15 16:44:47 UTC 2024 - 22.2K bytes - Viewed (0)