- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 41 for 32kb (0.05 sec)
-
src/cmd/vendor/github.com/google/pprof/internal/driver/driver_focus.go
// contained on the range described by a string. It can recognize // strings of the form: // "32kb" -- matches values == 32kb // ":64kb" -- matches values <= 64kb // "4mb:" -- matches values >= 4mb // "12kb:64mb" -- matches values between 12kb and 64mb (both included). func parseTagFilterRange(filter string) func(int64, string) bool { ranges := tagFilterRangeRx.FindAllStringSubmatch(filter, 2)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Oct 22 23:33:06 UTC 2020 - 6.4K bytes - Viewed (0) -
src/runtime/testdata/testprogcgo/stackswitch.c
#if defined(__linux__) && !defined(__GLIBC__) #define MUSL 1 #endif #if defined(MUSL) void callStackSwitchCallbackFromThread(void) { printf("SKIP\n"); exit(0); } #else // Use a stack size larger than the 32kb estimate in // runtime.callbackUpdateSystemStack. This ensures that a second stack // allocation won't accidentally count as in bounds of the first stack #define STACK_SIZE (64ull << 10)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Oct 26 15:17:33 UTC 2023 - 3.9K bytes - Viewed (0) -
src/cmd/internal/buildid/note.go
} // It should be in the first few bytes, but read a lot just in case, // especially given our past problems on OS X with the build ID moving. // There shouldn't be much difference between reading 4kB and 32kB: // the hard part is getting to the data, not transferring it. n := sect.Size if n > uint64(readSize) { n = uint64(readSize) } buf := make([]byte, n)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Aug 17 20:40:42 UTC 2023 - 6K bytes - Viewed (0) -
src/net/http/request.go
func (l *maxBytesReader) Read(p []byte) (n int, err error) { if l.err != nil { return 0, l.err } if len(p) == 0 { return 0, nil } // If they asked for a 32KB byte read but only 5 bytes are // remaining, no need to read 32KB. 6 bytes will answer the // question of the whether we hit the limit or go past it. // 0 < len(p) < 2^63 if int64(len(p))-1 > l.n { p = p[:l.n+1] }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 49.4K bytes - Viewed (0) -
src/cmd/go/internal/work/exec.go
cleanup = func() {} // no cleanup by default var argLen int for _, arg := range cmd.Args { argLen += len(arg) } // If we're not approaching 32KB of args, just pass args normally. // (use 30KB instead to be conservative; not sure how accounting is done) if !useResponseFile(cmd.Path, argLen) { return } tf, err := os.CreateTemp("", "args") if err != nil {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Jun 10 14:46:37 UTC 2024 - 105.6K bytes - Viewed (0) -
src/runtime/memmove_ppc64x.s
MOVBZ TMP, -1(TGT) SUB $1,TGT BDNZ backwardtailloop nobackwardtail: BC 4, 5, LR // blelr cr1, return if DWORDS == 0 SRDCC $2,DWORDS,QWORDS // Compute number of 32B blocks and compare to 0 BNE backward32setup // If QWORDS != 0, start the 32B copy loop. backward24: // DWORDS is a value between 1-3. CMP DWORDS, $2 MOVD -8(SRC), TMP MOVD TMP, -8(TGT) BC 12, 0, LR // bltlr, return if DWORDS == 1
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Apr 21 16:47:45 UTC 2023 - 4.9K bytes - Viewed (0) -
cmd/handler-api.go
if legacy { // ram_per_request is (1MiB+32KiB) * driveCount \ // + 2 * 10MiB (default erasure block size v1) + 2 * 1MiB (default erasure block size v2) apiRequestsMaxPerNode = int(maxMem / uint64(maxSetDrives*blockSize+int(blockSizeV1*2+blockSizeV2*2))) } else { // ram_per_request is (1MiB+32KiB) * driveCount \ // + 2 * 1MiB (default erasure block size v2)
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Wed Jun 12 08:13:12 UTC 2024 - 10.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/end2end/fake_quant_without_identity.pbtxt
211E\276\344\035O\276\242\220l\276w\307\020\2773`\376\276\331\377\202\276\374\016Y\276\224\273^\2767`~\276V\016>\276\226\333g\276*\241K\276\224g\177\276\320\315(\276t\376\177\276B\025B\276@\214\226\276\377s\376\276\373\365\377\277\343\211~\276\205\320b\276\213\236\377\2763\0136\276u\237:\276\325\373~\276\344\r\037\276>\035(\276\247\364!\276\2613%\276\207]!\276\320S\376\301\354 _\276\337PT\276\342\226[\276\304\307\177\276\360\"x\276\312\023\200\276Wfp\2768$W\276\235\014\202\276\230\210m\2761\342L...
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 13.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/end2end/fake_quant_without_identity_4bit.pbtxt
211E\276\344\035O\276\242\220l\276w\307\020\2773`\376\276\331\377\202\276\374\016Y\276\224\273^\2767`~\276V\016>\276\226\333g\276*\241K\276\224g\177\276\320\315(\276t\376\177\276B\025B\276@\214\226\276\377s\376\276\373\365\377\277\343\211~\276\205\320b\276\213\236\377\2763\0136\276u\237:\276\325\373~\276\344\r\037\276>\035(\276\247\364!\276\2613%\276\207]!\276\320S\376\301\354 _\276\337PT\276\342\226[\276\304\307\177\276\360\"x\276\312\023\200\276Wfp\2768$W\276\235\014\202\276\230\210m\2761\342L...
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 13.8K bytes - Viewed (0) -
src/runtime/memclr_amd64.s
JNE skip_erms // If the size is less than 2kb, do not use ERMS as it has a big start-up cost. // Table 3-4. Relative Performance of Memcpy() Using ERMSB Vs. 128-bit AVX // in the Intel Optimization Guide shows better performance for ERMSB starting // from 2KB. Benchmarks show the similar threshold for REP STOS vs AVX. CMPQ BX, $2048 JAE loop_preheader_erms skip_erms:
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue May 10 20:52:34 UTC 2022 - 4.9K bytes - Viewed (0)