- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 70 for zeroed (0.11 sec)
-
src/runtime/mheap.go
arenaBase := base % heapArenaBytes if arenaBase < zeroedBase { // We extended into the non-zeroed part of the // arena, so this region needs to be zeroed before use. // // zeroedBase is monotonically increasing, so if we see this now then // we can be sure we need to zero this memory region. // // We still need to update zeroedBase for this arena, and // potentially more arenas.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 22:31:00 UTC 2024 - 78K bytes - Viewed (0) -
src/runtime/malloc.go
// already zeroed. Otherwise if needzero is true, objects are zeroed as // they are allocated. There are various benefits to delaying zeroing // this way: // // 1. Stack frame allocation can avoid zeroing altogether. // // 2. It exhibits better temporal locality, since the program is // probably about to write to the memory. // // 3. We don't zero pages that never get reused.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 59.6K bytes - Viewed (0) -
src/runtime/mbitmap.go
// does not execute write barriers for [dst, dst+size). // // In addition to the requirements of bulkBarrierPreWrite // callers need to ensure [dst, dst+size) is zeroed. // // This is used for special cases where e.g. dst was just // created and zeroed with malloc. // // The type of the space can be provided purely as an optimization. // See bulkBarrierPreWrite's comment for more details -- use this
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 00:18:55 UTC 2024 - 60K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/_gen/ARM64Ops.go
// insert low width bits of arg0 into the result starting at bit lsb, bits to the left and right of the inserted bit field are zeroed {name: "UBFIZ", argLength: 1, reg: gp11, asm: "UBFIZ", aux: "ARM64BitField"}, // extract width bits of arg0 starting at bit lsb and insert at low end of result, remaining high bits are zeroed {name: "UBFX", argLength: 1, reg: gp11, asm: "UBFX", aux: "ARM64BitField"}, // moves
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 15:49:20 UTC 2024 - 58.8K bytes - Viewed (0) -
src/cmd/compile/internal/ppc64/ssa.go
p.From.Type = obj.TYPE_REG p.From.Reg = ppc64.REG_VS32 p.To.Type = obj.TYPE_MEM p.To.Reg = v.Args[0].Reg() p.To.Offset = 48 // Increment address for the // 64 bytes just zeroed. p = s.Prog(ppc64.AADD) p.Reg = v.Args[0].Reg() p.From.Type = obj.TYPE_CONST p.From.Offset = 64 p.To.Type = obj.TYPE_REG p.To.Reg = v.Args[0].Reg()
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 19:59:38 UTC 2024 - 55.4K bytes - Viewed (0) -
src/math/big/int_test.go
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 18:42:28 UTC 2024 - 58.5K bytes - Viewed (0) -
src/cmd/cgo/out.go
// The results part of the argument structure must be // initialized to 0 so the write barriers generated by // the assignments to these fields in Go are safe. // // We use a local static variable to get the zeroed // value of the argument type. This avoids including // string.h for memset, and is also robust to C++ // types with constructors. Both GCC and LLVM optimize // this into just zeroing _cgo_a.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Mar 29 16:41:10 UTC 2024 - 59.6K bytes - Viewed (0) -
src/runtime/map.go
} func mapaccess1_fat(t *maptype, h *hmap, key, zero unsafe.Pointer) unsafe.Pointer { e := mapaccess1(t, h, key) if e == unsafe.Pointer(&zeroVal[0]) { return zero } return e } func mapaccess2_fat(t *maptype, h *hmap, key, zero unsafe.Pointer) (unsafe.Pointer, bool) { e := mapaccess1(t, h, key) if e == unsafe.Pointer(&zeroVal[0]) { return zero, false } return e, true }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 57.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/merge_control_flow.mlir
tf_device.return }) {cluster_attr = "cluster_attr"} : () -> () func.return } // Check that 3 IfRegions where 1st and 3rd IfRegions // can be merged but not 2nd IfRegion and 2nd IfRegion should be moved after // newly merged IfRegion. // CHECK-LABEL: func @same_predicate_3_ifregions_reorder func.func @same_predicate_3_ifregions_reorder() { // CHECK: tf_device.cluster
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Oct 30 06:52:55 UTC 2023 - 63.7K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/rewrite.go
} // Test if RLWINM feeding into SRDconst can be merged. Return the encoded RLIWNM constant, // or 0 if they cannot be merged. func mergePPC64SldiRlwinm(sldi, rlw int64) int64 { r_1, mb, me, mask_1 := DecodePPC64RotateMask(rlw) if mb > me || mb < sldi { // Wrapping masks cannot be merged as the upper 32 bits are effectively undefined in this case.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 19:02:52 UTC 2024 - 64.2K bytes - Viewed (0)