- Sort Score
- Result 10 results
- Languages All
Results 61 - 70 of 168 for Malloc (0.23 sec)
-
tensorflow/compiler/mlir/tensorflow/analysis/side_effect_analysis.cc
class SideEffects { enum Type { kAlloc = 0, kFree = 1, kRead = 2, kWrite = 3 }; public: bool IsAlloc() const { return effects_.test(kAlloc); } bool IsFree() const { return effects_.test(kFree); } bool IsRead() const { return effects_.test(kRead); } bool IsWrite() const { return effects_.test(kWrite); } bool IsAllocOnly() const { return IsAlloc() && effects_.count() == 1; }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 15 09:04:13 UTC 2024 - 41.2K bytes - Viewed (0) -
src/runtime/mpagecache_test.go
test := test t.Run(name, func(t *testing.T) { c := test.cache for i, h := range test.hits { b, s := c.Alloc(h.npages) if b != h.base { t.Fatalf("bad alloc base #%d: got 0x%x, want 0x%x", i, b, h.base) } if s != h.scav { t.Fatalf("bad alloc scav #%d: got %d, want %d", i, s, h.scav) } } }) } } func TestPageCacheFlush(t *testing.T) {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Dec 06 19:16:48 UTC 2021 - 10.8K bytes - Viewed (0) -
pkg/registry/core/service/portallocator/controller/repair.go
return &Repair{ interval: interval, serviceClient: serviceClient, portRange: portRange, alloc: alloc, leaks: map[int]int{}, broadcaster: eventBroadcaster, recorder: recorder, } } // RunUntil starts the controller until the provided ch is closed. func (c *Repair) RunUntil(onFirstSuccess func(), stopCh chan struct{}) { c.broadcaster.StartRecordingToSink(stopCh)
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Sat Sep 30 15:46:06 UTC 2023 - 10.4K bytes - Viewed (0) -
src/internal/trace/event/go122/event.go
) // Experiments. const ( // AllocFree is the alloc-free events experiment. AllocFree event.Experiment = 1 + iota ) // Experimental events. const ( _ event.Type = 127 + iota // Experimental events for AllocFree. // Experimental heap span events. Added in Go 1.23. EvSpan // heap span exists [timestamp, id, npages, type/class] EvSpanAlloc // heap span alloc [timestamp, id, npages, type/class]
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 17 18:48:18 UTC 2024 - 14.8K bytes - Viewed (0) -
src/runtime/export_test.go
if s.isUnusedUserArenaChunk() { continue } if sizeclass := s.spanclass.sizeclass(); sizeclass == 0 { slow.Mallocs++ slow.Alloc += uint64(s.elemsize) } else { slow.Mallocs += uint64(s.allocCount) slow.Alloc += uint64(s.allocCount) * uint64(s.elemsize) bySize[sizeclass].Mallocs += uint64(s.allocCount) } } // Add in frees by just reading the stats for those directly.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 30 17:50:53 UTC 2024 - 46.1K bytes - Viewed (0) -
src/runtime/mgcpacer_test.go
t.Parallel() const initialHeapBytes = 256 << 10 for _, e := range []*gcExecTest{ { // The most basic test case: a steady-state heap. // Growth to an O(MiB) heap, then constant heap size, alloc/scan rates. name: "Steady", gcPercent: 100, memoryLimit: math.MaxInt64, globalsBytes: 32 << 10, nCores: 8, allocRate: constant(33.0), scanRate: constant(1024.0),
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 19 13:53:21 UTC 2023 - 39.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfr/passes/raise_to_tf.cc
// the new TF op is an unranked tensor with element type derived. class RewriteTFRCallOp : public OpRewritePattern<CallOp> { using OpRewritePattern<CallOp>::OpRewritePattern; public: explicit RewriteTFRCallOp(MLIRContext* context, const SymbolTable& table, bool materialize_derived_attrs) : OpRewritePattern<CallOp>(context), symbol_table_(table), materialize_derived_attrs_(materialize_derived_attrs) {}
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 21.8K bytes - Viewed (0) -
pkg/controller/nodeipam/ipam/cidrset/cidr_set_test.go
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Thu May 11 08:53:03 UTC 2023 - 29.5K bytes - Viewed (0) -
pkg/registry/core/service/allocator/bitmap_test.go
// Allocate all the values in both the dynamic and reserved blocks for i := 0; i < max; i++ { alloc, ok, _ := m.AllocateNext() if !ok { t.Fatalf("unexpected error") } if _, ok := uniqueAllocated[alloc]; ok { t.Fatalf("unexpected allocated value %d", alloc) } else { uniqueAllocated[alloc] = true } } if max != len(uniqueAllocated) {
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Tue May 10 08:56:31 UTC 2022 - 13.3K bytes - Viewed (0) -
pkg/registry/core/service/ipallocator/controller/repair.go
} } return nil } func (c *Repair) saveSnapShot(rebuilt *ipallocator.Range, alloc rangeallocation.RangeRegistry, snapshot *api.RangeAllocation) error { if err := rebuilt.Snapshot(snapshot); err != nil { return fmt.Errorf("unable to snapshot the updated service IP allocations: %v", err) } if err := alloc.CreateOrUpdate(snapshot); err != nil { if errors.IsConflict(err) { return err }
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Sat Sep 30 15:46:06 UTC 2023 - 13.8K bytes - Viewed (0)