- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 26 for Allocate (0.14 sec)
-
tensorflow/c/experimental/stream_executor/stream_executor.cc
platform_fns_->destroy_device(platform_, &device_); } absl::Status Init() override { return absl::OkStatus(); } DeviceMemoryBase Allocate(uint64 size, int64_t memory_space) override { SP_DeviceMemoryBase mem = {SP_DEVICE_MEMORY_BASE_STRUCT_SIZE}; stream_executor_->allocate(&device_, size, memory_space, &mem); absl::Status status = ValidateSPDeviceMemoryBase(mem); if (!status.ok()) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Jun 14 07:39:19 UTC 2024 - 27.1K bytes - Viewed (0) -
tests/integration/ambient/main_test.go
apps.Mesh = inMesh.GetMatches(echos) apps.MeshExternal = match.Not(inMesh).GetMatches(echos) // TODO(https://github.com/istio/istio/issues/51083) remove manually allocate if err := cdeployment.DeployExternalServiceEntry(t.ConfigIstio(), apps.Namespace, apps.ExternalNamespace, true). Apply(apply.CleanupConditionally); err != nil { return err } if apps.WaypointProxies == nil {
Registered: Fri Jun 14 15:00:06 UTC 2024 - Last Modified: Wed Jun 12 00:07:28 UTC 2024 - 10.8K bytes - Viewed (0) -
cmd/erasure-multipart.go
if err != nil { return pi, toObjectErr(err, bucket, object) } // Fetch buffer for I/O, returns from the pool if not allocates a new one and returns. var buffer []byte switch size := data.Size(); { case size == 0: buffer = make([]byte, 1) // Allocate at least a byte to reach EOF case size == -1: if size := data.ActualSize(); size > 0 && size < fi.Erasure.BlockSize {
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Thu Jun 13 06:56:12 UTC 2024 - 44.8K bytes - Viewed (0) -
cmd/iam-object-store.go
func listIAMConfigItems(ctx context.Context, objAPI ObjectLayer, pathPrefix string) <-chan itemOrErr[string] { ch := make(chan itemOrErr[string]) go func() { defer xioutil.SafeClose(ch) // Allocate new results channel to receive ObjectInfo. objInfoCh := make(chan itemOrErr[ObjectInfo]) if err := objAPI.Walk(ctx, minioMetaBucket, pathPrefix, objInfoCh, WalkOptions{}); err != nil { select {
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Thu Jun 13 22:26:38 UTC 2024 - 19.5K bytes - Viewed (0) -
pkg/scheduler/framework/types.go
r.SetScalar(name, r.ScalarResources[name]+quantity) } // SetScalar sets a resource by a scalar value of this resource. func (r *Resource) SetScalar(name v1.ResourceName, quantity int64) { // Lazily allocate scalar resource map. if r.ScalarResources == nil { r.ScalarResources = map[v1.ResourceName]int64{} } r.ScalarResources[name] = quantity }
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed Jun 12 19:28:17 UTC 2024 - 36.7K bytes - Viewed (0) -
src/go/types/unify.go
func newUnifier(tparams []*TypeParam, targs []Type, enableInterfaceInference bool) *unifier { assert(len(tparams) >= len(targs)) handles := make(map[*TypeParam]*Type, len(tparams)) // Allocate all handles up-front: in a correct program, all type parameters // must be resolved and thus eventually will get a handle. // Also, sharing of handles caused by unified type parameters is rare and
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Jun 11 16:24:39 UTC 2024 - 27.9K bytes - Viewed (0) -
src/cmd/compile/internal/types2/unify.go
func newUnifier(tparams []*TypeParam, targs []Type, enableInterfaceInference bool) *unifier { assert(len(tparams) >= len(targs)) handles := make(map[*TypeParam]*Type, len(tparams)) // Allocate all handles up-front: in a correct program, all type parameters // must be resolved and thus eventually will get a handle. // Also, sharing of handles caused by unified type parameters is rare and
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Jun 11 16:24:39 UTC 2024 - 27.8K bytes - Viewed (0) -
src/cmd/compile/internal/liveness/plive.go
// parameters are live all the time. In addition, any locals // that are pointers to heap-allocated output parameters are // also always live (post-deferreturn code needs these // pointers to copy values back to the stack). // TODO: if the output parameter is heap-allocated, then we // don't need to keep the stack copy live? if lv.fn.HasDefer() { for i, n := range lv.vars {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 15:22:22 UTC 2024 - 45.2K bytes - Viewed (0) -
src/os/exec/exec.go
Dir: c.Dir, Files: childFiles, Env: env, Sys: c.SysProcAttr, }) if err != nil { return err } started = true // Don't allocate the goroutineErr channel unless there are goroutines to start. if len(c.goroutine) > 0 { goroutineErr := make(chan error, 1) c.goroutineErr = goroutineErr type goroutineStatus struct { running int
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 20:13:53 UTC 2024 - 41.4K bytes - Viewed (0) -
src/runtime/mfinal.go
// objects in memory. See https://go.dev/ref/spec#Size_and_alignment_guarantees. // // It is not guaranteed that a finalizer will run for objects allocated // in initializers for package-level variables. Such objects may be // linker-allocated, not heap-allocated. // // Note that because finalizers may execute arbitrarily far into the future // after an object is no longer referenced, the runtime is allowed to perform
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 01:56:56 UTC 2024 - 19K bytes - Viewed (0)