- Sort Score
- Result 10 results
- Languages All
Results 91 - 100 of 1,032 for Allocate (0.1 sec)
-
src/runtime/rt0_linux_ppc64le.s
XOR R0, R0 // Make sure R0 is zero before _main BR _main<>(SB) TEXT _rt0_ppc64le_linux_lib(SB),NOSPLIT|NOFRAME,$0 // This is called with ELFv2 calling conventions. Convert to Go. // Allocate argument storage for call to newosproc0. STACK_AND_SAVE_HOST_TO_GO_ABI(16) MOVD R3, _rt0_ppc64le_linux_lib_argc<>(SB) MOVD R4, _rt0_ppc64le_linux_lib_argv<>(SB) // Synchronous initialization.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 18:17:17 UTC 2024 - 2.9K bytes - Viewed (0) -
tensorflow/c/kernels_test.cc
} REGISTER_OP("AllocateOutputOp1").Output("output1: float"); TEST_F(DeviceKernelOpTest, TestAllocateOutputSizeOne) { auto my_compute_func = [](void* kernel, TF_OpKernelContext* ctx) { // Allocate output TF_Status* s = TF_NewStatus(); int64_t dim = 1; size_t tensor_size_bytes = TF_DataTypeSize(TF_FLOAT); TF_Tensor* output = TF_AllocateOutput(
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Sep 06 19:12:29 UTC 2023 - 50.4K bytes - Viewed (0) -
cmd/kubeadm/app/phases/controlplane/manifests_test.go
"--client-ca-file=" + filepath.Join(testCertsDir, "ca.crt"), "--requestheader-client-ca-file=" + filepath.Join(testCertsDir, "front-proxy-ca.crt"), "--allocate-node-cidrs=true", "--cluster-cidr=10.0.1.15/16", }, }, { name: "custom service-cluster-ip-range for " + cpVersion, cfg: &kubeadmapi.ClusterConfiguration{ Networking: kubeadmapi.Networking{
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Sun Mar 03 14:43:47 UTC 2024 - 48.5K bytes - Viewed (0) -
internal/s3select/csv/reader.go
if r.err != nil { return nil, r.err } // Move to next block item, ok := <-r.queue if !ok { r.err = io.EOF return nil, r.err } //nolint:staticcheck // SA6002 Using pointer would allocate more since we would have to copy slice header before taking a pointer. r.csvDstPool.Put(r.current) r.current = <-item.dst r.err = item.err r.recordsRead = 0 } csvRecord := r.current[r.recordsRead]
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Thu Feb 22 06:26:06 UTC 2024 - 8.9K bytes - Viewed (0) -
pkg/kubelet/cm/memorymanager/policy_static.go
return err } return nil } // Allocate call is idempotent func (p *staticPolicy) Allocate(s state.State, pod *v1.Pod, container *v1.Container) (rerr error) { // allocate the memory only for guaranteed pods if v1qos.GetPodQOS(pod) != v1.PodQOSGuaranteed { return nil } podUID := string(pod.UID) klog.InfoS("Allocate", "pod", klog.KObj(pod), "containerName", container.Name)
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Sun Nov 12 07:34:55 UTC 2023 - 34K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_tensor.cc
uint64 size = client->backend().transfer_manager()->GetByteSizeRequirement(subshape); TF_ASSIGN_OR_RETURN(se::OwningDeviceMemory buffer, client->backend().memory_allocator()->Allocate( device_ordinal, size, /*retry_on_failure=*/false, subshape.layout().memory_space())); // Move our buffer into shaped_buffer, which takes ownership of it.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 22 08:47:20 UTC 2024 - 4.5K bytes - Viewed (0) -
src/runtime/testdata/testprogcgo/stackswitch.c
stackSwitchCallback(); // Next, verify that switching stacks doesn't break callbacks. char *stack1 = malloc(STACK_SIZE); if (stack1 == NULL) { perror("malloc"); exit(1); } // Allocate the second stack before freeing the first to ensure we don't get // the same address from malloc. // // Will be freed in stackSwitchThread2. stack2 = malloc(STACK_SIZE); if (stack1 == NULL) { perror("malloc");
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Oct 26 15:17:33 UTC 2023 - 3.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/common/quantization_lib/quantization_driver.h
// result without creating new entry in the state vector. Otherwise, allocate // a new entry in the state vector. void InitializeArgState(BlockArgument arg, Value arg_value); // Sets the state of the index-th operand of the op. If this operand is // cached, uses the cached result without creating new entry in the state // vector. Otherwise, allocate a new entry in the state vector.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Mar 20 11:42:17 UTC 2024 - 16.8K bytes - Viewed (0) -
pilot/pkg/serviceregistry/serviceentry/controller.go
// IPs allocated to a service entry may differ from istiod to istiod // but it does not matter because these IPs only affect the listener // IPs on a given proxy managed by a given istiod. // // NOTE: If DNS capture is not enabled by the proxy, the automatically // allocated IP addresses do not take effect. // // The current algorithm to allocate IPs is deterministic across all istiods.
Registered: Fri Jun 14 15:00:06 UTC 2024 - Last Modified: Wed May 29 15:31:09 UTC 2024 - 36.8K bytes - Viewed (0) -
pkg/scheduler/framework/plugins/dynamicresources/structuredparameters_test.go
driverName, allocation, err := controller.allocate(tCtx, nodeName, tc.resources) if err != nil { if expect.allocateErr == "" { tCtx.Fatalf("unexpected allocate error: %v", err) } require.Equal(tCtx, expect.allocateErr, err.Error()) return } if expect.allocateErr != "" { tCtx.Fatalf("did not get expected allocate error: %v", expect.allocateErr) }
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed Mar 27 09:27:01 UTC 2024 - 34.8K bytes - Viewed (0)