- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 16 for cpu (0.11 sec)
-
tensorflow/c/eager/parallel_device/parallel_device_remote_test.cc
EXPECT_EQ(TF_OK, TF_GetCode(status.get())) << TF_Message(status.get()); BasicTestsForTwoDevices(context.get(), "/job:worker/replica:0/task:1/device:CPU:0", "/job:worker/replica:0/task:2/device:CPU:0"); worker_server1.release(); worker_server2.release(); } TEST(PARALLEL_DEVICE, TestAsyncCopyOff) {
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Apr 27 22:09:57 GMT 2023 - 6.7K bytes - Viewed (0) -
tensorflow/c/eager/parallel_device/parallel_device_lib_test.cc
ASSERT_TRUE(TF_GetCode(status.get()) == TF_OK) << TF_Message(status.get()); std::vector<std::string> devices{ "/job:localhost/replica:0/task:0/device:CPU:0", "/job:localhost/replica:0/task:0/device:CPU:1"}; ParallelDevice parallel_device(std::move(devices)); std::unique_ptr<TFE_Op, decltype(&TFE_DeleteOp)> handle_op( TFE_NewOp(context.get(), "VarHandleOp", status.get()), TFE_DeleteOp);
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Jul 08 23:47:35 GMT 2021 - 15.3K bytes - Viewed (0) -
tensorflow/c/eager/c_api_cluster_test.cc
EXPECT_EQ(TF_OK, TF_GetCode(status)) << TF_Message(status); const char remote_device_name[] = "/job:localhost/replica:0/task:1/device:CPU:0"; const char local_device_name[] = "/job:localhost/replica:0/task:0/device:CPU:0"; CheckRemoteMatMulExecutesOK(ctx, remote_device_name, local_device_name); TFE_Executor* executor = TFE_ContextGetExecutorForThread(ctx);
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Fri Apr 14 10:03:59 GMT 2023 - 19.3K bytes - Viewed (0) -
tensorflow/c/c_api_experimental.cc
flags->tf_xla_cpu_global_jit = true; flags->tf_xla_min_cluster_size = 1; } else { optimizer_options->set_global_jit_level(tensorflow::OptimizerOptions::OFF); } auto* gpu_options = config.mutable_gpu_options(); gpu_options->set_allow_growth(gpu_memory_allow_growth); (*config.mutable_device_count())["CPU"] = num_cpu_devices;
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 03:35:10 GMT 2024 - 29.4K bytes - Viewed (0) -
tensorflow/c/experimental/next_pluggable_device/tensor_pjrt_buffer_util_test.cc
StatusIs(error::INTERNAL, HasSubstr(absl::StrCat("PjRtClient for ", DEVICE_CPU, " is not type PjRtCApiClient")))); } TEST(TensorPjRtBufferUtilTest, GetPjRtCApiClientSuccess) { auto status = pjrt::PjrtApi(DEVICE_CPU); if (!status.ok()) { TF_ASSERT_OK(pjrt::SetPjrtApi(DEVICE_CPU, GetPjrtApi())); }
C++ - Registered: Tue Feb 27 12:39:08 GMT 2024 - Last Modified: Mon Oct 30 19:20:20 GMT 2023 - 7.2K bytes - Viewed (0) -
tensorflow/c/eager/parallel_device/parallel_device_test.cc
"/job:localhost/replica:0/task:0/device:CPU:0", "/job:localhost/replica:0/task:0/device:CPU:1"}; RegisterParallelDevice(context.get(), first_device_name, first_underlying_devices, status.get()); ASSERT_EQ(TF_GetCode(status.get()), TF_OK) << TF_Message(status.get()); // Create a second parallel device with the first parallel device and one // additional CPU. const char* second_device_name =
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Jul 08 23:47:35 GMT 2021 - 29.3K bytes - Viewed (1) -
tensorflow/c/eager/dlpack.cc
std::string device_type = parsed_name.type; int device_id = 0; if (parsed_name.has_id) { device_id = parsed_name.id; } ctx.device_id = device_id; if (device_type == "CPU") { ctx.device_type = DLDeviceType::kDLCPU; } else if (device_type == "GPU") { #if TENSORFLOW_USE_ROCM ctx.device_type = DLDeviceType::kDLROCM; #else ctx.device_type = DLDeviceType::kDLCUDA; #endif
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Feb 15 09:49:45 GMT 2024 - 12.8K bytes - Viewed (0) -
tensorflow/c/eager/c_api_distributed_test.cc
EXPECT_EQ(TF_GetCode(status), TF_OK) << TF_Message(status); const char task0_name[] = "/job:localhost/replica:0/task:0/device:CPU:0"; const char task1_name[] = "/job:localhost/replica:0/task:1/device:CPU:0"; const char task2_name[] = "/job:localhost/replica:0/task:2/device:CPU:0"; // Create one variable per task. TFE_TensorHandle* h0 = TestVariable(ctx, 1.0, task1_name);
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Feb 15 09:49:45 GMT 2024 - 23.5K bytes - Viewed (0) -
tensorflow/c/c_test_util.cc
TF_SetAttrType(desc, "T", TF_INT32); // Set device to CPU since there is no version of split for int32 on GPU // TODO(iga): Convert all these helpers and tests to use floats because // they are usually available on GPUs. After doing this, remove TF_SetDevice // call in c_api_function_test.cc TF_SetDevice(desc, "/cpu:0"); *op = TF_FinishOperation(desc, s); ASSERT_EQ(TF_OK, TF_GetCode(s)) << TF_Message(s);
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Fri Oct 15 03:16:52 GMT 2021 - 17.8K bytes - Viewed (2) -
tensorflow/c/eager/parallel_device/parallel_device_testlib.cc
// device. { TensorHandlePtr initial_value_cpu = FloatTensorHandle(20., status.get()); ASSERT_TRUE(TF_GetCode(status.get()) == TF_OK) << TF_Message(status.get()); std::array<TFE_TensorHandle*, 2> components{initial_value_cpu.get(), initial_value_cpu.get()}; TensorHandlePtr initial_value =
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Tue Jun 15 15:44:44 GMT 2021 - 12.5K bytes - Viewed (0)