- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 39 for That (0.15 sec)
-
tensorflow/c/c_test_util.cc
values[0] = v; return TF_NewTensor(TF_FLOAT, nullptr, 0, values, num_bytes, &FloatDeallocator, nullptr); } // All the *Helper methods are used as a workaround for the restrictions that // one cannot call ASSERT_* methods in non-void-returning functions (when // exceptions are disabled during compilation) void PlaceholderHelper(TF_Graph* graph, TF_Status* s, const char* name,
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Fri Oct 15 03:16:52 GMT 2021 - 17.8K bytes - Viewed (2) -
tensorflow/c/eager/dlpack.cc
bool valid = true; int64_t expected_stride = 1; for (int i = ndim - 1; i >= 0; --i) { // Empty tensors are always compact regardless of strides. if (shape_arr[i] == 0) return true; // Note that dimensions with size=1 can have any stride. if (shape_arr[i] != 1 && stride_arr[i] != expected_stride) { valid = false; } expected_stride *= shape_arr[i]; } return valid; }
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Feb 15 09:49:45 GMT 2024 - 12.8K bytes - Viewed (0) -
tensorflow/c/experimental/grappler/grappler.cc
==============================================================================*/ // This file extends/implements core graph optimizer base classes in terms of // the C API defined in grappler.h. A class "CSomething" represents a // "Something" that can be manipulated via calls in the C interface and a C // struct called "TP_Something". #include "tensorflow/c/experimental/grappler/grappler.h" #include <algorithm> #include <cstddef> #include <cstring>
C++ - Registered: Tue Feb 27 12:39:08 GMT 2024 - Last Modified: Wed Sep 06 19:12:29 GMT 2023 - 15K bytes - Viewed (1) -
tensorflow/c/experimental/next_pluggable_device/tensor_pjrt_buffer_util_test.cc
} TEST(TensorPjRtBufferUtilTest, GetPjRtCBufferFromTensorNoBuffer) { auto allocator = std::make_unique<AsyncValueAllocator>(); tensorflow::Tensor tensor(allocator.get(), DT_FLOAT, {1}); EXPECT_THAT( GetPjRtCBufferFromTensor(&tensor), StatusIs(error::INTERNAL, HasSubstr(absl::StrCat( "Input tensor does not have PjRtBuffer")))); }
C++ - Registered: Tue Feb 27 12:39:08 GMT 2024 - Last Modified: Mon Oct 30 19:20:20 GMT 2023 - 7.2K bytes - Viewed (0) -
tensorflow/c/eager/parallel_device/parallel_device_lib.cc
ExecutionState::kIdle; // Tells the worker thread that there is new work. tensorflow::condition_variable start_execute_; // The worker thread notifies that work has finished. tensorflow::condition_variable finished_execute_; // Notifies a StartExecute that the previous Join has finished. tensorflow::condition_variable finished_join_;
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Fri Feb 09 07:47:20 GMT 2024 - 25.4K bytes - Viewed (1) -
tensorflow/c/eager/c_api_distributed_test.cc
TFE_TensorHandle* h0 = TestVariable(ctx, 1.0, task1_name); TFE_TensorHandle* h1 = TestVariable(ctx, 2.0, task2_name); TFE_TensorHandle* h2 = TestVariable(ctx, 3.0, task0_name); // Add a sync point to make sure that variables have been initialized // before the function execution starts. TFE_ContextAsyncWait(ctx, status); EXPECT_EQ(TF_OK, TF_GetCode(status)) << TF_Message(status);
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Feb 15 09:49:45 GMT 2024 - 23.5K bytes - Viewed (0) -
tensorflow/c/eager/parallel_device/parallel_device_test.cc
ExpectScalarEq<float>(result_components[1].get(), 3.); } TEST(PARALLEL_DEVICE, TestCollectiveSync) { TestCollective(/*async=*/false); } // Note that ops on the parallel device currently don't execute // asynchronously. The test is just that we don't get deadlocks. TEST(PARALLEL_DEVICE, TestCollectiveAsync) { TestCollective(/*async=*/true); } void RegisterCollectiveMulFunction(TFE_Context* context,
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Jul 08 23:47:35 GMT 2021 - 29.3K bytes - Viewed (1) -
tensorflow/c/eager/parallel_device/parallel_device.cc
// // TODO(allenl): There may be smarter ways to do this copy in some // cases, i.e. with a collective broadcast. We'll need to be careful // about things that are taken as inputs on the host or on their // existing device (for multi-device functions). std::unique_ptr<ParallelTensor> parallel_tensor( parallel_device.CopyToParallelDevice(
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Wed Mar 29 22:05:31 GMT 2023 - 18.3K bytes - Viewed (0) -
tensorflow/c/eager/parallel_device/parallel_device_testlib.cc
variable->Assign(context, initial_value.get(), status.get()); ASSERT_TRUE(TF_GetCode(status.get()) == TF_OK) << TF_Message(status.get()); } // Read from the variable and verify that we have a parallel tensor. { TensorHandlePtr read = variable->Read(context, status.get()); ASSERT_TRUE(TF_GetCode(status.get()) == TF_OK) << TF_Message(status.get()); std::array<TensorHandlePtr, 2> components;
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Tue Jun 15 15:44:44 GMT 2021 - 12.5K bytes - Viewed (0) -
tensorflow/c/c_api_function.cc
fn_name, "'"); output_tensors->emplace_back(node, idx); } return absl::OkStatus(); } // Populates `body_nodes` with the nodes that will become function's body. // Performs various checks. Status ComputeBodyNodes( const TF_Graph* fn_body, const char* fn_name, int num_opers, const TF_Operation* const* opers,
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 03:35:10 GMT 2024 - 13.6K bytes - Viewed (2)