- Sort Score
- Result 10 results
- Languages All
Results 1 - 7 of 7 for backend (0.17 sec)
-
CONTRIBUTING.md
* [JavaScript/TypeScript license example](https://github.com/tensorflow/tensorboard/blob/master/tensorboard/components/tf_backend/backend.ts#L1) Bazel BUILD files also need to include a license section, e.g., [BUILD example](https://github.com/tensorflow/tensorflow/blob/master/tensorflow/core/BUILD#L61). #### C++ coding style
Plain Text - Registered: Tue May 07 12:40:20 GMT 2024 - Last Modified: Thu Mar 21 11:45:51 GMT 2024 - 15.6K bytes - Viewed (0) -
tensorflow/c/c_api_test.cc
} void NoOpDeallocator(void* data, size_t, void*) {} TEST(CAPI, MalformedTensor) { // See https://github.com/tensorflow/tensorflow/issues/7394 // num_dims = 0 implies a scalar, so should be backed by at least 4 bytes of // data. TF_Tensor* t = TF_NewTensor(TF_FLOAT, nullptr, 0, nullptr, 0, &NoOpDeallocator, nullptr); ASSERT_TRUE(t == nullptr); } TEST(CAPI, AllocateTensor) {
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 03:35:10 GMT 2024 - 96.9K bytes - Viewed (3) -
.bazelrc
build:mkl --define=build_with_mkl=true --define=enable_mkl=true build:mkl --define=tensorflow_mkldnn_contraction_kernel=0 build:mkl --define=build_with_openmp=true build:mkl -c opt # config to build OneDNN backend with a user specified threadpool. build:mkl_threadpool --define=build_with_mkl=true --define=enable_mkl=true build:mkl_threadpool --define=tensorflow_mkldnn_contraction_kernel=0
Plain Text - Registered: Tue May 07 12:40:20 GMT 2024 - Last Modified: Thu May 02 19:34:20 GMT 2024 - 52.8K bytes - Viewed (2) -
RELEASE.md
`tf.data.experimental.enable_debug_mode()`. * `tf.lite` * Enabled the new MLIR-based quantization backend by default * The new backend is used for 8 bits full integer post-training quantization * The new backend removes the redundant rescales and fixes some bugs (shared weight/bias, extremely small scales, etc)
Plain Text - Registered: Tue May 07 12:40:20 GMT 2024 - Last Modified: Mon Apr 29 19:17:57 GMT 2024 - 727.7K bytes - Viewed (8) -
tensorflow/c/eager/c_api_experimental.h
TF_CAPI_EXPORT TFE_TensorHandle* TFE_NewTensorHandleFromTensor( TFE_Context* ctx, TF_Tensor* t, TF_Status* status); // Create a packed TensorHandle with the given list of TensorHandles. // If `handles` are on the same device, assign the same device to the packed // handle; if `handles` are on different deivces, assign a CompositeDevice to // it. TF_CAPI_EXPORT extern TFE_TensorHandle* TFE_CreatePackedTensorHandle(
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Wed Feb 21 22:37:46 GMT 2024 - 39.5K bytes - Viewed (0) -
tensorflow/c/eager/tape.h
// While the Accumulate method is running (accumulating is True), any op // executions not forwarded to backward_tape should be ignored. bool accumulating; }; // A deque-backed stack, whose element references are not invalidated by // pushes and pops at the back. std::stack<AccumulatorCallState> call_state_; }; // Template instantiations here
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Tue Apr 02 12:40:29 GMT 2024 - 47.2K bytes - Viewed (1) -
tensorflow/c/eager/c_api_distributed_test.cc
TFE_TensorHandle* packed_handle = TFE_CreatePackedTensorHandle(ctx, handles.data(), &num_replicas, status); ASSERT_EQ(TF_GetCode(status), TF_OK) << TF_Message(status); EXPECT_EQ(TFE_TensorHandleDataType(packed_handle), TF_RESOURCE); EXPECT_EQ(TFE_TensorHandleNumDims(packed_handle, status), 0); EXPECT_EQ(TFE_TensorHandleNumElements(packed_handle, status), 1);
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Feb 15 09:49:45 GMT 2024 - 23.5K bytes - Viewed (0)