- Sort Score
- Result 10 results
- Languages All
Results 11 - 16 of 16 for MatMul (0.18 sec)
-
tensorflow/c/eager/c_api_test_util.h
const tensorflow::string& device_name = ""); // Return an add op multiplying `a` by `b`. TFE_Op* AddOp(TFE_Context* ctx, TFE_TensorHandle* a, TFE_TensorHandle* b); // Return a matmul op multiplying `a` by `b`. TFE_Op* MatMulOp(TFE_Context* ctx, TFE_TensorHandle* a, TFE_TensorHandle* b); // Return an identity op. TFE_Op* IdentityOp(TFE_Context* ctx, TFE_TensorHandle* a);
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Mon Jul 17 23:43:59 UTC 2023 - 7.7K bytes - Viewed (0) -
tensorflow/c/eager/c_api_experimental_test.cc
TFE_TensorHandle* m = TestMatrixTensorHandle(ctx); TFE_Op* matmul = MatMulOp(ctx, m, m); TFE_TensorHandle* retvals[2] = {nullptr, nullptr}; int num_retvals = 2; TFE_Execute(matmul, &retvals[0], &num_retvals, status); EXPECT_EQ(1, num_retvals); EXPECT_EQ(TF_OK, TF_GetCode(status)) << TF_Message(status); TFE_DeleteOp(matmul); TFE_DeleteTensorHandle(m);
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Thu Aug 03 03:14:26 UTC 2023 - 31.5K bytes - Viewed (0) -
tensorflow/c/c_api_experimental_test.cc
TFE_ContextOptions* tfe_context_options_; TFE_Context* tfe_context_; }; TEST_F(ShapeInferenceTest, InfersShapesFromInputShapes) { TFE_Op* matmul_op; matmul_op = TFE_NewOp(tfe_context_, "MatMul", status_); CHECK_EQ(TF_OK, TF_GetCode(status_)) << TF_Message(status_); // Infer shape when everything is known. CheckOutputShapes(matmul_op,
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Tue Jan 17 22:27:52 UTC 2023 - 13.1K bytes - Viewed (0) -
tensorflow/c/eager/c_api_test_util.cc
return op; } TFE_Op* MatMulOp(TFE_Context* ctx, TFE_TensorHandle* a, TFE_TensorHandle* b) { TF_Status* status = TF_NewStatus(); TFE_Op* op = TFE_NewOp(ctx, "MatMul", status); CHECK_EQ(TF_OK, TF_GetCode(status)) << TF_Message(status); TFE_OpAddInput(op, a, status); CHECK_EQ(TF_OK, TF_GetCode(status)) << TF_Message(status); TFE_OpAddInput(op, b, status);
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Wed Feb 21 22:37:46 UTC 2024 - 23.5K bytes - Viewed (0) -
RELEASE.md
* `tf.config.experimental.enable_tensor_float_32_execution` * Disabling TensorFloat-32 execution now causes TPUs to use float32 precision for float32 matmuls and other ops. TPUs have always used bfloat16 precision for certain ops, like matmul, when such ops had float32 inputs. Now, disabling TensorFloat-32 by calling `tf.config.experimental.enable_tensor_float_32_execution(False)` will cause TPUs to use float32 precision for such ops instead of bfloat16.
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Tue Oct 22 14:33:53 UTC 2024 - 735.3K bytes - Viewed (0) -
okhttp/src/test/resources/okhttp3/internal/publicsuffix/public_suffix_list.dat
// Submitted by Alexander Lodin <******@****.***> yandexcloud.net storage.yandexcloud.net website.yandexcloud.net // YesCourse Pty Ltd : https://yescourse.com // Submitted by Atul Bhouraskar <atul@yescourse.com> official.academy // Yola : https://www.yola.com/ // Submitted by Stefano Rivera <******@****.***> yolasite.com // Yombo : https://yombo.net
Registered: Fri Nov 01 11:42:11 UTC 2024 - Last Modified: Wed Dec 20 23:27:07 UTC 2023 - 240.3K bytes - Viewed (0)