- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 17 for compute (0.86 sec)
-
tensorflow/c/c_test.c
TF_Status* s = TF_NewStatus(); TF_OpKernelConstruction_GetAttrType(ctx, "foobar", &type, s); TF_DeleteStatus(s); return NULL; } // A compute function. This will never actually get called in this test, it's // just nice to know that it compiles. void compute(void* kernel, TF_OpKernelContext* ctx) { TF_Tensor* input; TF_Status* s = TF_NewStatus(); TF_GetInput(ctx, 0, &input, s); TF_DeleteTensor(input);
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Wed Apr 24 20:50:35 GMT 2024 - 2.8K bytes - Viewed (0) -
tensorflow/c/experimental/gradients/nn_grad.cc
: forward_outputs_(f_outputs) { for (auto output : forward_outputs_) { if (output) { output->Ref(); } } } Status Compute(AbstractContext* ctx, absl::Span<AbstractTensorHandle* const> grad_outputs, absl::Span<AbstractTensorHandle*> grad_inputs) override { AbstractTensorHandle* upstream_grad = grad_outputs[0];
C++ - Registered: Tue Mar 26 12:39:09 GMT 2024 - Last Modified: Wed Feb 28 13:53:47 GMT 2024 - 5.7K bytes - Viewed (0) -
tensorflow/c/experimental/gradients/math_grad.cc
"Expected numeric or variant tensor, got dtype ", dtype); } return tensorflow::ops::Conj(ctx, input, output, name); } class AddGradientFunction : public GradientFunction { public: Status Compute(AbstractContext* ctx, absl::Span<AbstractTensorHandle* const> grad_outputs, absl::Span<AbstractTensorHandle*> grad_inputs) override { // TODO(b/161805092): Support broadcasting.
C++ - Registered: Tue Mar 26 12:39:09 GMT 2024 - Last Modified: Wed Feb 28 13:53:47 GMT 2024 - 15.2K bytes - Viewed (0) -
configure.py
native_cuda_compute_capabilities = get_native_cuda_compute_capabilities( environ_cp) if not native_cuda_compute_capabilities: default_cuda_compute_capabilities = _DEFAULT_CUDA_COMPUTE_CAPABILITIES else: default_cuda_compute_capabilities = native_cuda_compute_capabilities ask_cuda_compute_capabilities = ( 'Please specify a list of comma-separated CUDA compute capabilities '
Python - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 18:25:36 GMT 2024 - 53.8K bytes - Viewed (1) -
.bazelrc
build:cross_compile_linux_arm64_xla --crosstool_top=//tools/toolchains/cross_compile/cc:cross_compile_toolchain_suite # RBE cross-compile configs for Linux Aarch64 build:rbe_cross_compile_linux_arm64 --config=cross_compile_linux_arm64 build:rbe_cross_compile_linux_arm64 --config=rbe_cross_compile_base test:rbe_cross_compile_linux_arm64 --config=rbe_cross_compile_base # RBE cross-compile configs for XLA Linux Aarch64
Plain Text - Registered: Tue May 07 12:40:20 GMT 2024 - Last Modified: Thu May 02 19:34:20 GMT 2024 - 52.8K bytes - Viewed (2) -
tensorflow/c/experimental/gradients/array_grad.cc
#include "tensorflow/c/eager/abstract_context.h" namespace tensorflow { namespace gradients { namespace { class IdentityNGradientFunction : public GradientFunction { public: Status Compute(AbstractContext* ctx, absl::Span<AbstractTensorHandle* const> grad_outputs, absl::Span<AbstractTensorHandle*> grad_inputs) override { for (int i = 0; i < grad_outputs.size(); i++) {
C++ - Registered: Tue Apr 09 12:39:09 GMT 2024 - Last Modified: Wed Feb 28 13:53:47 GMT 2024 - 1.6K bytes - Viewed (0) -
tensorflow/c/eager/tape.h
// If provided, a non-null `forward_function` will be used instead of the // backward function (`backward_function_getter` / // `backward_function_deleter`) to compute jvps for this operation. If // `forward_function` is null, a GradientTape is used on the backward function // to compute the jvp, which will waste computation when executing eagerly. // // Unlike GradientTape::RecordOperation, Accumulate runs gradient computation
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Tue Apr 02 12:40:29 GMT 2024 - 47.2K bytes - Viewed (1) -
tensorflow/c/experimental/gradients/custom_gradient_test.cc
Status s = StatusFromTF_Status(status.get()); CHECK_EQ(errors::OK, s.code()) << s.message(); } }; class PassThroughGradientFunction : public GradientFunction { public: Status Compute(AbstractContext* ctx, absl::Span<AbstractTensorHandle* const> grad_outputs, absl::Span<AbstractTensorHandle*> grad_inputs) override { CHECK_EQ(grad_outputs.size(), 1);
C++ - Registered: Tue Mar 26 12:39:09 GMT 2024 - Last Modified: Wed Feb 28 13:53:47 GMT 2024 - 4.8K bytes - Viewed (0) -
tensorflow/c/c_api_test.cc
EXPECT_EQ(TF_OK, TF_GetCode(s_)) << TF_Message(s_); // Compare that the graphs match. GraphDef expected_gdef; GraphDef gdef; EXPECT_TRUE(GetGraphDef(expected_graph_, &expected_gdef)); EXPECT_TRUE(GetGraphDef(graph_, &gdef)); TF_EXPECT_GRAPH_EQ(expected_gdef, gdef); // Compare that the output of the gradients of both graphs match.
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 03:35:10 GMT 2024 - 96.9K bytes - Viewed (3) -
tensorflow/c/c_api_function.cc
control_output_names_vec.push_back(string(control_output_names[i])); } } // Compute body nodes. std::vector<const Node*> body_nodes; status->status = tensorflow::ComputeBodyNodes( fn_body, fn_name, num_opers, opers, input_nodes, &body_nodes); if (TF_GetCode(status) != TF_OK) return nullptr; // Compute body nodes. std::vector<const Node*> control_output_nodes;
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 03:35:10 GMT 2024 - 13.6K bytes - Viewed (2)